Sample records for event analysis atheana

  1. Philosophy of ATHEANA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bley, D.C.; Cooper, S.E.; Forester, J.A.

    ATHEANA, a second-generation Human Reliability Analysis (HRA) method integrates advances in psychology with engineering, human factors, and Probabilistic Risk Analysis (PRA) disciplines to provide an HRA quantification process and PRA modeling interface that can accommodate and represent human performance in real nuclear power plant events. The method uses the characteristics of serious accidents identified through retrospective analysis of serious operational events to set priorities in a search process for significant human failure events, unsafe acts, and error-forcing context (unfavorable plant conditions combined with negative performance-shaping factors). ATHEANA has been tested in a demonstration project at an operating pressurized water reactor.

  2. Results of a nuclear power plant application of A New Technique for Human Error Analysis (ATHEANA)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Whitehead, D.W.; Forester, J.A.; Bley, D.C.

    1998-03-01

    A new method to analyze human errors has been demonstrated at a pressurized water reactor (PWR) nuclear power plant. This was the first application of the new method referred to as A Technique for Human Error Analysis (ATHEANA). The main goals of the demonstration were to test the ATHEANA process as described in the frame-of-reference manual and the implementation guideline, test a training package developed for the method, test the hypothesis that plant operators and trainers have significant insight into the error-forcing-contexts (EFCs) that can make unsafe actions (UAs) more likely, and to identify ways to improve the method andmore » its documentation. A set of criteria to evaluate the success of the ATHEANA method as used in the demonstration was identified. A human reliability analysis (HRA) team was formed that consisted of an expert in probabilistic risk assessment (PRA) with some background in HRA (not ATHEANA) and four personnel from the nuclear power plant. Personnel from the plant included two individuals from their PRA staff and two individuals from their training staff. Both individuals from training are currently licensed operators and one of them was a senior reactor operator on shift until a few months before the demonstration. The demonstration was conducted over a 5-month period and was observed by members of the Nuclear Regulatory Commission`s ATHEANA development team, who also served as consultants to the HRA team when necessary. Example results of the demonstration to date, including identified human failure events (HFEs), UAs, and EFCs are discussed. Also addressed is how simulator exercises are used in the ATHEANA demonstration project.« less

  3. Technical Basis and Implementation Guidelines for a Technique for Human Event Analysis (ATHEANA)

    DTIC Science & Technology

    2000-05-01

    posted at NRC’s Web site address www.nrc.gov/NRC/NUREGS/indexnum.html are updated regularly and may differ from the last printed version. Non-NRC...distinctly different in that it provides structured search schemes for finding such EFCs, by using and integrating knowledge and experience in...Learned from Serious Accidents The record of significant incidents in nuclear power plant NPP operations shows a substantially different picture of

  4. EventThread: Visual Summarization and Stage Analysis of Event Sequence Data.

    PubMed

    Guo, Shunan; Xu, Ke; Zhao, Rongwen; Gotz, David; Zha, Hongyuan; Cao, Nan

    2018-01-01

    Event sequence data such as electronic health records, a person's academic records, or car service records, are ordered series of events which have occurred over a period of time. Analyzing collections of event sequences can reveal common or semantically important sequential patterns. For example, event sequence analysis might reveal frequently used care plans for treating a disease, typical publishing patterns of professors, and the patterns of service that result in a well-maintained car. It is challenging, however, to visually explore large numbers of event sequences, or sequences with large numbers of event types. Existing methods focus on extracting explicitly matching patterns of events using statistical analysis to create stages of event progression over time. However, these methods fail to capture latent clusters of similar but not identical evolutions of event sequences. In this paper, we introduce a novel visualization system named EventThread which clusters event sequences into threads based on tensor analysis and visualizes the latent stage categories and evolution patterns by interactively grouping the threads by similarity into time-specific clusters. We demonstrate the effectiveness of EventThread through usage scenarios in three different application domains and via interviews with an expert user.

  5. Bayesian analysis of rare events

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Straub, Daniel, E-mail: straub@tum.de; Papaioannou, Iason; Betz, Wolfgang

    2016-06-01

    In many areas of engineering and science there is an interest in predicting the probability of rare events, in particular in applications related to safety and security. Increasingly, such predictions are made through computer models of physical systems in an uncertainty quantification framework. Additionally, with advances in IT, monitoring and sensor technology, an increasing amount of data on the performance of the systems is collected. This data can be used to reduce uncertainty, improve the probability estimates and consequently enhance the management of rare events and associated risks. Bayesian analysis is the ideal method to include the data into themore » probabilistic model. It ensures a consistent probabilistic treatment of uncertainty, which is central in the prediction of rare events, where extrapolation from the domain of observation is common. We present a framework for performing Bayesian updating of rare event probabilities, termed BUS. It is based on a reinterpretation of the classical rejection-sampling approach to Bayesian analysis, which enables the use of established methods for estimating probabilities of rare events. By drawing upon these methods, the framework makes use of their computational efficiency. These methods include the First-Order Reliability Method (FORM), tailored importance sampling (IS) methods and Subset Simulation (SuS). In this contribution, we briefly review these methods in the context of the BUS framework and investigate their applicability to Bayesian analysis of rare events in different settings. We find that, for some applications, FORM can be highly efficient and is surprisingly accurate, enabling Bayesian analysis of rare events with just a few model evaluations. In a general setting, BUS implemented through IS and SuS is more robust and flexible.« less

  6. Bayesian analysis of rare events

    NASA Astrophysics Data System (ADS)

    Straub, Daniel; Papaioannou, Iason; Betz, Wolfgang

    2016-06-01

    In many areas of engineering and science there is an interest in predicting the probability of rare events, in particular in applications related to safety and security. Increasingly, such predictions are made through computer models of physical systems in an uncertainty quantification framework. Additionally, with advances in IT, monitoring and sensor technology, an increasing amount of data on the performance of the systems is collected. This data can be used to reduce uncertainty, improve the probability estimates and consequently enhance the management of rare events and associated risks. Bayesian analysis is the ideal method to include the data into the probabilistic model. It ensures a consistent probabilistic treatment of uncertainty, which is central in the prediction of rare events, where extrapolation from the domain of observation is common. We present a framework for performing Bayesian updating of rare event probabilities, termed BUS. It is based on a reinterpretation of the classical rejection-sampling approach to Bayesian analysis, which enables the use of established methods for estimating probabilities of rare events. By drawing upon these methods, the framework makes use of their computational efficiency. These methods include the First-Order Reliability Method (FORM), tailored importance sampling (IS) methods and Subset Simulation (SuS). In this contribution, we briefly review these methods in the context of the BUS framework and investigate their applicability to Bayesian analysis of rare events in different settings. We find that, for some applications, FORM can be highly efficient and is surprisingly accurate, enabling Bayesian analysis of rare events with just a few model evaluations. In a general setting, BUS implemented through IS and SuS is more robust and flexible.

  7. Survival analysis: Part I — analysis of time-to-event

    PubMed Central

    2018-01-01

    Length of time is a variable often encountered during data analysis. Survival analysis provides simple, intuitive results concerning time-to-event for events of interest, which are not confined to death. This review introduces methods of analyzing time-to-event. The Kaplan-Meier survival analysis, log-rank test, and Cox proportional hazards regression modeling method are described with examples of hypothetical data. PMID:29768911

  8. Event-based analysis of free-living behaviour.

    PubMed

    Granat, Malcolm H

    2012-11-01

    The quantification of free-living physical activities is important in understanding how physical activity and sedentary behaviour impact on health and also on how interventions might modify free-living behaviour to enhance health. Quantification, and the terminology used, has in many ways been determined by the choice of measurement technique. The inter-related issues around measurement devices and terminology used are explored. This paper proposes a terminology and a systematic approach for the analysis of free-living activity information using event-based activity data. The event-based approach uses a flexible hierarchical classification of events and, dependent on the research question, analysis can then be undertaken on a selection of these events. The quantification of free-living behaviour is therefore the result of the analysis on the patterns of these chosen events. The application of this approach is illustrated with results from a range of published studies by our group showing how event-based analysis provides a flexible yet robust method of addressing the research question(s) and provides a deeper insight into free-living behaviour. It is proposed that it is through event-based analysis we can more clearly understand how behaviour is related to health and also how we can produce more relevant outcome measures.

  9. Research on Visual Analysis Methods of Terrorism Events

    NASA Astrophysics Data System (ADS)

    Guo, Wenyue; Liu, Haiyan; Yu, Anzhu; Li, Jing

    2016-06-01

    Under the situation that terrorism events occur more and more frequency throughout the world, improving the response capability of social security incidents has become an important aspect to test governments govern ability. Visual analysis has become an important method of event analysing for its advantage of intuitive and effective. To analyse events' spatio-temporal distribution characteristics, correlations among event items and the development trend, terrorism event's spatio-temporal characteristics are discussed. Suitable event data table structure based on "5W" theory is designed. Then, six types of visual analysis are purposed, and how to use thematic map and statistical charts to realize visual analysis on terrorism events is studied. Finally, experiments have been carried out by using the data provided by Global Terrorism Database, and the results of experiments proves the availability of the methods.

  10. Joint Attributes and Event Analysis for Multimedia Event Detection.

    PubMed

    Ma, Zhigang; Chang, Xiaojun; Xu, Zhongwen; Sebe, Nicu; Hauptmann, Alexander G

    2017-06-15

    Semantic attributes have been increasingly used the past few years for multimedia event detection (MED) with promising results. The motivation is that multimedia events generally consist of lower level components such as objects, scenes, and actions. By characterizing multimedia event videos with semantic attributes, one could exploit more informative cues for improved detection results. Much existing work obtains semantic attributes from images, which may be suboptimal for video analysis since these image-inferred attributes do not carry dynamic information that is essential for videos. To address this issue, we propose to learn semantic attributes from external videos using their semantic labels. We name them video attributes in this paper. In contrast with multimedia event videos, these external videos depict lower level contents such as objects, scenes, and actions. To harness video attributes, we propose an algorithm established on a correlation vector that correlates them to a target event. Consequently, we could incorporate video attributes latently as extra information into the event detector learnt from multimedia event videos in a joint framework. To validate our method, we perform experiments on the real-world large-scale TRECVID MED 2013 and 2014 data sets and compare our method with several state-of-the-art algorithms. The experiments show that our method is advantageous for MED.

  11. Negated bio-events: analysis and identification

    PubMed Central

    2013-01-01

    Background Negation occurs frequently in scientific literature, especially in biomedical literature. It has previously been reported that around 13% of sentences found in biomedical research articles contain negation. Historically, the main motivation for identifying negated events has been to ensure their exclusion from lists of extracted interactions. However, recently, there has been a growing interest in negative results, which has resulted in negation detection being identified as a key challenge in biomedical relation extraction. In this article, we focus on the problem of identifying negated bio-events, given gold standard event annotations. Results We have conducted a detailed analysis of three open access bio-event corpora containing negation information (i.e., GENIA Event, BioInfer and BioNLP’09 ST), and have identified the main types of negated bio-events. We have analysed the key aspects of a machine learning solution to the problem of detecting negated events, including selection of negation cues, feature engineering and the choice of learning algorithm. Combining the best solutions for each aspect of the problem, we propose a novel framework for the identification of negated bio-events. We have evaluated our system on each of the three open access corpora mentioned above. The performance of the system significantly surpasses the best results previously reported on the BioNLP’09 ST corpus, and achieves even better results on the GENIA Event and BioInfer corpora, both of which contain more varied and complex events. Conclusions Recently, in the field of biomedical text mining, the development and enhancement of event-based systems has received significant interest. The ability to identify negated events is a key performance element for these systems. We have conducted the first detailed study on the analysis and identification of negated bio-events. Our proposed framework can be integrated with state-of-the-art event extraction systems. The

  12. Event coincidence analysis for quantifying statistical interrelationships between event time series. On the role of flood events as triggers of epidemic outbreaks

    NASA Astrophysics Data System (ADS)

    Donges, J. F.; Schleussner, C.-F.; Siegmund, J. F.; Donner, R. V.

    2016-05-01

    Studying event time series is a powerful approach for analyzing the dynamics of complex dynamical systems in many fields of science. In this paper, we describe the method of event coincidence analysis to provide a framework for quantifying the strength, directionality and time lag of statistical interrelationships between event series. Event coincidence analysis allows to formulate and test null hypotheses on the origin of the observed interrelationships including tests based on Poisson processes or, more generally, stochastic point processes with a prescribed inter-event time distribution and other higher-order properties. Applying the framework to country-level observational data yields evidence that flood events have acted as triggers of epidemic outbreaks globally since the 1950s. Facing projected future changes in the statistics of climatic extreme events, statistical techniques such as event coincidence analysis will be relevant for investigating the impacts of anthropogenic climate change on human societies and ecosystems worldwide.

  13. Interpretation Analysis as a Competitive Event.

    ERIC Educational Resources Information Center

    Nading, Robert M.

    Interpretation analysis is a new and interesting event on the forensics horizon which appears to be attracting an ever larger number of supporters. This event, developed by Larry Lambert of Ball State University in 1989, requires a student to perform all three disciplines of forensic competition (interpretation, public speaking, and limited…

  14. Second-Order Analysis of Semiparametric Recurrent Event Processes

    PubMed Central

    Guan, Yongtao

    2011-01-01

    Summary A typical recurrent event dataset consists of an often large number of recurrent event processes, each of which contains multiple event times observed from an individual during a followup period. Such data have become increasingly available in medical and epidemiological studies. In this paper, we introduce novel procedures to conduct second-order analysis for a flexible class of semiparametric recurrent event processes. Such an analysis can provide useful information regarding the dependence structure within each recurrent event process. Specifically, we will use the proposed procedures to test whether the individual recurrent event processes are all Poisson processes and to suggest sensible alternative models for them if they are not. We apply these procedures to a well-known recurrent event dataset on chronic granulomatous disease and an epidemiological dataset on Meningococcal disease cases in Merseyside, UK to illustrate their practical value. PMID:21361885

  15. Poisson-event-based analysis of cell proliferation.

    PubMed

    Summers, Huw D; Wills, John W; Brown, M Rowan; Rees, Paul

    2015-05-01

    A protocol for the assessment of cell proliferation dynamics is presented. This is based on the measurement of cell division events and their subsequent analysis using Poisson probability statistics. Detailed analysis of proliferation dynamics in heterogeneous populations requires single cell resolution within a time series analysis and so is technically demanding to implement. Here, we show that by focusing on the events during which cells undergo division rather than directly on the cells themselves a simplified image acquisition and analysis protocol can be followed, which maintains single cell resolution and reports on the key metrics of cell proliferation. The technique is demonstrated using a microscope with 1.3 μm spatial resolution to track mitotic events within A549 and BEAS-2B cell lines, over a period of up to 48 h. Automated image processing of the bright field images using standard algorithms within the ImageJ software toolkit yielded 87% accurate recording of the manually identified, temporal, and spatial positions of the mitotic event series. Analysis of the statistics of the interevent times (i.e., times between observed mitoses in a field of view) showed that cell division conformed to a nonhomogeneous Poisson process in which the rate of occurrence of mitotic events, λ exponentially increased over time and provided values of the mean inter mitotic time of 21.1 ± 1.2 hours for the A549 cells and 25.0 ± 1.1 h for the BEAS-2B cells. Comparison of the mitotic event series for the BEAS-2B cell line to that predicted by random Poisson statistics indicated that temporal synchronisation of the cell division process was occurring within 70% of the population and that this could be increased to 85% through serum starvation of the cell culture. © 2015 International Society for Advancement of Cytometry.

  16. Surface Management System Departure Event Data Analysis

    NASA Technical Reports Server (NTRS)

    Monroe, Gilena A.

    2010-01-01

    This paper presents a data analysis of the Surface Management System (SMS) performance of departure events, including push-back and runway departure events.The paper focuses on the detection performance, or the ability to detect departure events, as well as the prediction performance of SMS. The results detail a modest overall detection performance of push-back events and a significantly high overall detection performance of runway departure events. The overall detection performance of SMS for push-back events is approximately 55%.The overall detection performance of SMS for runway departure events nears 100%. This paper also presents the overall SMS prediction performance for runway departure events as well as the timeliness of the Aircraft Situation Display for Industry data source for SMS predictions.

  17. Time-to-first-event versus recurrent-event analysis: points to consider for selecting a meaningful analysis strategy in clinical trials with composite endpoints.

    PubMed

    Rauch, Geraldine; Kieser, Meinhard; Binder, Harald; Bayes-Genis, Antoni; Jahn-Eimermacher, Antje

    2018-05-01

    Composite endpoints combining several event types of clinical interest often define the primary efficacy outcome in cardiologic trials. They are commonly evaluated as time-to-first-event, thereby following the recommendations of regulatory agencies. However, to assess the patient's full disease burden and to identify preventive factors or interventions, subsequent events following the first one should be considered as well. This is especially important in cohort studies and RCTs with a long follow-up leading to a higher number of observed events per patients. So far, there exist no recommendations which approach should be preferred. Recently, the Cardiovascular Round Table of the European Society of Cardiology indicated the need to investigate "how to interpret results if recurrent-event analysis results differ […] from time-to-first-event analysis" (Anker et al., Eur J Heart Fail 18:482-489, 2016). This work addresses this topic by means of a systematic simulation study. This paper compares two common analysis strategies for composite endpoints differing with respect to the incorporation of recurrent events for typical data scenarios motivated by a clinical trial. We show that the treatment effects estimated from a time-to-first-event analysis (Cox model) and a recurrent-event analysis (Andersen-Gill model) can systematically differ, particularly in cardiovascular trials. Moreover, we provide guidance on how to interpret these results and recommend points to consider for the choice of a meaningful analysis strategy. When planning trials with a composite endpoint, researchers, and regulatory agencies should be aware that the model choice affects the estimated treatment effect and its interpretation.

  18. Second-order analysis of semiparametric recurrent event processes.

    PubMed

    Guan, Yongtao

    2011-09-01

    A typical recurrent event dataset consists of an often large number of recurrent event processes, each of which contains multiple event times observed from an individual during a follow-up period. Such data have become increasingly available in medical and epidemiological studies. In this article, we introduce novel procedures to conduct second-order analysis for a flexible class of semiparametric recurrent event processes. Such an analysis can provide useful information regarding the dependence structure within each recurrent event process. Specifically, we will use the proposed procedures to test whether the individual recurrent event processes are all Poisson processes and to suggest sensible alternative models for them if they are not. We apply these procedures to a well-known recurrent event dataset on chronic granulomatous disease and an epidemiological dataset on meningococcal disease cases in Merseyside, United Kingdom to illustrate their practical value. © 2011, The International Biometric Society.

  19. Event Reports Promoting Root Cause Analysis.

    PubMed

    Pandit, Swananda; Gong, Yang

    2016-01-01

    Improving health is the sole objective of medical care. Unfortunately, mishaps or patient safety events happen during the care. If the safety events were collected effectively, they would help identify patterns, underlying causes, and ultimately generate proactive and remedial solutions for prevention of recurrence. Based on the AHRQ Common Formats, we examine the quality of patient safety incident reports and describe the initial data requirement that can support and accelerate effective root cause analysis. The ultimate goal is to develop a knowledge base of patient safety events and their common solutions which can be readily available for sharing and learning.

  20. Parallel Event Analysis Under Unix

    NASA Astrophysics Data System (ADS)

    Looney, S.; Nilsson, B. S.; Oest, T.; Pettersson, T.; Ranjard, F.; Thibonnier, J.-P.

    The ALEPH experiment at LEP, the CERN CN division and Digital Equipment Corp. have, in a joint project, developed a parallel event analysis system. The parallel physics code is identical to ALEPH's standard analysis code, ALPHA, only the organisation of input/output is changed. The user may switch between sequential and parallel processing by simply changing one input "card". The initial implementation runs on an 8-node DEC 3000/400 farm, using the PVM software, and exhibits a near-perfect speed-up linearity, reducing the turn-around time by a factor of 8.

  1. Hydrometeorological Analysis of Flooding Events in San Antonio, TX

    NASA Astrophysics Data System (ADS)

    Chintalapudi, S.; Sharif, H.; Elhassan, A.

    2008-12-01

    South Central Texas is particularly vulnerable to floods due to: proximity to a moist air source (the Gulf of Mexico); the Balcones Escarpment, which concentrates rainfall runoff; a tendency for synoptic scale features to become cut-off and stall over the area; and decaying tropical cyclones stalling over the area. The San Antonio Metropolitan Area is the 7th largest city in the nation, one of the most flash-flood prone regions in North America, and has experienced a number of flooding events in the last decade (1998, 2002, 2004, and 2007). Research is being conducted to characterize the meteorological conditions that lead to these events and apply the rainfall and watershed characteristics data to recreate the runoff events using a two- dimensional, physically-based, distributed-parameter hydrologic model. The physically based, distributed-parameter Gridded Surface Subsurface Hydrologic Analysis (GSSHA) hydrological model was used for simulating the watershed response to these storm events. Finally observed discharges were compared to GSSHA model discharges for these storm events. Analysis of the some of these events will be presented.

  2. Glaucoma progression detection: agreement, sensitivity, and specificity of expert visual field evaluation, event analysis, and trend analysis.

    PubMed

    Antón, Alfonso; Pazos, Marta; Martín, Belén; Navero, José Manuel; Ayala, Miriam Eleonora; Castany, Marta; Martínez, Patricia; Bardavío, Javier

    2013-01-01

    To assess sensitivity, specificity, and agreement among automated event analysis, automated trend analysis, and expert evaluation to detect glaucoma progression. This was a prospective study that included 37 eyes with a follow-up of 36 months. All had glaucomatous disks and fields and performed reliable visual fields every 6 months. Each series of fields was assessed with 3 different methods: subjective assessment by 2 independent teams of glaucoma experts, glaucoma/guided progression analysis (GPA) event analysis, and GPA (visual field index-based) trend analysis. Kappa agreement coefficient between methods and sensitivity and specificity for each method using expert opinion as gold standard were calculated. The incidence of glaucoma progression was 16% to 18% in 3 years but only 3 cases showed progression with all 3 methods. Kappa agreement coefficient was high (k=0.82) between subjective expert assessment and GPA event analysis, and only moderate between these two and GPA trend analysis (k=0.57). Sensitivity and specificity for GPA event and GPA trend analysis were 71% and 96%, and 57% and 93%, respectively. The 3 methods detected similar numbers of progressing cases. The GPA event analysis and expert subjective assessment showed high agreement between them and moderate agreement with GPA trend analysis. In a period of 3 years, both methods of GPA analysis offered high specificity, event analysis showed 83% sensitivity, and trend analysis had a 66% sensitivity.

  3. Event shape analysis of deep inelastic scattering events with a large rapidity gap at HERA

    NASA Astrophysics Data System (ADS)

    ZEUS Collaboration; Breitweg, J.; Derrick, M.; Krakauer, D.; Magill, S.; Mikunas, D.; Musgrave, B.; Repond, J.; Stanek, R.; Talaga, R. L.; Yoshida, R.; Zhang, H.; Mattingly, M. C. K.; Anselmo, F.; Antonioli, P.; Bari, G.; Basile, M.; Bellagamba, L.; Boscherini, D.; Bruni, A.; Bruni, G.; Cara Romeo, G.; Castellini, G.; Cifarelli, L.; Cindolo, F.; Contin, A.; Corradi, M.; de Pasquale, S.; Gialas, I.; Giusti, P.; Iacobucci, G.; Laurenti, G.; Levi, G.; Margotti, A.; Massam, T.; Nania, R.; Palmonari, F.; Pesci, A.; Polini, A.; Ricci, F.; Sartorelli, G.; Zamora Garcia, Y.; Zichichi, A.; Amelung, C.; Bornheim, A.; Brock, I.; Coböken, K.; Crittenden, J.; Deffner, R.; Eckert, M.; Grothe, M.; Hartmann, H.; Heinloth, K.; Heinz, L.; Hilger, E.; Jakob, H.-P.; Katz, U. F.; Kerger, R.; Paul, E.; Pfeiffer, M.; Rembser, Ch.; Stamm, J.; Wedemeyer, R.; Wieber, H.; Bailey, D. S.; Campbell-Robson, S.; Cottingham, W. N.; Foster, B.; Hall-Wilton, R.; Hayes, M. E.; Heath, G. P.; Heath, H. F.; McFall, J. D.; Piccioni, D.; Roff, D. G.; Tapper, R. J.; Arneodo, M.; Ayad, R.; Capua, M.; Garfagnini, A.; Iannotti, L.; Schioppa, M.; Susinno, G.; Kim, J. Y.; Lee, J. H.; Lim, I. T.; Pac, M. Y.; Caldwell, A.; Cartiglia, N.; Jing, Z.; Liu, W.; Mellado, B.; Parsons, J. A.; Ritz, S.; Sampson, S.; Sciulli, F.; Straub, P. B.; Zhu, Q.; Borzemski, P.; Chwastowski, J.; Eskreys, A.; Figiel, J.; Klimek, K.; Przybycień , M. B.; Zawiejski, L.; Adamczyk, L.; Bednarek, B.; Bukowy, M.; Jeleń , K.; Kisielewska, D.; Kowalski, T.; Przybycień , M.; Rulikowska-Zarȩ Bska, E.; Suszycki, L.; Zaja C, J.; Duliń Ski, Z.; Kotań Ski, A.; Abbiendi, G.; Bauerdick, L. A. T.; Behrens, U.; Beier, H.; Bienlein, J. K.; Cases, G.; Deppe, O.; Desler, K.; Drews, G.; Fricke, U.; Gilkinson, D. J.; Glasman, C.; Göttlicher, P.; Haas, T.; Hain, W.; Hasell, D.; Johnson, K. F.; Kasemann, M.; Koch, W.; Kötz, U.; Kowalski, H.; Labs, J.; Lindemann, L.; Löhr, B.; Löwe, M.; Mań Czak, O.; Milewski, J.; Monteiro, T.; Ng, J. S. T.; Notz, D.; Ohrenberg, K.; Park, I. H.; Pellegrino, A.; Pelucchi, F.; Piotrzkowski, K.; Roco, M.; Rohde, M.; Roldán, J.; Ryan, J. J.; Savin, A. A.; Schneekloth, U.; Selonke, F.; Surrow, B.; Tassi, E.; Voß, T.; Westphal, D.; Wolf, G.; Wollmer, U.; Youngman, C.; Zsolararnecki, A. F.; Zeuner, W.; Burow, B. D.; Grabosch, H. J.; Meyer, A.; Schlenstedt, S.; Barbagli, G.; Gallo, E.; Pelfer, P.; Maccarrone, G.; Votano, L.; Bamberger, A.; Eisenhardt, S.; Markun, P.; Trefzger, T.; Wölfle, S.; Bromley, J. T.; Brook, N. H.; Bussey, P. J.; Doyle, A. T.; MacDonald, N.; Saxon, D. H.; Sinclair, L. E.; Strickland, E.; Waugh, R.; Bohnet, I.; Gendner, N.; Holm, U.; Meyer-Larsen, A.; Salehi, H.; Wick, K.; Gladilin, L. K.; Horstmann, D.; Kçira, D.; Klanner, R.; Lohrmann, E.; Poelz, G.; Schott, W.; Zetsche, F.; Bacon, T. C.; Butterworth, I.; Cole, J. E.; Howell, G.; Hung, B. H. Y.; Lamberti, L.; Long, K. R.; Miller, D. B.; Pavel, N.; Prinias, A.; Sedgbeer, J. K.; Sideris, D.; Walker, R.; Mallik, U.; Wang, S. M.; Wu, J. T.; Cloth, P.; Filges, D.; Fleck, J. I.; Ishii, T.; Kuze, M.; Suzuki, I.; Tokushuku, K.; Yamada, S.; Yamauchi, K.; Yamazaki, Y.; Hong, S. J.; Lee, S. B.; Nam, S. W.; Park, S. K.; Barreiro, F.; Fernández, J. P.; García, G.; Graciani, R.; Hernández, J. M.; Hervás, L.; Labarga, L.; Martínez, M.; del Peso, J.; Puga, J.; Terrón, J.; de Trocóniz, J. F.; Corriveau, F.; Hanna, D. S.; Hartmann, J.; Hung, L. W.; Murray, W. N.; Ochs, A.; Riveline, M.; Stairs, D. G.; St-Laurent, M.; Ullmann, R.; Tsurugai, T.; Bashkirov, V.; Dolgoshein, B. A.; Stifutkin, A.; Bashindzhagyan, G. L.; Ermolov, P. F.; Golubkov, Yu. A.; Khein, L. A.; Korotkova, N. A.; Korzhavina, I. A.; Kuzmin, V. A.; Lukina, O. Yu.; Proskuryakov, A. S.; Shcheglova, L. M.; Solomin, A. N.; Zotkin, S. A.; Bokel, C.; Botje, M.; Brümmer, N.; Chlebana, F.; Engelen, J.; Koffeman, E.; Kooijman, P.; van Sighem, A.; Tiecke, H.; Tuning, N.; Verkerke, W.; Vossebeld, J.; Vreeswijk, M.; Wiggers, L.; de Wolf, E.; Acosta, D.; Bylsma, B.; Durkin, L. S.; Gilmore, J.; Ginsburg, C. M.; Kim, C. L.; Ling, T. Y.; Nylander, P.; Romanowski, T. A.; Blaikley, H. E.; Cashmore, R. J.; Cooper-Sarkar, A. M.; Devenish, R. C. E.; Edmonds, J. K.; Große-Knetter, J.; Harnew, N.; Nath, C.; Noyes, V. A.; Quadt, A.; Ruske, O.; Tickner, J. R.; Uijterwaal, H.; Walczak, R.; Waters, D. S.; Bertolin, A.; Brugnera, R.; Carlin, R.; dal Corso, F.; Dosselli, U.; Limentani, S.; Morandin, M.; Posocco, M.; Stanco, L.; Stroili, R.; Voci, C.; Bulmahn, J.; Oh, B. Y.; Okrasiń Ski, J. R.; Toothacker, W. S.; Whitmore, J. J.; Iga, Y.; D'Agostini, G.; Marini, G.; Nigro, A.; Raso, M.; Hart, J. C.; McCubbin, N. A.; Shah, T. P.; Epperson, D.; Heusch, C.; Rahn, J. T.; Sadrozinski, H. F.-W.; Seiden, A.; Wichmann, R.; Williams, D. C.; Schwarzer, O.; Walenta, A. H.; Abramowicz, H.; Briskin, G.; Dagan, S.; Kananov, S.; Levy, A.; Abe, T.; Fusayasu, T.; Inuzuka, M.; Nagano, K.; Umemori, K.; Yamashita, T.; Hamatsu, R.; Hirose, T.; Homma, K.; Kitamura, S.; Matsushita, T.; Cirio, R.; Costa, M.; Ferrero, M. I.; Maselli, S.; Monaco, V.; Peroni, C.; Petrucci, M. C.; Ruspa, M.; Sacchi, R.; Solano, A.; Staiano, A.; Dardo, M.; Bailey, D. C.; Fagerstroem, C.-P.; Galea, R.; Hartner, G. F.; Joo, K. K.; Levman, G. M.; Martin, J. F.; Orr, R. S.; Polenz, S.; Sabetfakhri, A.; Simmons, D.; Teuscher, R. J.; Butterworth, J. M.; Catterall, C. D.; Jones, T. W.; Lane, J. B.; Saunders, R. L.; Sutton, M. R.; Wing, M.; Ciborowski, J.; Grzelak, G.; Kasprzak, M.; Muchorowski, K.; Nowak, R. J.; Pawlak, J. M.; Pawlak, R.; Tymieniecka, T.; Wróblewski, A. K.; Zakrzewski, J. A.; Adamus, M.; Coldewey, C.; Eisenberg, Y.; Hochman, D.; Karshon, U.; Badgett, W. F.; Chapin, D.; Cross, R.; Dasu, S.; Foudas, C.; Loveless, R. J.; Mattingly, S.; Reeder, D. D.; Smith, W. H.; Vaiciulis, A.; Wodarczyk, M.; Deshpande, A.; Dhawan, S.; Hughes, V. W.; Bhadra, S.; Frisken, W. R.; Khakzad, M.; Schmidke, W. B.

    1998-03-01

    A global event shape analysis of the multihadronic final states observed in neutral current deep inelastic scattering events with a large rapidity gap with respect to the proton direction is presented. The analysis is performed in the range 5<=Q2<=185 GeV2 and 160<=W<=250 GeV, where Q2 is the virtuality of the photon and W is the virtual-photon proton centre of mass energy. Particular emphasis is placed on the dependence of the shape variables, measured in the γ*-pomeron rest frame, on the mass of the hadronic final state, MX. With increasing MX the multihadronic final state becomes more collimated and planar. The experimental results are compared with several models which attempt to describe diffractive events. The broadening effects exhibited by the data require in these models a significant gluon component of the pomeron.

  4. Combining conversation analysis and event sequencing to study health communication.

    PubMed

    Pecanac, Kristen E

    2018-06-01

    Good communication is essential in patient-centered care. The purpose of this paper is to describe conversation analysis and event sequencing and explain how integrating these methods strengthened the analysis in a study of communication between clinicians and surrogate decision makers in an intensive care unit. Conversation analysis was first used to determine how clinicians introduced the need for decision-making regarding life-sustaining treatment and how surrogate decision makers responded. Event sequence analysis then was used to determine the transitional probability (probability of one event leading to another in the interaction) that a given type of clinician introduction would lead to surrogate resistance or alignment. Conversation analysis provides a detailed analysis of the interaction between participants in a conversation. When combined with a quantitative analysis of the patterns of communication in an interaction, these data add information on the communication strategies that produce positive outcomes. Researchers can apply this mixed-methods approach to identify beneficial conversational practices and design interventions to improve health communication. © 2018 Wiley Periodicals, Inc.

  5. Video analysis of motor events in REM sleep behavior disorder.

    PubMed

    Frauscher, Birgit; Gschliesser, Viola; Brandauer, Elisabeth; Ulmer, Hanno; Peralta, Cecilia M; Müller, Jörg; Poewe, Werner; Högl, Birgit

    2007-07-30

    In REM sleep behavior disorder (RBD), several studies focused on electromyographic characterization of motor activity, whereas video analysis has remained more general. The aim of this study was to undertake a detailed and systematic video analysis. Nine polysomnographic records from 5 Parkinson patients with RBD were analyzed and compared with sex- and age-matched controls. Each motor event in the video during REM sleep was classified according to duration, type of movement, and topographical distribution. In RBD, a mean of 54 +/- 23.2 events/10 minutes of REM sleep (total 1392) were identified and visually analyzed. Seventy-five percent of all motor events lasted <2 seconds. Of these events, 1,155 (83.0%) were classified as elementary, 188 (13.5%) as complex behaviors, 50 (3.6%) as violent, and 146 (10.5%) as vocalizations. In the control group, 3.6 +/- 2.3 events/10 minutes (total 264) of predominantly elementary simple character (n = 240, 90.9%) were identified. Number and types of motor events differed significantly between patients and controls (P < 0.05). This study shows a very high number and great variety of motor events during REM sleep in symptomatic RBD. However, most motor events are minor, and violent episodes represent only a small fraction. Copyright 2007 Movement Disorder Society

  6. Using the DOE Knowledge Base for Special Event Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Armstrong, H.M.; Harris, J.M.; Young, C.J.

    1998-10-20

    The DOE Knowledge Base is a library of detailed information whose purpose is to support the United States National Data Center (USNDC) in its mission to monitor compliance with the Comprehensive Test Ban Treaty (CTBT). One of the important tasks which the USNDC must accomplish is to periodically perform detailed analysis of events of high interest, so-called "Special Events", to provide the national authority with information needed to make policy decisions. In this paper we investigate some possible uses of the Knowledge Base for Special Event Analysis (SEA), and make recommendations for improving Knowledge Base support for SEA. To analyzemore » an event in detail, there are two basic types of data which must be used sensor-derived data (wave- forms, arrivals, events, etc.) and regiohalized contextual data (known sources, geological characteristics, etc.). Cur- rently there is no single package which can provide full access to both types of data, so for our study we use a separate package for each MatSeis, the Sandia Labs-developed MATLAB-based seismic analysis package, for wave- form data analysis, and ArcView, an ESRI product, for contextual data analysis. Both packages are well-suited to pro- totyping because they provide a rich set of currently available functionality and yet are also flexible and easily extensible, . Using these tools and Phase I Knowledge Base data sets, we show how the Knowledge Base can improve both the speed and the quality of SEA. Empirically-derived interpolated correction information can be accessed to improve both location estimates and associated error estimates. This information can in turn be used to identi~ any known nearby sources (e.g. mines, volcanos), which may then trigger specialized processing of the sensor data. Based on the location estimate, preferred magnitude formulas and discriminants can be retrieved, and any known blockages can be identified to prevent miscalculations. Relevant historic events can be identilled

  7. ANALYSIS OF INPATIENT HOSPITAL STAFF MENTAL WORKLOAD BY MEANS OF DISCRETE-EVENT SIMULATION

    DTIC Science & Technology

    2016-03-24

    ANALYSIS OF INPATIENT HOSPITAL STAFF MENTAL WORKLOAD BY MEANS OF DISCRETE -EVENT SIMULATION...in the United States. AFIT-ENV-MS-16-M-166 ANALYSIS OF INPATIENT HOSPITAL STAFF MENTAL WORKLOAD BY MEANS OF DISCRETE -EVENT SIMULATION...UNLIMITED. AFIT-ENV-MS-16-M-166 ANALYSIS OF INPATIENT HOSPITAL STAFF MENTAL WORKLOAD BY MEANS OF DISCRETE -EVENT SIMULATION Erich W

  8. Regression analysis of mixed recurrent-event and panel-count data

    PubMed Central

    Zhu, Liang; Tong, Xinwei; Sun, Jianguo; Chen, Manhua; Srivastava, Deo Kumar; Leisenring, Wendy; Robison, Leslie L.

    2014-01-01

    In event history studies concerning recurrent events, two types of data have been extensively discussed. One is recurrent-event data (Cook and Lawless, 2007. The Analysis of Recurrent Event Data. New York: Springer), and the other is panel-count data (Zhao and others, 2010. Nonparametric inference based on panel-count data. Test 20, 1–42). In the former case, all study subjects are monitored continuously; thus, complete information is available for the underlying recurrent-event processes of interest. In the latter case, study subjects are monitored periodically; thus, only incomplete information is available for the processes of interest. In reality, however, a third type of data could occur in which some study subjects are monitored continuously, but others are monitored periodically. When this occurs, we have mixed recurrent-event and panel-count data. This paper discusses regression analysis of such mixed data and presents two estimation procedures for the problem. One is a maximum likelihood estimation procedure, and the other is an estimating equation procedure. The asymptotic properties of both resulting estimators of regression parameters are established. Also, the methods are applied to a set of mixed recurrent-event and panel-count data that arose from a Childhood Cancer Survivor Study and motivated this investigation. PMID:24648408

  9. Event time analysis of longitudinal neuroimage data.

    PubMed

    Sabuncu, Mert R; Bernal-Rusiel, Jorge L; Reuter, Martin; Greve, Douglas N; Fischl, Bruce

    2014-08-15

    This paper presents a method for the statistical analysis of the associations between longitudinal neuroimaging measurements, e.g., of cortical thickness, and the timing of a clinical event of interest, e.g., disease onset. The proposed approach consists of two steps, the first of which employs a linear mixed effects (LME) model to capture temporal variation in serial imaging data. The second step utilizes the extended Cox regression model to examine the relationship between time-dependent imaging measurements and the timing of the event of interest. We demonstrate the proposed method both for the univariate analysis of image-derived biomarkers, e.g., the volume of a structure of interest, and the exploratory mass-univariate analysis of measurements contained in maps, such as cortical thickness and gray matter density. The mass-univariate method employs a recently developed spatial extension of the LME model. We applied our method to analyze structural measurements computed using FreeSurfer, a widely used brain Magnetic Resonance Image (MRI) analysis software package. We provide a quantitative and objective empirical evaluation of the statistical performance of the proposed method on longitudinal data from subjects suffering from Mild Cognitive Impairment (MCI) at baseline. Copyright © 2014 Elsevier Inc. All rights reserved.

  10. Reaching Out: A Break from Traditional Forensic Events. "On Interpretation Analysis."

    ERIC Educational Resources Information Center

    Seney, Ronald J.

    In recent years a new event called "Interpretation Analysis" has appeared at certain forensic events. The objective is for the student, through analysis and performance, to study a piece of literature and to communicate his or her understanding of that literature to a specific audience. Perhaps there is room within the established…

  11. An analysis of the 2016 Hitomi breakup event

    NASA Astrophysics Data System (ADS)

    Flegel, Sven; Bennett, James; Lachut, Michael; Möckel, Marek; Smith, Craig

    2017-04-01

    The breakup of Hitomi (ASTRO-H) on 26 March 2016 is analysed. Debris from the fragmentation is used to estimate the time of the event by propagating backwards and estimating the close approach with the parent object. Based on this method, the breakup event is predicted to have occurred at approximately 01:42 UTC on 26 March 2016. The Gaussian variation of parameters equations based on the instantaneous orbits at the predicted time of the event are solved to gain additional insight into the on-orbit position of Hitomi at the time of the event and to test an alternate approach of determining the event epoch and location. A conjunction analysis is carried out between Hitomi and all catalogued objects which were in orbit around the estimated time of the anomaly. Several debris objects have close approaches with Hitomi; however, there is no evidence to support the breakup was caused by a catalogued object. Debris from both of the largest fragmentation events—the Iridium 33-Cosmos 2251 conjunction in 2009 and the intentional destruction of Fengyun 1C in 2007—is involved in close approaches with Hitomi indicating the persistent threat these events have caused in subsequent space missions. To quantify the magnitude of a potential conjunction, the fragmentation resulting from a collision with the debris is modelled using the EVOLVE-4 breakup model. The debris characteristics are estimated from two-line element data. This analysis is indicative of the threat to space assets that mission planners face due to the growing debris population. The impact of the actual event to the environment is investigated based on the debris associated with Hitomi which is currently contained in the United States Strategic Command's catalogue. A look at the active missions in the orbital vicinity of Hitomi reveals that the Hubble Space Telescope is among the spacecraft which may be immediately affected by the new debris.[Figure not available: see fulltext.

  12. Civil protection and Damaging Hydrogeological Events: comparative analysis of the 2000 and 2015 events in Calabria (southern Italy)

    NASA Astrophysics Data System (ADS)

    Petrucci, Olga; Caloiero, Tommaso; Aurora Pasqua, Angela; Perrotta, Piero; Russo, Luigi; Tansi, Carlo

    2017-11-01

    Calabria (southern Italy) is a flood prone region, due to both its rough orography and fast hydrologic response of most watersheds. During the rainy season, intense rain affects the region, triggering floods and mass movements that cause economic damage and fatalities. This work presents a methodological approach to perform the comparative analysis of two events affecting the same area at a distance of 15 years, by collecting all the qualitative and quantitative features useful to describe both rain and damage. The aim is to understand if similar meteorological events affecting the same area can have different outcomes in terms of damage. The first event occurred between 8 and 10 September 2000, damaged 109 out of 409 municipalities of the region and killed 13 people in a campsite due to a flood. The second event, which occurred between 30 October and 1 November 2015, damaged 79 municipalities, and killed a man due to a flood. The comparative analysis highlights that, despite the exceptionality of triggering daily rain was higher in the 2015 event, the damage caused by the 2000 event to both infrastructures and belongings was higher, and it was strongly increased due to the 13 flood victims. We concluded that, in the 2015 event, the management of pre-event phases, with the issuing of meteorological alert, and the emergency management, with the preventive evacuation of people in hazardous situations due to landslides or floods, contributed to reduce the number of victims.

  13. Regression analysis of mixed recurrent-event and panel-count data.

    PubMed

    Zhu, Liang; Tong, Xinwei; Sun, Jianguo; Chen, Manhua; Srivastava, Deo Kumar; Leisenring, Wendy; Robison, Leslie L

    2014-07-01

    In event history studies concerning recurrent events, two types of data have been extensively discussed. One is recurrent-event data (Cook and Lawless, 2007. The Analysis of Recurrent Event Data. New York: Springer), and the other is panel-count data (Zhao and others, 2010. Nonparametric inference based on panel-count data. Test 20: , 1-42). In the former case, all study subjects are monitored continuously; thus, complete information is available for the underlying recurrent-event processes of interest. In the latter case, study subjects are monitored periodically; thus, only incomplete information is available for the processes of interest. In reality, however, a third type of data could occur in which some study subjects are monitored continuously, but others are monitored periodically. When this occurs, we have mixed recurrent-event and panel-count data. This paper discusses regression analysis of such mixed data and presents two estimation procedures for the problem. One is a maximum likelihood estimation procedure, and the other is an estimating equation procedure. The asymptotic properties of both resulting estimators of regression parameters are established. Also, the methods are applied to a set of mixed recurrent-event and panel-count data that arose from a Childhood Cancer Survivor Study and motivated this investigation. © The Author 2014. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  14. Initiating Event Analysis of a Lithium Fluoride Thorium Reactor

    NASA Astrophysics Data System (ADS)

    Geraci, Nicholas Charles

    The primary purpose of this study is to perform an Initiating Event Analysis for a Lithium Fluoride Thorium Reactor (LFTR) as the first step of a Probabilistic Safety Assessment (PSA). The major objective of the research is to compile a list of key initiating events capable of resulting in failure of safety systems and release of radioactive material from the LFTR. Due to the complex interactions between engineering design, component reliability and human reliability, probabilistic safety assessments are most useful when the scope is limited to a single reactor plant. Thus, this thesis will study the LFTR design proposed by Flibe Energy. An October 2015 Electric Power Research Institute report on the Flibe Energy LFTR asked "what-if?" questions of subject matter experts and compiled a list of key hazards with the most significant consequences to the safety or integrity of the LFTR. The potential exists for unforeseen hazards to pose additional risk for the LFTR, but the scope of this thesis is limited to evaluation of those key hazards already identified by Flibe Energy. These key hazards are the starting point for the Initiating Event Analysis performed in this thesis. Engineering evaluation and technical study of the plant using a literature review and comparison to reference technology revealed four hazards with high potential to cause reactor core damage. To determine the initiating events resulting in realization of these four hazards, reference was made to previous PSAs and existing NRC and EPRI initiating event lists. Finally, fault tree and event tree analyses were conducted, completing the logical classification of initiating events. Results are qualitative as opposed to quantitative due to the early stages of system design descriptions and lack of operating experience or data for the LFTR. In summary, this thesis analyzes initiating events using previous research and inductive and deductive reasoning through traditional risk management techniques to

  15. Root Cause Analysis: Learning from Adverse Safety Events.

    PubMed

    Brook, Olga R; Kruskal, Jonathan B; Eisenberg, Ronald L; Larson, David B

    2015-10-01

    Serious adverse events continue to occur in clinical practice, despite our best preventive efforts. It is essential that radiologists, both as individuals and as a part of organizations, learn from such events and make appropriate changes to decrease the likelihood that such events will recur. Root cause analysis (RCA) is a process to (a) identify factors that underlie variation in performance or that predispose an event toward undesired outcomes and (b) allow for development of effective strategies to decrease the likelihood of similar adverse events occurring in the future. An RCA process should be performed within the environment of a culture of safety, focusing on underlying system contributors and, in a confidential manner, taking into account the emotional effects on the staff involved. The Joint Commission now requires that a credible RCA be performed within 45 days for all sentinel or major adverse events, emphasizing the need for all radiologists to understand the processes with which an effective RCA can be performed. Several RCA-related tools that have been found to be useful in the radiology setting include the "five whys" approach to determine causation; cause-and-effect, or Ishikawa, diagrams; causal tree mapping; affinity diagrams; and Pareto charts. © RSNA, 2015.

  16. Meta-analysis: Association between hypoglycaemia and serious adverse events in older patients.

    PubMed

    Mattishent, Katharina; Loke, Yoon Kong

    2016-07-01

    We aimed to conduct a meta-analysis of serious adverse events (macro- and microvascular events, falls and fractures, death) associated with hypoglycaemia in older patients. We searched MEDLINE and EMBASE spanning a ten-year period up to March 2015 (with automated PubMed updates to October 2015). We selected observational studies reporting on hypoglycaemia and associated serious adverse events, and conducted a meta-analysis. We assessed study validity based on ascertainment of hypoglycaemia, adverse events and adjustment for confounders. We included 17 studies involving 1.86 million participants. Meta-analysis of eight studies demonstrated that hypoglycemic episodes were associated with macrovascular complications, odds ratio (OR) 1.83 (95% confidence interval [CI] 1.64, 2.05), and microvascular complications in two studies OR 1.77 (95% CI 1.49, 2.10). Meta-analysis of four studies demonstrated an association between hypoglycaemia and falls or fractures, OR 1.89 (95% CI 1.54, 2.32) and 1.92 (95% CI 1.56, 2.38) respectively. Hypoglycaemia was associated with increased likelihood of death in a meta-analysis of eight studies, OR 2.04 (95% Confidence Interval 1.68, 2.47). Our meta-analysis raises major concerns about a range of serious adverse events associated with hypoglycaemia. Clinicians should prioritize individualized therapy and closer monitoring strategies to avoid hypoglycaemia in susceptible older patients. Copyright © 2016 Elsevier Inc. All rights reserved.

  17. Performance Analysis: Work Control Events Identified January - August 2010

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    De Grange, C E; Freeman, J W; Kerr, C E

    2011-01-14

    This performance analysis evaluated 24 events that occurred at LLNL from January through August 2010. The analysis identified areas of potential work control process and/or implementation weaknesses and several common underlying causes. Human performance improvement and safety culture factors were part of the causal analysis of each event and were analyzed. The collective significance of all events in 2010, as measured by the occurrence reporting significance category and by the proportion of events that have been reported to the DOE ORPS under the ''management concerns'' reporting criteria, does not appear to have increased in 2010. The frequency of reporting inmore » each of the significance categories has not changed in 2010 compared to the previous four years. There is no change indicating a trend in the significance category and there has been no increase in the proportion of occurrences reported in the higher significance category. Also, the frequency of events, 42 events reported through August 2010, is not greater than in previous years and is below the average of 63 occurrences per year at LLNL since 2006. Over the previous four years, an average of 43% of the LLNL's reported occurrences have been reported as either ''management concerns'' or ''near misses.'' In 2010, 29% of the occurrences have been reported as ''management concerns'' or ''near misses.'' This rate indicates that LLNL is now reporting fewer ''management concern'' and ''near miss'' occurrences compared to the previous four years. From 2008 to the present, LLNL senior management has undertaken a series of initiatives to strengthen the work planning and control system with the primary objective to improve worker safety. In 2008, the LLNL Deputy Director established the Work Control Integrated Project Team to develop the core requirements and graded elements of an institutional work planning and control system. By the end of that year this system was documented and implementation had begun

  18. An analysis of post-event processing in social anxiety disorder.

    PubMed

    Brozovich, Faith; Heimberg, Richard G

    2008-07-01

    Research has demonstrated that self-focused thoughts and negative affect have a reciprocal relationship [Mor, N., Winquist, J. (2002). Self-focused attention and negative affect: A meta-analysis. Psychological Bulletin, 128, 638-662]. In the anxiety disorder literature, post-event processing has emerged as a specific construction of repetitive self-focused thoughts that pertain to social anxiety disorder. Post-event processing can be defined as an individual's repeated consideration and potential reconstruction of his performance following a social situation. Post-event processing can also occur when an individual anticipates a social or performance event and begins to brood about other, past social experiences. The present review examined the post-event processing literature in an attempt to organize and highlight the significant results. The methodologies employed to study post-event processing have included self-report measures, daily diaries, social or performance situations created in the laboratory, and experimental manipulations of post-event processing or anticipation of an upcoming event. Directions for future research on post-event processing are discussed.

  19. Integrating natural language processing expertise with patient safety event review committees to improve the analysis of medication events.

    PubMed

    Fong, Allan; Harriott, Nicole; Walters, Donna M; Foley, Hanan; Morrissey, Richard; Ratwani, Raj R

    2017-08-01

    Many healthcare providers have implemented patient safety event reporting systems to better understand and improve patient safety. Reviewing and analyzing these reports is often time consuming and resource intensive because of both the quantity of reports and length of free-text descriptions in the reports. Natural language processing (NLP) experts collaborated with clinical experts on a patient safety committee to assist in the identification and analysis of medication related patient safety events. Different NLP algorithmic approaches were developed to identify four types of medication related patient safety events and the models were compared. Well performing NLP models were generated to categorize medication related events into pharmacy delivery delays, dispensing errors, Pyxis discrepancies, and prescriber errors with receiver operating characteristic areas under the curve of 0.96, 0.87, 0.96, and 0.81 respectively. We also found that modeling the brief without the resolution text generally improved model performance. These models were integrated into a dashboard visualization to support the patient safety committee review process. We demonstrate the capabilities of various NLP models and the use of two text inclusion strategies at categorizing medication related patient safety events. The NLP models and visualization could be used to improve the efficiency of patient safety event data review and analysis. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. Regression Analysis of Mixed Panel Count Data with Dependent Terminal Events

    PubMed Central

    Yu, Guanglei; Zhu, Liang; Li, Yang; Sun, Jianguo; Robison, Leslie L.

    2017-01-01

    Event history studies are commonly conducted in many fields and a great deal of literature has been established for the analysis of the two types of data commonly arising from these studies: recurrent event data and panel count data. The former arises if all study subjects are followed continuously, while the latter means that each study subject is observed only at discrete time points. In reality, a third type of data, a mixture of the two types of the data above, may occur and furthermore, as with the first two types of the data, there may exist a dependent terminal event, which may preclude the occurrences of recurrent events of interest. This paper discusses regression analysis of mixed recurrent event and panel count data in the presence of a terminal event and an estimating equation-based approach is proposed for estimation of regression parameters of interest. In addition, the asymptotic properties of the proposed estimator are established and a simulation study conducted to assess the finite-sample performance of the proposed method suggests that it works well in practical situations. Finally the methodology is applied to a childhood cancer study that motivated this study. PMID:28098397

  1. Re-Evaluation of Event Correlations in Virtual California Using Statistical Analysis

    NASA Astrophysics Data System (ADS)

    Glasscoe, M. T.; Heflin, M. B.; Granat, R. A.; Yikilmaz, M. B.; Heien, E.; Rundle, J.; Donnellan, A.

    2010-12-01

    Fusing the results of simulation tools with statistical analysis methods has contributed to our better understanding of the earthquake process. In a previous study, we used a statistical method to investigate emergent phenomena in data produced by the Virtual California earthquake simulator. The analysis indicated that there were some interesting fault interactions and possible triggering and quiescence relationships between events. We have converted the original code from Matlab to python/C++ and are now evaluating data from the most recent version of Virtual California in order to analyze and compare any new behavior exhibited by the model. The Virtual California earthquake simulator can be used to study fault and stress interaction scenarios for realistic California earthquakes. The simulation generates a synthetic earthquake catalog of events with a minimum size of ~M 5.8 that can be evaluated using statistical analysis methods. Virtual California utilizes realistic fault geometries and a simple Amontons - Coulomb stick and slip friction law in order to drive the earthquake process by means of a back-slip model where loading of each segment occurs due to the accumulation of a slip deficit at the prescribed slip rate of the segment. Like any complex system, Virtual California may generate emergent phenomena unexpected even by its designers. In order to investigate this, we have developed a statistical method that analyzes the interaction between Virtual California fault elements and thereby determine whether events on any given fault elements show correlated behavior. Our method examines events on one fault element and then determines whether there is an associated event within a specified time window on a second fault element. Note that an event in our analysis is defined as any time an element slips, rather than any particular “earthquake” along the entire fault length. Results are then tabulated and then differenced with an expected correlation

  2. Detection and analysis of microseismic events using a Matched Filtering Algorithm (MFA)

    NASA Astrophysics Data System (ADS)

    Caffagni, Enrico; Eaton, David W.; Jones, Joshua P.; van der Baan, Mirko

    2016-07-01

    A new Matched Filtering Algorithm (MFA) is proposed for detecting and analysing microseismic events recorded by downhole monitoring of hydraulic fracturing. This method requires a set of well-located template (`parent') events, which are obtained using conventional microseismic processing and selected on the basis of high signal-to-noise (S/N) ratio and representative spatial distribution of the recorded microseismicity. Detection and extraction of `child' events are based on stacked, multichannel cross-correlation of the continuous waveform data, using the parent events as reference signals. The location of a child event relative to its parent is determined using an automated process, by rotation of the multicomponent waveforms into the ray-centred co-ordinates of the parent and maximizing the energy of the stacked amplitude envelope within a search volume around the parent's hypocentre. After correction for geometrical spreading and attenuation, the relative magnitude of the child event is obtained automatically using the ratio of stacked envelope peak with respect to its parent. Since only a small number of parent events require interactive analysis such as picking P- and S-wave arrivals, the MFA approach offers the potential for significant reduction in effort for downhole microseismic processing. Our algorithm also facilitates the analysis of single-phase child events, that is, microseismic events for which only one of the S- or P-wave arrivals is evident due to unfavourable S/N conditions. A real-data example using microseismic monitoring data from four stages of an open-hole slickwater hydraulic fracture treatment in western Canada demonstrates that a sparse set of parents (in this case, 4.6 per cent of the originally located events) yields a significant (more than fourfold increase) in the number of located events compared with the original catalogue. Moreover, analysis of the new MFA catalogue suggests that this approach leads to more robust interpretation

  3. Web Video Event Recognition by Semantic Analysis From Ubiquitous Documents.

    PubMed

    Yu, Litao; Yang, Yang; Huang, Zi; Wang, Peng; Song, Jingkuan; Shen, Heng Tao

    2016-12-01

    In recent years, the task of event recognition from videos has attracted increasing interest in multimedia area. While most of the existing research was mainly focused on exploring visual cues to handle relatively small-granular events, it is difficult to directly analyze video content without any prior knowledge. Therefore, synthesizing both the visual and semantic analysis is a natural way for video event understanding. In this paper, we study the problem of Web video event recognition, where Web videos often describe large-granular events and carry limited textual information. Key challenges include how to accurately represent event semantics from incomplete textual information and how to effectively explore the correlation between visual and textual cues for video event understanding. We propose a novel framework to perform complex event recognition from Web videos. In order to compensate the insufficient expressive power of visual cues, we construct an event knowledge base by deeply mining semantic information from ubiquitous Web documents. This event knowledge base is capable of describing each event with comprehensive semantics. By utilizing this base, the textual cues for a video can be significantly enriched. Furthermore, we introduce a two-view adaptive regression model, which explores the intrinsic correlation between the visual and textual cues of the videos to learn reliable classifiers. Extensive experiments on two real-world video data sets show the effectiveness of our proposed framework and prove that the event knowledge base indeed helps improve the performance of Web video event recognition.

  4. External events analysis for the Savannah River Site K reactor

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brandyberry, M.D.; Wingo, H.E.

    1990-01-01

    The probabilistic external events analysis performed for the Savannah River Site K-reactor PRA considered many different events which are generally perceived to be external'' to the reactor and its systems, such as fires, floods, seismic events, and transportation accidents (as well as many others). Events which have been shown to be significant contributors to risk include seismic events, tornados, a crane failure scenario, fires and dam failures. The total contribution to the core melt frequency from external initiators has been found to be 2.2 {times} 10{sup {minus}4} per year, from which seismic events are the major contributor (1.2 {times} 10{supmore » {minus}4} per year). Fire initiated events contribute 1.4 {times} 10{sup {minus}7} per year, tornados 5.8 {times} 10{sup {minus}7} per year, dam failures 1.5 {times} 10{sup {minus}6} per year and the crane failure scenario less than 10{sup {minus}4} per year to the core melt frequency. 8 refs., 3 figs., 5 tabs.« less

  5. Regression analysis of mixed panel count data with dependent terminal events.

    PubMed

    Yu, Guanglei; Zhu, Liang; Li, Yang; Sun, Jianguo; Robison, Leslie L

    2017-05-10

    Event history studies are commonly conducted in many fields, and a great deal of literature has been established for the analysis of the two types of data commonly arising from these studies: recurrent event data and panel count data. The former arises if all study subjects are followed continuously, while the latter means that each study subject is observed only at discrete time points. In reality, a third type of data, a mixture of the two types of the data earlier, may occur and furthermore, as with the first two types of the data, there may exist a dependent terminal event, which may preclude the occurrences of recurrent events of interest. This paper discusses regression analysis of mixed recurrent event and panel count data in the presence of a terminal event and an estimating equation-based approach is proposed for estimation of regression parameters of interest. In addition, the asymptotic properties of the proposed estimator are established, and a simulation study conducted to assess the finite-sample performance of the proposed method suggests that it works well in practical situations. Finally, the methodology is applied to a childhood cancer study that motivated this study. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.

  6. Arenal-type pyroclastic flows: A probabilistic event tree risk analysis

    NASA Astrophysics Data System (ADS)

    Meloy, Anthony F.

    2006-09-01

    A quantitative hazard-specific scenario-modelling risk analysis is performed at Arenal volcano, Costa Rica for the newly recognised Arenal-type pyroclastic flow (ATPF) phenomenon using an event tree framework. These flows are generated by the sudden depressurisation and fragmentation of an active basaltic andesite lava pool as a result of a partial collapse of the crater wall. The deposits of this type of flow include angular blocks and juvenile clasts, which are rarely found in other types of pyroclastic flow. An event tree analysis (ETA) is a useful tool and framework in which to analyse and graphically present the probabilities of the occurrence of many possible events in a complex system. Four event trees are created in the analysis, three of which are extended to investigate the varying individual risk faced by three generic representatives of the surrounding community: a resident, a worker, and a tourist. The raw numerical risk estimates determined by the ETA are converted into a set of linguistic expressions (i.e. VERY HIGH, HIGH, MODERATE etc.) using an established risk classification scale. Three individually tailored semi-quantitative risk maps are then created from a set of risk conversion tables to show how the risk varies for each individual in different areas around the volcano. In some cases, by relocating from the north to the south, the level of risk can be reduced by up to three classes. While the individual risk maps may be broadly applicable, and therefore of interest to the general community, the risk maps and associated probability values generated in the ETA are intended to be used by trained professionals and government agencies to evaluate the risk and effectively manage the long-term development of infrastructure and habitation. With the addition of fresh monitoring data, the combination of both long- and short-term event trees would provide a comprehensive and consistent method of risk analysis (both during and pre-crisis), and as such

  7. Markov chains and semi-Markov models in time-to-event analysis.

    PubMed

    Abner, Erin L; Charnigo, Richard J; Kryscio, Richard J

    2013-10-25

    A variety of statistical methods are available to investigators for analysis of time-to-event data, often referred to as survival analysis. Kaplan-Meier estimation and Cox proportional hazards regression are commonly employed tools but are not appropriate for all studies, particularly in the presence of competing risks and when multiple or recurrent outcomes are of interest. Markov chain models can accommodate censored data, competing risks (informative censoring), multiple outcomes, recurrent outcomes, frailty, and non-constant survival probabilities. Markov chain models, though often overlooked by investigators in time-to-event analysis, have long been used in clinical studies and have widespread application in other fields.

  8. Markov chains and semi-Markov models in time-to-event analysis

    PubMed Central

    Abner, Erin L.; Charnigo, Richard J.; Kryscio, Richard J.

    2014-01-01

    A variety of statistical methods are available to investigators for analysis of time-to-event data, often referred to as survival analysis. Kaplan-Meier estimation and Cox proportional hazards regression are commonly employed tools but are not appropriate for all studies, particularly in the presence of competing risks and when multiple or recurrent outcomes are of interest. Markov chain models can accommodate censored data, competing risks (informative censoring), multiple outcomes, recurrent outcomes, frailty, and non-constant survival probabilities. Markov chain models, though often overlooked by investigators in time-to-event analysis, have long been used in clinical studies and have widespread application in other fields. PMID:24818062

  9. Fault and event tree analyses for process systems risk analysis: uncertainty handling formulations.

    PubMed

    Ferdous, Refaul; Khan, Faisal; Sadiq, Rehan; Amyotte, Paul; Veitch, Brian

    2011-01-01

    Quantitative risk analysis (QRA) is a systematic approach for evaluating likelihood, consequences, and risk of adverse events. QRA based on event (ETA) and fault tree analyses (FTA) employs two basic assumptions. The first assumption is related to likelihood values of input events, and the second assumption is regarding interdependence among the events (for ETA) or basic events (for FTA). Traditionally, FTA and ETA both use crisp probabilities; however, to deal with uncertainties, the probability distributions of input event likelihoods are assumed. These probability distributions are often hard to come by and even if available, they are subject to incompleteness (partial ignorance) and imprecision. Furthermore, both FTA and ETA assume that events (or basic events) are independent. In practice, these two assumptions are often unrealistic. This article focuses on handling uncertainty in a QRA framework of a process system. Fuzzy set theory and evidence theory are used to describe the uncertainties in the input event likelihoods. A method based on a dependency coefficient is used to express interdependencies of events (or basic events) in ETA and FTA. To demonstrate the approach, two case studies are discussed. © 2010 Society for Risk Analysis.

  10. Links between Characteristics of Collaborative Peer Video Analysis Events and Literacy Teachers' Outcomes

    ERIC Educational Resources Information Center

    Arya, Poonam; Christ, Tanya; Chiu, Ming

    2015-01-01

    This study examined how characteristics of Collaborative Peer Video Analysis (CPVA) events are related to teachers' pedagogical outcomes. Data included 39 transcribed literacy video events, in which 14 in-service teachers engaged in discussions of their video clips. Emergent coding and Statistical Discourse Analysis were used to analyze the data.…

  11. Fine-Scale Event Location and Error Analysis in NET-VISA

    NASA Astrophysics Data System (ADS)

    Arora, N. S.; Russell, S.

    2016-12-01

    NET-VISA is a generative probabilistic model for the occurrence of seismic, hydro, and atmospheric events, and the propagation of energy from these events through various mediums and phases before being detected, or misdetected, by IMS stations. It is built on top of the basic station, and arrival detection processing at the IDC, and is currently being tested in the IDC network processing pipelines. A key distinguishing feature of NET-VISA is that it is easy to incorporate prior scientific knowledge and historical data into the probabilistic model. The model accounts for both detections and mis-detections when forming events, and this allows it to make more accurate event hypothesis. It has been continuously evaluated since 2012, and in each year it makes a roughly 60% reduction in the number of missed events without increasing the false event rate as compared to the existing GA algorithm. More importantly the model finds large numbers of events that have been confirmed by regional seismic bulletins but missed by the IDC analysts using the same data. In this work we focus on enhancements to the model to improve the location accuracy, and error ellipses. We will present a new version of the model that focuses on the fine scale around the event location, and present error ellipses and analysis of recent important events.

  12. Analysis and visualization of single-trial event-related potentials

    NASA Technical Reports Server (NTRS)

    Jung, T. P.; Makeig, S.; Westerfield, M.; Townsend, J.; Courchesne, E.; Sejnowski, T. J.

    2001-01-01

    In this study, a linear decomposition technique, independent component analysis (ICA), is applied to single-trial multichannel EEG data from event-related potential (ERP) experiments. Spatial filters derived by ICA blindly separate the input data into a sum of temporally independent and spatially fixed components arising from distinct or overlapping brain or extra-brain sources. Both the data and their decomposition are displayed using a new visualization tool, the "ERP image," that can clearly characterize single-trial variations in the amplitudes and latencies of evoked responses, particularly when sorted by a relevant behavioral or physiological variable. These tools were used to analyze data from a visual selective attention experiment on 28 control subjects plus 22 neurological patients whose EEG records were heavily contaminated with blink and other eye-movement artifacts. Results show that ICA can separate artifactual, stimulus-locked, response-locked, and non-event-related background EEG activities into separate components, a taxonomy not obtained from conventional signal averaging approaches. This method allows: (1) removal of pervasive artifacts of all types from single-trial EEG records, (2) identification and segregation of stimulus- and response-locked EEG components, (3) examination of differences in single-trial responses, and (4) separation of temporally distinct but spatially overlapping EEG oscillatory activities with distinct relationships to task events. The proposed methods also allow the interaction between ERPs and the ongoing EEG to be investigated directly. We studied the between-subject component stability of ICA decomposition of single-trial EEG epochs by clustering components with similar scalp maps and activation power spectra. Components accounting for blinks, eye movements, temporal muscle activity, event-related potentials, and event-modulated alpha activities were largely replicated across subjects. Applying ICA and ERP image

  13. [Analysis on the adverse events of cupping therapy in the application].

    PubMed

    Zhou, Xin; Ruan, Jing-wen; Xing, Bing-feng

    2014-10-01

    The deep analysis has been done on the cases of adverse events and common injury of cupping therapy encountered in recent years in terms of manipulation and patient's constitution. The adverse events of cupping therapy are commonly caused by improper manipulation of medical practitioners, ignoring contraindication and patient's constitution. Clinical practitioners should use cupping therapy cautiously, follow strictly the rules of standard manipulation and medical core system, pay attention to the contraindication and take strict precautions against the occurrence of adverse events.

  14. An unjustified benefit: immortal time bias in the analysis of time-dependent events.

    PubMed

    Gleiss, Andreas; Oberbauer, Rainer; Heinze, Georg

    2018-02-01

    Immortal time bias is a problem arising from methodologically wrong analyses of time-dependent events in survival analyses. We illustrate the problem by analysis of a kidney transplantation study. Following patients from transplantation to death, groups defined by the occurrence or nonoccurrence of graft failure during follow-up seemingly had equal overall mortality. Such naive analysis assumes that patients were assigned to the two groups at time of transplantation, which actually are a consequence of occurrence of a time-dependent event later during follow-up. We introduce landmark analysis as the method of choice to avoid immortal time bias. Landmark analysis splits the follow-up time at a common, prespecified time point, the so-called landmark. Groups are then defined by time-dependent events having occurred before the landmark, and outcome events are only considered if occurring after the landmark. Landmark analysis can be easily implemented with common statistical software. In our kidney transplantation example, landmark analyses with landmarks set at 30 and 60 months clearly identified graft failure as a risk factor for overall mortality. We give further typical examples from transplantation research and discuss strengths and limitations of landmark analysis and other methods to address immortal time bias such as Cox regression with time-dependent covariables. © 2017 Steunstichting ESOT.

  15. Twitter data analysis: temporal and term frequency analysis with real-time event

    NASA Astrophysics Data System (ADS)

    Yadav, Garima; Joshi, Mansi; Sasikala, R.

    2017-11-01

    From the past few years, World Wide Web (www) has become a prominent and huge source for user generated content and opinionative data. Among various social media, Twitter gained popularity as it offers a fast and effective way of sharing users’ perspective towards various critical and other issues in different domain. As the data is hugely generated on cloud, it has opened doors for the researchers in the field of data science and analysis. There are various domains such as ‘Political’ domain, ‘Entertainment’ domain and ‘Business’ domain. Also there are various APIs that Twitter provides for developers 1) Search API, focus on the old tweets 2) Rest API, focuses on user details and allow to collect the user profile, friends and followers 3) Streaming API, which collects details like tweets, hashtags, geo locations. In our work we are accessing Streaming API in order to fetch real-time tweets for the dynamic happening event. For this we are focusing on ‘Entertainment’ domain especially ‘Sports’ as IPL-T20 is currently the trending on-going event. We are collecting these numerous amounts of tweets and storing them in MongoDB database where the tweets are stored in JSON document format. On this document we are performing time-series analysis and term frequency analysis using different techniques such as filtering, information extraction for text-mining that fulfils our objective of finding interesting moments for temporal data in the event and finding the ranking among the players or the teams based on popularity which helps people in understanding key influencers on the social media platform.

  16. Analysis hierarchical model for discrete event systems

    NASA Astrophysics Data System (ADS)

    Ciortea, E. M.

    2015-11-01

    The This paper presents the hierarchical model based on discrete event network for robotic systems. Based on the hierarchical approach, Petri network is analysed as a network of the highest conceptual level and the lowest level of local control. For modelling and control of complex robotic systems using extended Petri nets. Such a system is structured, controlled and analysed in this paper by using Visual Object Net ++ package that is relatively simple and easy to use, and the results are shown as representations easy to interpret. The hierarchical structure of the robotic system is implemented on computers analysed using specialized programs. Implementation of hierarchical model discrete event systems, as a real-time operating system on a computer network connected via a serial bus is possible, where each computer is dedicated to local and Petri model of a subsystem global robotic system. Since Petri models are simplified to apply general computers, analysis, modelling, complex manufacturing systems control can be achieved using Petri nets. Discrete event systems is a pragmatic tool for modelling industrial systems. For system modelling using Petri nets because we have our system where discrete event. To highlight the auxiliary time Petri model using transport stream divided into hierarchical levels and sections are analysed successively. Proposed robotic system simulation using timed Petri, offers the opportunity to view the robotic time. Application of goods or robotic and transmission times obtained by measuring spot is obtained graphics showing the average time for transport activity, using the parameters sets of finished products. individually.

  17. Radar rainfall estimation in the context of post-event analysis of flash-flood events

    NASA Astrophysics Data System (ADS)

    Delrieu, G.; Bouilloud, L.; Boudevillain, B.; Kirstetter, P.-E.; Borga, M.

    2009-09-01

    This communication is about a methodology for radar rainfall estimation in the context of post-event analysis of flash-flood events developed within the HYDRATE project. For such extreme events, some raingauge observations (operational, amateur) are available at the event time scale, while few raingauge time series are generally available at the hydrologic time steps. Radar data is therefore the only way to access to the rainfall space-time organization, but the quality of the radar data may be highly variable as a function of (1) the relative locations of the event and the radar(s) and (2) the radar operating protocol(s) and maintenance. A positive point: heavy rainfall is associated with convection implying better visibility and lesser bright band contamination compared with more current situations. In parallel with the development of a regionalized and adaptive radar data processing system (TRADHy; Delrieu et al. 2009), a pragmatic approach is proposed here to make best use of the available radar and raingauge data for a given flash-flood event by: (1) Identifying and removing residual ground clutter, (2) Applying the "hydrologic visibility" concept (Pellarin et al. 2002) to correct for range-dependent errors (screening and VPR effects for non-attenuating wavelengths, (3) Estimating an effective Z-R relationship through a radar-raingauge optimization approach to remove the mean field bias (Dinku et al. 2002) A sensitivity study, based on the high-quality volume radar datasets collected during two intense rainfall events of the Bollène 2002 experiment (Delrieu et al. 2009), is first proposed. Then the method is implemented for two other historical events occurred in France (Avène 1997 and Aude 1999) with datasets of lesser quality. References: Delrieu, G., B. Boudevillain, J. Nicol, B. Chapon, P.-E. Kirstetter, H. Andrieu, and D. Faure, 2009: Bollène 2002 experiment: radar rainfall estimation in the Cévennes-Vivarais region, France. Journal of Applied

  18. Formal analysis of imprecise system requirements with Event-B.

    PubMed

    Le, Hong Anh; Nakajima, Shin; Truong, Ninh Thuan

    2016-01-01

    Formal analysis of functional properties of system requirements needs precise descriptions. However, the stakeholders sometimes describe the system with ambiguous, vague or fuzzy terms, hence formal frameworks for modeling and verifying such requirements are desirable. The Fuzzy If-Then rules have been used for imprecise requirements representation, but verifying their functional properties still needs new methods. In this paper, we propose a refinement-based modeling approach for specification and verification of such requirements. First, we introduce a representation of imprecise requirements in the set theory. Then we make use of Event-B refinement providing a set of translation rules from Fuzzy If-Then rules to Event-B notations. After that, we show how to verify both safety and eventuality properties with RODIN/Event-B. Finally, we illustrate the proposed method on the example of Crane Controller.

  19. Meta-Analysis of Rare Binary Adverse Event Data

    PubMed Central

    Bhaumik, Dulal K.; Amatya, Anup; Normand, Sharon-Lise; Greenhouse, Joel; Kaizar, Eloise; Neelon, Brian; Gibbons, Robert D.

    2013-01-01

    We examine the use of fixed-effects and random-effects moment-based meta-analytic methods for analysis of binary adverse event data. Special attention is paid to the case of rare adverse events which are commonly encountered in routine practice. We study estimation of model parameters and between-study heterogeneity. In addition, we examine traditional approaches to hypothesis testing of the average treatment effect and detection of the heterogeneity of treatment effect across studies. We derive three new methods, simple (unweighted) average treatment effect estimator, a new heterogeneity estimator, and a parametric bootstrapping test for heterogeneity. We then study the statistical properties of both the traditional and new methods via simulation. We find that in general, moment-based estimators of combined treatment effects and heterogeneity are biased and the degree of bias is proportional to the rarity of the event under study. The new methods eliminate much, but not all of this bias. The various estimators and hypothesis testing methods are then compared and contrasted using an example dataset on treatment of stable coronary artery disease. PMID:23734068

  20. Regression analysis of mixed recurrent-event and panel-count data with additive rate models.

    PubMed

    Zhu, Liang; Zhao, Hui; Sun, Jianguo; Leisenring, Wendy; Robison, Leslie L

    2015-03-01

    Event-history studies of recurrent events are often conducted in fields such as demography, epidemiology, medicine, and social sciences (Cook and Lawless, 2007, The Statistical Analysis of Recurrent Events. New York: Springer-Verlag; Zhao et al., 2011, Test 20, 1-42). For such analysis, two types of data have been extensively investigated: recurrent-event data and panel-count data. However, in practice, one may face a third type of data, mixed recurrent-event and panel-count data or mixed event-history data. Such data occur if some study subjects are monitored or observed continuously and thus provide recurrent-event data, while the others are observed only at discrete times and hence give only panel-count data. A more general situation is that each subject is observed continuously over certain time periods but only at discrete times over other time periods. There exists little literature on the analysis of such mixed data except that published by Zhu et al. (2013, Statistics in Medicine 32, 1954-1963). In this article, we consider the regression analysis of mixed data using the additive rate model and develop some estimating equation-based approaches to estimate the regression parameters of interest. Both finite sample and asymptotic properties of the resulting estimators are established, and the numerical studies suggest that the proposed methodology works well for practical situations. The approach is applied to a Childhood Cancer Survivor Study that motivated this study. © 2014, The International Biometric Society.

  1. Application of a temporal reasoning framework tool in analysis of medical device adverse events.

    PubMed

    Clark, Kimberly K; Sharma, Deepak K; Chute, Christopher G; Tao, Cui

    2011-01-01

    The Clinical Narrative Temporal Relation Ontology (CNTRO)1 project offers a semantic-web based reasoning framework, which represents temporal events and relationships within clinical narrative texts, and infer new knowledge over them. In this paper, the CNTRO reasoning framework is applied to temporal analysis of medical device adverse event files. One specific adverse event was used as a test case: late stent thrombosis. Adverse event narratives were obtained from the Food and Drug Administration's (FDA) Manufacturing and User Facility Device Experience (MAUDE) database2. 15 adverse event files in which late stent thrombosis was confirmed were randomly selected across multiple drug eluting stent devices. From these files, 81 events and 72 temporal relations were annotated. 73 temporal questions were generated, of which 65 were correctly answered by the CNTRO system. This results in an overall accuracy of 89%. This system should be pursued further to continue assessing its potential benefits in temporal analysis of medical device adverse events.

  2. Urbanization and Fertility: An Event-History Analysis of Coastal Ghana

    PubMed Central

    WHITE, MICHAEL J.; MUHIDIN, SALUT; ANDRZEJEWSKI, CATHERINE; TAGOE, EVA; KNIGHT, RODNEY; REED, HOLLY

    2008-01-01

    In this article, we undertake an event-history analysis of fertility in Ghana. We exploit detailed life history calendar data to conduct a more refined and definitive analysis of the relationship among personal traits, urban residence, and fertility. Although urbanization is generally associated with lower fertility in developing countries, inferences in most studies have been hampered by a lack of information about the timing of residence in relationship to childbearing. We find that the effect of urbanization itself is strong, evident, and complex, and persists after we control for the effects of age, cohort, union status, and education. Our discrete-time event-history analysis shows that urban women exhibit fertility rates that are, on average, 11% lower than those of rural women, but the effects vary by parity. Differences in urban population traits would augment the effects of urban adaptation itself. Extensions of the analysis point to the operation of a selection effect in rural-to-urban mobility but provide limited evidence for disruption effects. The possibility of further selection of urbanward migrants on unmeasured traits remains. The analysis also demonstrates the utility of an annual life history calendar for collecting such data in the field. PMID:19110898

  3. Study of the peculiarities of multiparticle production via event-by-event analysis in asymmetric nucleus-nucleus interactions

    NASA Astrophysics Data System (ADS)

    Fedosimova, Anastasiya; Gaitinov, Adigam; Grushevskaya, Ekaterina; Lebedev, Igor

    2017-06-01

    In this work the study on the peculiarities of multiparticle production in interactions of asymmetric nuclei to search for unusual features of such interactions, is performed. A research of long-range and short-range multiparticle correlations in the pseudorapidity distribution of secondary particles on the basis of analysis of individual interactions of nuclei of 197 Au at energy 10.7 AGeV with photoemulsion nuclei, is carried out. Events with long-range multiparticle correlations (LC), short-range multiparticle correlations (SC) and mixed type (MT) in pseudorapidity distribution of secondary particles, are selected by the Hurst method in accordance with Hurst curve behavior. These types have significantly different characteristics. At first, they have different fragmentation parameters. Events of LC type are processes of full destruction of the projectile nucleus, in which multicharge fragments are absent. In events of mixed type several multicharge fragments of projectile nucleus are discovered. Secondly, these two types have significantly different multiplicity distribution. The mean multiplicity of LC type events is significantly more than in mixed type events. On the basis of research of the dependence of multiplicity versus target-nuclei fragments number for events of various types it is revealed, that the most considerable multiparticle correlations are observed in interactions of the mixed type, which correspond to the central collisions of gold nuclei and nuclei of CNO-group, i.e. nuclei with strongly asymmetric volume, nuclear mass, charge, etc. Such events are characterised by full destruction of the target-nucleus and the disintegration of the projectile-nucleus on several multi-charged fragments.

  4. Development and assessment of stressful life events subscales - A preliminary analysis.

    PubMed

    Buccheri, Teresa; Musaad, Salma; Bost, Kelly K; Fiese, Barbara H

    2018-01-15

    Stress affects people of all ages, genders, and cultures and is associated with physical and psychological complications. Stressful life events are an important research focus and a psychometrically valid measure could provide useful clinical information. The purpose of the study was to develop a reliable and valid measurement of stressful life events and to assess its reliability and validity using established measures of social support, stress, depression, anxiety and maternal and child health. The authors used an adaptation from the Social Readjustment Rating Scale (SRRS) to describe the prevalence of life events; they developed a 4-factor stressful life events subscales and used Medical Outcomes Social Support Scale, Social Support Scale, Depression, Anxiety and Stress Scale and 14 general health items for validity analysis. Analyses were performed with descriptive statistics, Cronbach's alpha, Spearman's rho, Chi-square test or Fisher's exact test and Wilcoxon 2-sample test. The 4-factor stressful life events subscales showed acceptable reliability. The resulting subscale scores were significantly associated with established measures of social support, depression, anxiety, stress, and caregiver health indicators. The study presented a number of limitations in terms of design and recall bias. Despite the presence of a number of limitations, the study provided valuable insight and suggested that further investigation is needed in order to determine the effectiveness of the measures in revealing the family's wellbeing and to develop and strengthen a more detailed analysis of the stressful life events/health association. Copyright © 2017 Elsevier B.V. All rights reserved.

  5. Idaho National Laboratory Quarterly Event Performance Analysis FY 2013 4th Quarter

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mitchell, Lisbeth A.

    2013-11-01

    This report is published quarterly by the Idaho National Laboratory (INL) Performance Assurance Organization. The Department of Energy Occurrence Reporting and Processing System (ORPS) as prescribed in DOE Order 232.2 “Occurrence Reporting and Processing of Operations Information” requires a quarterly analysis of events, both reportable and not reportable for the previous twelve months. This report is the analysis of occurrence reports and deficiency reports (including not reportable events) identified at the Idaho National Laboratory (INL) during the period of October 2012 through September 2013.

  6. Normalization Strategies for Enhancing Spatio-Temporal Analysis of Social Media Responses during Extreme Events: A Case Study based on Analysis of Four Extreme Events using Socio-Environmental Data Explorer (SEDE)

    NASA Astrophysics Data System (ADS)

    Ajayakumar, J.; Shook, E.; Turner, V. K.

    2017-10-01

    With social media becoming increasingly location-based, there has been a greater push from researchers across various domains including social science, public health, and disaster management, to tap in the spatial, temporal, and textual data available from these sources to analyze public response during extreme events such as an epidemic outbreak or a natural disaster. Studies based on demographics and other socio-economic factors suggests that social media data could be highly skewed based on the variations of population density with respect to place. To capture the spatio-temporal variations in public response during extreme events we have developed the Socio-Environmental Data Explorer (SEDE). SEDE collects and integrates social media, news and environmental data to support exploration and assessment of public response to extreme events. For this study, using SEDE, we conduct spatio-temporal social media response analysis on four major extreme events in the United States including the "North American storm complex" in December 2015, the "snowstorm Jonas" in January 2016, the "West Virginia floods" in June 2016, and the "Hurricane Matthew" in October 2016. Analysis is conducted on geo-tagged social media data from Twitter and warnings from the storm events database provided by National Centers For Environmental Information (NCEI) for analysis. Results demonstrate that, to support complex social media analyses, spatial and population-based normalization and filtering is necessary. The implications of these results suggests that, while developing software solutions to support analysis of non-conventional data sources such as social media, it is quintessential to identify the inherent biases associated with the data sources, and adapt techniques and enhance capabilities to mitigate the bias. The normalization strategies that we have developed and incorporated to SEDE will be helpful in reducing the population bias associated with social media data and will be useful

  7. Analysis of Loss-of-Offsite-Power Events 1997-2015

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, Nancy Ellen; Schroeder, John Alton

    2016-07-01

    Loss of offsite power (LOOP) can have a major negative impact on a power plant’s ability to achieve and maintain safe shutdown conditions. LOOP event frequencies and times required for subsequent restoration of offsite power are important inputs to plant probabilistic risk assessments. This report presents a statistical and engineering analysis of LOOP frequencies and durations at U.S. commercial nuclear power plants. The data used in this study are based on the operating experience during calendar years 1997 through 2015. LOOP events during critical operation that do not result in a reactor trip, are not included. Frequencies and durations weremore » determined for four event categories: plant-centered, switchyard-centered, grid-related, and weather-related. Emergency diesel generator reliability is also considered (failure to start, failure to load and run, and failure to run more than 1 hour). There is an adverse trend in LOOP durations. The previously reported adverse trend in LOOP frequency was not statistically significant for 2006-2015. Grid-related LOOPs happen predominantly in the summer. Switchyard-centered LOOPs happen predominantly in winter and spring. Plant-centered and weather-related LOOPs do not show statistically significant seasonality. The engineering analysis of LOOP data shows that human errors have been much less frequent since 1997 than in the 1986 -1996 time period.« less

  8. Analysis of event data recorder data for vehicle safety improvement

    DOT National Transportation Integrated Search

    2008-04-01

    The Volpe Center performed a comprehensive engineering analysis of Event Data Recorder (EDR) data supplied by the National Highway Traffic Safety Administration (NHTSA) to assess its accuracy and usefulness in crash reconstruction and improvement of ...

  9. Biometrical issues in the analysis of adverse events within the benefit assessment of drugs.

    PubMed

    Bender, Ralf; Beckmann, Lars; Lange, Stefan

    2016-07-01

    The analysis of adverse events plays an important role in the benefit assessment of drugs. Consequently, results on adverse events are an integral part of reimbursement dossiers submitted by pharmaceutical companies to health policy decision-makers. Methods applied in the analysis of adverse events commonly include simple standard methods for contingency tables. However, the results produced may be misleading if observations are censored at the time of discontinuation due to treatment switching or noncompliance, resulting in unequal follow-up periods. In this paper, we present examples to show that the application of inadequate methods for the analysis of adverse events in the reimbursement dossier can lead to a downgrading of the evidence on a drug's benefit in the subsequent assessment, as greater harm from the drug cannot be excluded with sufficient certainty. Legal regulations on the benefit assessment of drugs in Germany are presented, in particular, with regard to the analysis of adverse events. Differences in safety considerations between the drug approval process and the benefit assessment are discussed. We show that the naive application of simple proportions in reimbursement dossiers frequently leads to uninterpretable results if observations are censored and the average follow-up periods differ between treatment groups. Likewise, the application of incidence rates may be misleading in the case of recurrent events and unequal follow-up periods. To allow for an appropriate benefit assessment of drugs, adequate survival time methods accounting for time dependencies and duration of follow-up are required, not only for time-to-event efficacy endpoints but also for adverse events. © 2016 The Authors. Pharmaceutical Statistics published by John Wiley & Sons Ltd. © 2016 The Authors. Pharmaceutical Statistics published by John Wiley & Sons Ltd.

  10. Classification and Space-Time Analysis of Precipitation Events in Manizales, Caldas, Colombia.

    NASA Astrophysics Data System (ADS)

    Suarez Hincapie, J. N.; Vélez, J.; Romo Melo, L.; Chang, P.

    2015-12-01

    Manizales is a mid-mountain Andean city located near the Nevado del Ruiz volcano in west-central Colombia, this location exposes it to earthquakes, floods, landslides and volcanic eruptions. It is located in the intertropical convergence zone (ITCZ) and presents a climate with a bimodal rainfall regime (Cortés, 2010). Its mean annual rainfall is 2000 mm, one may observe precipitation 70% of the days over a year. This rain which favors the formation of large masses of clouds and the presence of macroclimatic phenomenon as "El Niño South Oscillation", has historically caused great impacts in the region (Vélez et al, 2012). For example the geographical location coupled with rain events results in a high risk of landslides in the city. Manizales has a hydrometeorological network of 40 stations that measure and transmit data of up to eight climate variables. Some of these stations keep 10 years of historical data. However, until now this information has not been used for space-time classification of precipitation events, nor has the meteorological variables that influence them been thoroughly researched. The purpose of this study was to classify historical events of rain in an urban area of Manizales and investigate patterns of atmospheric behavior that influence or trigger such events. Classification of events was performed by calculating the "n" index of the heavy rainfall, describing the behavior of precipitation as a function of time throughout the event (Monjo, 2009). The analysis of meteorological variables was performed using statistical quantification over variable time periods before each event. The proposed classification allowed for an analysis of the evolution of rainfall events. Specially, it helped to look for the influence of different meteorological variables triggering rainfall events in hazardous areas as the city of Manizales.

  11. Random-Effects Meta-Analysis of Time-to-Event Data Using the Expectation-Maximisation Algorithm and Shrinkage Estimators

    ERIC Educational Resources Information Center

    Simmonds, Mark C.; Higgins, Julian P. T.; Stewart, Lesley A.

    2013-01-01

    Meta-analysis of time-to-event data has proved difficult in the past because consistent summary statistics often cannot be extracted from published results. The use of individual patient data allows for the re-analysis of each study in a consistent fashion and thus makes meta-analysis of time-to-event data feasible. Time-to-event data can be…

  12. Service-Learning and Graduation: Evidence from Event History Analysis

    ERIC Educational Resources Information Center

    Yue, Hongtao; Hart, Steven M.

    2017-01-01

    This research employed Event History Analysis to understand how service-learning participation is related to students' graduation within six years. The longitudinal dataset includes 31,074 new undergraduate students who enrolled in a large western U.S. public university from Fall 2002 to Fall 2009. The study revealed that service-learning…

  13. Magnesium and the Risk of Cardiovascular Events: A Meta-Analysis of Prospective Cohort Studies

    PubMed Central

    Hao, Yongqiang; Li, Huiwu; Tang, Tingting; Wang, Hao; Yan, Weili; Dai, Kerong

    2013-01-01

    Background Prospective studies that have examined the association between dietary magnesium intake and serum magnesium concentrations and the risk of cardiovascular disease (CVD) events have reported conflicting findings. We undertook a meta-analysis to evaluate the association between dietary magnesium intake and serum magnesium concentrations and the risk of total CVD events. Methodology/Principal Findings We performed systematic searches on MEDLINE, EMBASE, and OVID up to February 1, 2012 without limits. Categorical, linear, and nonlinear, dose-response, heterogeneity, publication bias, subgroup, and meta-regression analysis were performed. The analysis included 532,979 participants from 19 studies (11 studies on dietary magnesium intake, 6 studies on serum magnesium concentrations, and 2 studies on both) with 19,926 CVD events. The pooled relative risks of total CVD events for the highest vs. lowest category of dietary magnesium intake and serum magnesium concentrations were 0.85 (95% confidence interval 0.78 to 0.92) and 0.77 (0.66 to 0.87), respectively. In linear dose-response analysis, only serum magnesium concentrations ranging from 1.44 to 1.8 mEq/L were significantly associated with total CVD events risk (0.91, 0.85 to 0.97) per 0.1 mEq/L (Pnonlinearity = 0.465). However, significant inverse associations emerged in nonlinear models for dietary magnesium intake (Pnonlinearity = 0.024). The greatest risk reduction occurred when intake increased from 150 to 400 mg/d. There was no evidence of publication bias. Conclusions/Significance There is a statistically significant nonlinear inverse association between dietary magnesium intake and total CVD events risk. Serum magnesium concentrations are linearly and inversely associated with the risk of total CVD events. PMID:23520480

  14. Defining Human Failure Events for Petroleum Risk Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ronald L. Boring; Knut Øien

    2014-06-01

    In this paper, an identification and description of barriers and human failure events (HFEs) for human reliability analysis (HRA) is performed. The barriers, called target systems, are identified from risk significant accident scenarios represented as defined situations of hazard and accident (DSHAs). This report serves as the foundation for further work to develop petroleum HFEs compatible with the SPAR-H method and intended for reuse in future HRAs.

  15. The January 2001, El Salvador event: a multi-data analysis

    NASA Astrophysics Data System (ADS)

    Vallee, M.; Bouchon, M.; Schwartz, S. Y.

    2001-12-01

    On January 13, 2001, a large normal event (Mw=7.6) occured 100 kilometers away from the Salvadorian coast (Central America) with a centroid depth of about 50km. The size of this event is surprising according to the classical idea that such events have to be much weaker than thrust events in subduction zones. We analysed this earthquake with different types of data: because teleseismic waves are the only data which offer a good azimuthal coverage, we first built a kinematic source model with P and SH waves provided by the IRIS-GEOSCOPE networks. The ambiguity between the 30o plane (plunging toward Pacific Ocean) and the 60o degree plane (plunging toward Central America) leaded us to do a parallel analysis of the two possible planes. We used a simple point-source modelling in order to define the main characteristics of the event and then used an extended source to retrieve the kinematic features of the rupture. For the 2 possible planes, this analysis reveals a downdip and northwest rupture propagation but the difference of fit remains subtle even when using the extended source. In a second part we confronted our models for the two planes with other seismological data, which are (1) regional data, (2) surface wave data through an Empirical Green Function given by a similar but much weaker earthquake which occured in July 1996 and lastly (3) nearfield data provided by Universidad Centroamericana (UCA) and Centro de Investigationes Geotecnicas (CIG). Regional data do not allow to discriminate the 2 planes neither but surface waves and especially near field data confirm that the fault plane is the steepest one plunging toward Central America. Moreover, the slight directivity toward North is confirmed by surface waves.

  16. Many multicenter trials had few events per center, requiring analysis via random-effects models or GEEs.

    PubMed

    Kahan, Brennan C; Harhay, Michael O

    2015-12-01

    Adjustment for center in multicenter trials is recommended when there are between-center differences or when randomization has been stratified by center. However, common methods of analysis (such as fixed-effects, Mantel-Haenszel, or stratified Cox models) often require a large number of patients or events per center to perform well. We reviewed 206 multicenter randomized trials published in four general medical journals to assess the average number of patients and events per center and determine whether appropriate methods of analysis were used in trials with few patients or events per center. The median number of events per center/treatment arm combination for trials using a binary or survival outcome was 3 (interquartile range, 1-10). Sixteen percent of trials had less than 1 event per center/treatment combination, 50% fewer than 3, and 63% fewer than 5. Of the trials which adjusted for center using a method of analysis which requires a large number of events per center, 6% had less than 1 event per center-treatment combination, 25% fewer than 3, and 50% fewer than 5. Methods of analysis that allow for few events per center, such as random-effects models or generalized estimating equations (GEEs), were rarely used. Many multicenter trials contain few events per center. Adjustment for center using random-effects models or GEE with model-based (non-robust) standard errors may be beneficial in these scenarios. Copyright © 2015 Elsevier Inc. All rights reserved.

  17. Detection of Abnormal Events via Optical Flow Feature Analysis

    PubMed Central

    Wang, Tian; Snoussi, Hichem

    2015-01-01

    In this paper, a novel algorithm is proposed to detect abnormal events in video streams. The algorithm is based on the histogram of the optical flow orientation descriptor and the classification method. The details of the histogram of the optical flow orientation descriptor are illustrated for describing movement information of the global video frame or foreground frame. By combining one-class support vector machine and kernel principal component analysis methods, the abnormal events in the current frame can be detected after a learning period characterizing normal behaviors. The difference abnormal detection results are analyzed and explained. The proposed detection method is tested on benchmark datasets, then the experimental results show the effectiveness of the algorithm. PMID:25811227

  18. Identifying causes of adverse events detected by an automated trigger tool through in-depth analysis.

    PubMed

    Muething, S E; Conway, P H; Kloppenborg, E; Lesko, A; Schoettker, P J; Seid, M; Kotagal, U

    2010-10-01

    To describe how in-depth analysis of adverse events can reveal underlying causes. Triggers for adverse events were developed using the hospital's computerised medical record (naloxone for opiate-related oversedation and administration of a glucose bolus while on insulin for insulin-related hypoglycaemia). Triggers were identified daily. Based on information from the medical record and interviews, a subject expert determined if an adverse drug event had occurred and then conducted a real-time analysis to identify event characteristics. Expert groups, consisting of frontline staff and specialist physicians, examined event characteristics and determined the apparent cause. 30 insulin-related hypoglycaemia events and 34 opiate-related oversedation events were identified by the triggers over 16 and 21 months, respectively. In the opinion of the experts, patients receiving continuous-infusion insulin and those receiving dextrose only via parenteral nutrition were at increased risk for insulin-related hypoglycaemia. Lack of standardisation in insulin-dosing decisions and variation regarding when and how much to adjust insulin doses in response to changing glucose levels were identified as common causes of the adverse events. Opiate-related oversedation events often occurred within 48 h of surgery. Variation in pain management in the operating room and post-anaesthesia care unit was identified by the experts as potential causes. Variations in practice, multiple services writing orders, multidrug regimens and variations in interpretation of patient assessments were also noted as potential contributing causes. Identification of adverse drug events through an automated trigger system, supplemented by in-depth analysis, can help identify targets for intervention and improvement.

  19. Discrete event simulation tool for analysis of qualitative models of continuous processing systems

    NASA Technical Reports Server (NTRS)

    Malin, Jane T. (Inventor); Basham, Bryan D. (Inventor); Harris, Richard A. (Inventor)

    1990-01-01

    An artificial intelligence design and qualitative modeling tool is disclosed for creating computer models and simulating continuous activities, functions, and/or behavior using developed discrete event techniques. Conveniently, the tool is organized in four modules: library design module, model construction module, simulation module, and experimentation and analysis. The library design module supports the building of library knowledge including component classes and elements pertinent to a particular domain of continuous activities, functions, and behavior being modeled. The continuous behavior is defined discretely with respect to invocation statements, effect statements, and time delays. The functionality of the components is defined in terms of variable cluster instances, independent processes, and modes, further defined in terms of mode transition processes and mode dependent processes. Model construction utilizes the hierarchy of libraries and connects them with appropriate relations. The simulation executes a specialized initialization routine and executes events in a manner that includes selective inherency of characteristics through a time and event schema until the event queue in the simulator is emptied. The experimentation and analysis module supports analysis through the generation of appropriate log files and graphics developments and includes the ability of log file comparisons.

  20. Alternative splicing and trans-splicing events revealed by analysis of the Bombyx mori transcriptome

    PubMed Central

    Shao, Wei; Zhao, Qiong-Yi; Wang, Xiu-Ye; Xu, Xin-Yan; Tang, Qing; Li, Muwang; Li, Xuan; Xu, Yong-Zhen

    2012-01-01

    Alternative splicing and trans-splicing events have not been systematically studied in the silkworm Bombyx mori. Here, the silkworm transcriptome was analyzed by RNA-seq. We identified 320 novel genes, modified 1140 gene models, and found thousands of alternative splicing and 58 trans-splicing events. Studies of three SR proteins show that both their alternative splicing patterns and mRNA products are conserved from insect to human, and one isoform of Srsf6 with a retained intron is expressed sex-specifically in silkworm gonads. Trans-splicing of mod(mdg4) in silkworm was experimentally confirmed. We identified integrations from a common 5′-gene with 46 newly identified alternative 3′-exons that are located on both DNA strands over a 500-kb region. Other trans-splicing events in B. mori were predicted by bioinformatic analysis, in which 12 events were confirmed by RT-PCR, six events were further validated by chimeric SNPs, and two events were confirmed by allele-specific RT-PCR in F1 hybrids from distinct silkworm lines of JS and L10, indicating that trans-splicing is more widespread in insects than previously thought. Analysis of the B. mori transcriptome by RNA-seq provides valuable information of regulatory alternative splicing events. The conservation of splicing events across species and newly identified trans-splicing events suggest that B. mori is a good model for future studies. PMID:22627775

  1. Regression Analysis of Mixed Recurrent-Event and Panel-Count Data with Additive Rate Models

    PubMed Central

    Zhu, Liang; Zhao, Hui; Sun, Jianguo; Leisenring, Wendy; Robison, Leslie L.

    2015-01-01

    Summary Event-history studies of recurrent events are often conducted in fields such as demography, epidemiology, medicine, and social sciences (Cook and Lawless, 2007; Zhao et al., 2011). For such analysis, two types of data have been extensively investigated: recurrent-event data and panel-count data. However, in practice, one may face a third type of data, mixed recurrent-event and panel-count data or mixed event-history data. Such data occur if some study subjects are monitored or observed continuously and thus provide recurrent-event data, while the others are observed only at discrete times and hence give only panel-count data. A more general situation is that each subject is observed continuously over certain time periods but only at discrete times over other time periods. There exists little literature on the analysis of such mixed data except that published by Zhu et al. (2013). In this paper, we consider the regression analysis of mixed data using the additive rate model and develop some estimating equation-based approaches to estimate the regression parameters of interest. Both finite sample and asymptotic properties of the resulting estimators are established, and the numerical studies suggest that the proposed methodology works well for practical situations. The approach is applied to a Childhood Cancer Survivor Study that motivated this study. PMID:25345405

  2. Preterm Versus Term Children: Analysis of Sedation/Anesthesia Adverse Events and Longitudinal Risk.

    PubMed

    Havidich, Jeana E; Beach, Michael; Dierdorf, Stephen F; Onega, Tracy; Suresh, Gautham; Cravero, Joseph P

    2016-03-01

    Preterm and former preterm children frequently require sedation/anesthesia for diagnostic and therapeutic procedures. Our objective was to determine the age at which children who are born <37 weeks gestational age are no longer at increased risk for sedation/anesthesia adverse events. Our secondary objective was to describe the nature and incidence of adverse events. This is a prospective observational study of children receiving sedation/anesthesia for diagnostic and/or therapeutic procedures outside of the operating room by the Pediatric Sedation Research Consortium. A total of 57,227 patients 0 to 22 years of age were eligible for this study. All adverse events and descriptive terms were predefined. Logistic regression and locally weighted scatterplot regression were used for analysis. Preterm and former preterm children had higher adverse event rates (14.7% vs 8.5%) compared with children born at term. Our analysis revealed a biphasic pattern for the development of adverse sedation/anesthesia events. Airway and respiratory adverse events were most commonly reported. MRI scans were the most commonly performed procedures in both categories of patients. Patients born preterm are nearly twice as likely to develop sedation/anesthesia adverse events, and this risk continues up to 23 years of age. We recommend obtaining birth history during the formulation of an anesthetic/sedation plan, with heightened awareness that preterm and former preterm children may be at increased risk. Further prospective studies focusing on the etiology and prevention of adverse events in former preterm patients are warranted. Copyright © 2016 by the American Academy of Pediatrics.

  3. Re-presentations of space in Hollywood movies: an event-indexing analysis.

    PubMed

    Cutting, James; Iricinschi, Catalina

    2015-03-01

    Popular movies present chunk-like events (scenes and subscenes) that promote episodic, serial updating of viewers' representations of the ongoing narrative. Event-indexing theory would suggest that the beginnings of new scenes trigger these updates, which in turn require more cognitive processing. Typically, a new movie event is signaled by an establishing shot, one providing more background information and a longer look than the average shot. Our analysis of 24 films reconfirms this. More important, we show that, when returning to a previously shown location, the re-establishing shot reduces both context and duration while remaining greater than the average shot. In general, location shifts dominate character and time shifts in event segmentation of movies. In addition, over the last 70 years re-establishing shots have become more like the noninitial shots of a scene. Establishing shots have also approached noninitial shot scales, but not their durations. Such results suggest that film form is evolving, perhaps to suit more rapid encoding of narrative events. Copyright © 2014 Cognitive Science Society, Inc.

  4. Data driven analysis of rain events: feature extraction, clustering, microphysical /macro physical relationship

    NASA Astrophysics Data System (ADS)

    Djallel Dilmi, Mohamed; Mallet, Cécile; Barthes, Laurent; Chazottes, Aymeric

    2017-04-01

    The study of rain time series records is mainly carried out using rainfall rate or rain accumulation parameters estimated on a fixed integration time (typically 1 min, 1 hour or 1 day). In this study we used the concept of rain event. In fact, the discrete and intermittent natures of rain processes make the definition of some features inadequate when defined on a fixed duration. Long integration times (hour, day) lead to mix rainy and clear air periods in the same sample. Small integration time (seconds, minutes) will lead to noisy data with a great sensibility to detector characteristics. The analysis on the whole rain event instead of individual short duration samples of a fixed duration allows to clarify relationships between features, in particular between macro physical and microphysical ones. This approach allows suppressing the intra-event variability partly due to measurement uncertainties and allows focusing on physical processes. An algorithm based on Genetic Algorithm (GA) and Self Organising Maps (SOM) is developed to obtain a parsimonious characterisation of rain events using a minimal set of variables. The use of self-organizing map (SOM) is justified by the fact that it allows to map a high dimensional data space in a two-dimensional space while preserving as much as possible the initial space topology in an unsupervised way. The obtained SOM allows providing the dependencies between variables and consequently removing redundant variables leading to a minimal subset of only five features (the event duration, the rain rate peak, the rain event depth, the event rain rate standard deviation and the absolute rain rate variation of order 0.5). To confirm relevance of the five selected features the corresponding SOM is analyzed. This analysis shows clearly the existence of relationships between features. It also shows the independence of the inter-event time (IETp) feature or the weak dependence of the Dry percentage in event (Dd%e) feature. This confirms

  5. Event-scale power law recession analysis: quantifying methodological uncertainty

    NASA Astrophysics Data System (ADS)

    Dralle, David N.; Karst, Nathaniel J.; Charalampous, Kyriakos; Veenstra, Andrew; Thompson, Sally E.

    2017-01-01

    The study of single streamflow recession events is receiving increasing attention following the presentation of novel theoretical explanations for the emergence of power law forms of the recession relationship, and drivers of its variability. Individually characterizing streamflow recessions often involves describing the similarities and differences between model parameters fitted to each recession time series. Significant methodological sensitivity has been identified in the fitting and parameterization of models that describe populations of many recessions, but the dependence of estimated model parameters on methodological choices has not been evaluated for event-by-event forms of analysis. Here, we use daily streamflow data from 16 catchments in northern California and southern Oregon to investigate how combinations of commonly used streamflow recession definitions and fitting techniques impact parameter estimates of a widely used power law recession model. Results are relevant to watersheds that are relatively steep, forested, and rain-dominated. The highly seasonal mediterranean climate of northern California and southern Oregon ensures study catchments explore a wide range of recession behaviors and wetness states, ideal for a sensitivity analysis. In such catchments, we show the following: (i) methodological decisions, including ones that have received little attention in the literature, can impact parameter value estimates and model goodness of fit; (ii) the central tendencies of event-scale recession parameter probability distributions are largely robust to methodological choices, in the sense that differing methods rank catchments similarly according to the medians of these distributions; (iii) recession parameter distributions are method-dependent, but roughly catchment-independent, such that changing the choices made about a particular method affects a given parameter in similar ways across most catchments; and (iv) the observed correlative relationship

  6. Predicting analysis time in events-driven clinical trials using accumulating time-to-event surrogate information.

    PubMed

    Wang, Jianming; Ke, Chunlei; Yu, Zhinuan; Fu, Lei; Dornseif, Bruce

    2016-05-01

    For clinical trials with time-to-event endpoints, predicting the accrual of the events of interest with precision is critical in determining the timing of interim and final analyses. For example, overall survival (OS) is often chosen as the primary efficacy endpoint in oncology studies, with planned interim and final analyses at a pre-specified number of deaths. Often, correlated surrogate information, such as time-to-progression (TTP) and progression-free survival, are also collected as secondary efficacy endpoints. It would be appealing to borrow strength from the surrogate information to improve the precision of the analysis time prediction. Currently available methods in the literature for predicting analysis timings do not consider utilizing the surrogate information. In this article, using OS and TTP as an example, a general parametric model for OS and TTP is proposed, with the assumption that disease progression could change the course of the overall survival. Progression-free survival, related both to OS and TTP, will be handled separately, as it can be derived from OS and TTP. The authors seek to develop a prediction procedure using a Bayesian method and provide detailed implementation strategies under certain assumptions. Simulations are performed to evaluate the performance of the proposed method. An application to a real study is also provided. Copyright © 2015 John Wiley & Sons, Ltd. Copyright © 2015 John Wiley & Sons, Ltd.

  7. Two Point Autocorrelation Analysis of Auger Highest Energy Events Backtracked in Galactic Magnetic Field

    NASA Astrophysics Data System (ADS)

    Petrov, Yevgeniy

    2009-10-01

    Searches for sources of the highest-energy cosmic rays traditionally have included looking for clusters of event arrival directions on the sky. The smallest cluster is a pair of events falling within some angular window. In contrast to the standard two point (2-pt) autocorrelation analysis, this work takes into account influence of the galactic magnetic field (GMF). The highest energy events, those above 50EeV, collected by the surface detector of the Pierre Auger Observatory between January 1, 2004 and May 31, 2009 are used in the analysis. Having assumed protons as primaries, events are backtracked through BSS/S, BSS/A, ASS/S and ASS/A versions of Harari-Mollerach-Roulet (HMR) model of the GMF. For each version of the model, a 2-pt autocorrelation analysis is applied to the backtracked events and to 105 isotropic Monte Carlo realizations weighted by the Auger exposure. Scans in energy, separation angular window and different model parameters reveal clustering at different angular scales. Small angle clustering at 2-3 deg is particularly interesting and it is compared between different field scenarios. The strength of the autocorrelation signal at those angular scales differs between BSS and ASS versions of the HMR model. The BSS versions of the model tend to defocus protons as they arrive to Earth whereas for the ASS, in contrary, it is more likely to focus them.

  8. Advanced Reactor Passive System Reliability Demonstration Analysis for an External Event

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bucknor, Matthew D.; Grabaskas, David; Brunett, Acacia J.

    2016-01-01

    Many advanced reactor designs rely on passive systems to fulfill safety functions during accident sequences. These systems depend heavily on boundary conditions to induce a motive force, meaning the system can fail to operate as intended due to deviations in boundary conditions, rather than as the result of physical failures. Furthermore, passive systems may operate in intermediate or degraded modes. These factors make passive system operation difficult to characterize within a traditional probabilistic framework that only recognizes discrete operating modes and does not allow for the explicit consideration of time-dependent boundary conditions. Argonne National Laboratory has been examining various methodologiesmore » for assessing passive system reliability within a probabilistic risk assessment for a station blackout event at an advanced small modular reactor. This paper provides an overview of a passive system reliability demonstration analysis for an external event. Centering on an earthquake with the possibility of site flooding, the analysis focuses on the behavior of the passive reactor cavity cooling system following potential physical damage and system flooding. The assessment approach seeks to combine mechanistic and simulation-based methods to leverage the benefits of the simulation-based approach without the need to substantially deviate from conventional probabilistic risk assessment techniques. While this study is presented as only an example analysis, the results appear to demonstrate a high level of reliability for the reactor cavity cooling system (and the reactor system in general) to the postulated transient event.« less

  9. Advanced Reactor Passive System Reliability Demonstration Analysis for an External Event

    DOE PAGES

    Bucknor, Matthew; Grabaskas, David; Brunett, Acacia J.; ...

    2017-01-24

    We report that many advanced reactor designs rely on passive systems to fulfill safety functions during accident sequences. These systems depend heavily on boundary conditions to induce a motive force, meaning the system can fail to operate as intended because of deviations in boundary conditions, rather than as the result of physical failures. Furthermore, passive systems may operate in intermediate or degraded modes. These factors make passive system operation difficult to characterize within a traditional probabilistic framework that only recognizes discrete operating modes and does not allow for the explicit consideration of time-dependent boundary conditions. Argonne National Laboratory has beenmore » examining various methodologies for assessing passive system reliability within a probabilistic risk assessment for a station blackout event at an advanced small modular reactor. This paper provides an overview of a passive system reliability demonstration analysis for an external event. Considering an earthquake with the possibility of site flooding, the analysis focuses on the behavior of the passive Reactor Cavity Cooling System following potential physical damage and system flooding. The assessment approach seeks to combine mechanistic and simulation-based methods to leverage the benefits of the simulation-based approach without the need to substantially deviate from conventional probabilistic risk assessment techniques. Lastly, although this study is presented as only an example analysis, the results appear to demonstrate a high level of reliability of the Reactor Cavity Cooling System (and the reactor system in general) for the postulated transient event.« less

  10. Fault Tree Analysis: An Operations Research Tool for Identifying and Reducing Undesired Events in Training.

    ERIC Educational Resources Information Center

    Barker, Bruce O.; Petersen, Paul D.

    This paper explores the fault-tree analysis approach to isolating failure modes within a system. Fault tree investigates potentially undesirable events and then looks for failures in sequence that would lead to their occurring. Relationships among these events are symbolized by AND or OR logic gates, AND used when single events must coexist to…

  11. Element analysis: a wavelet-based method for analysing time-localized events in noisy time series.

    PubMed

    Lilly, Jonathan M

    2017-04-01

    A method is derived for the quantitative analysis of signals that are composed of superpositions of isolated, time-localized 'events'. Here, these events are taken to be well represented as rescaled and phase-rotated versions of generalized Morse wavelets, a broad family of continuous analytic functions. Analysing a signal composed of replicates of such a function using another Morse wavelet allows one to directly estimate the properties of events from the values of the wavelet transform at its own maxima. The distribution of events in general power-law noise is determined in order to establish significance based on an expected false detection rate. Finally, an expression for an event's 'region of influence' within the wavelet transform permits the formation of a criterion for rejecting spurious maxima due to numerical artefacts or other unsuitable events. Signals can then be reconstructed based on a small number of isolated points on the time/scale plane. This method, termed element analysis , is applied to the identification of long-lived eddy structures in ocean currents as observed by along-track measurements of sea surface elevation from satellite altimetry.

  12. Requirements analysis for a hardware, discrete-event, simulation engine accelerator

    NASA Astrophysics Data System (ADS)

    Taylor, Paul J., Jr.

    1991-12-01

    An analysis of a general Discrete Event Simulation (DES), executing on the distributed architecture of an eight mode Intel PSC/2 hypercube, was performed. The most time consuming portions of the general DES algorithm were determined to be the functions associated with message passing of required simulation data between processing nodes of the hypercube architecture. A behavioral description, using the IEEE standard VHSIC Hardware Description and Design Language (VHDL), for a general DES hardware accelerator is presented. The behavioral description specifies the operational requirements for a DES coprocessor to augment the hypercube's execution of DES simulations. The DES coprocessor design implements the functions necessary to perform distributed discrete event simulations using a conservative time synchronization protocol.

  13. Multi-spacecraft solar energetic particle analysis of FERMI gamma-ray flare events within the HESPERIA H2020 project

    NASA Astrophysics Data System (ADS)

    Tziotziou, Kostas; Malandraki, Olga; Valtonen, Eino; Heber, Bernd; Zucca, Pietro; Klein, Karl-Ludwig; Vainio, Rami; Tsiropoula, Georgia; Share, Gerald

    2017-04-01

    Multi-spacecraft observations of solar energetic particle (SEP) events are important for understanding the acceleration processes and the interplanetary propagation of particles released during eruptive events. In this work, we have carefully studied 25 gamma-ray flare events observed by FERMI and investigated possible associations with SEP-related events observed with STEREO and L1 spacecraft in the heliosphere. A data-driven velocity dispersion analysis (VDA) and Time-Shifting Analysis (TSA) are used for deriving the release times of protons and electrons at the Sun and for comparing them with the respective times stemming from the gamma-ray event analysis and their X-ray signatures, in an attempt to interconnect the SEPs and Fermi events and better understand the physics involved. Acknowledgements: This project has received funding from the European Union's Horizon 2020 research and innovation program under grant agreement No 637324.

  14. Modeling time-to-event (survival) data using classification tree analysis.

    PubMed

    Linden, Ariel; Yarnold, Paul R

    2017-12-01

    Time to the occurrence of an event is often studied in health research. Survival analysis differs from other designs in that follow-up times for individuals who do not experience the event by the end of the study (called censored) are accounted for in the analysis. Cox regression is the standard method for analysing censored data, but the assumptions required of these models are easily violated. In this paper, we introduce classification tree analysis (CTA) as a flexible alternative for modelling censored data. Classification tree analysis is a "decision-tree"-like classification model that provides parsimonious, transparent (ie, easy to visually display and interpret) decision rules that maximize predictive accuracy, derives exact P values via permutation tests, and evaluates model cross-generalizability. Using empirical data, we identify all statistically valid, reproducible, longitudinally consistent, and cross-generalizable CTA survival models and then compare their predictive accuracy to estimates derived via Cox regression and an unadjusted naïve model. Model performance is assessed using integrated Brier scores and a comparison between estimated survival curves. The Cox regression model best predicts average incidence of the outcome over time, whereas CTA survival models best predict either relatively high, or low, incidence of the outcome over time. Classification tree analysis survival models offer many advantages over Cox regression, such as explicit maximization of predictive accuracy, parsimony, statistical robustness, and transparency. Therefore, researchers interested in accurate prognoses and clear decision rules should consider developing models using the CTA-survival framework. © 2017 John Wiley & Sons, Ltd.

  15. Neural network approach in multichannel auditory event-related potential analysis.

    PubMed

    Wu, F Y; Slater, J D; Ramsay, R E

    1994-04-01

    Even though there are presently no clearly defined criteria for the assessment of P300 event-related potential (ERP) abnormality, it is strongly indicated through statistical analysis that such criteria exist for classifying control subjects and patients with diseases resulting in neuropsychological impairment such as multiple sclerosis (MS). We have demonstrated the feasibility of artificial neural network (ANN) methods in classifying ERP waveforms measured at a single channel (Cz) from control subjects and MS patients. In this paper, we report the results of multichannel ERP analysis and a modified network analysis methodology to enhance automation of the classification rule extraction process. The proposed methodology significantly reduces the work of statistical analysis. It also helps to standardize the criteria of P300 ERP assessment and facilitate the computer-aided analysis on neuropsychological functions.

  16. Root cause analysis of critical events in neurosurgery, New South Wales.

    PubMed

    Perotti, Vanessa; Sheridan, Mark M P

    2015-09-01

    Adverse events reportedly occur in 5% to 10% of health care episodes. Not all adverse events are the result of error; they may arise from systemic faults in the delivery of health care. Catastrophic events are not only physically devastating to patients, but they also attract medical liability and increase health care costs. Root cause analysis (RCA) has become a key tool for health care services to understand those adverse events. This study is a review of all the RCA case reports involving neurosurgical patients in New South Wales between 2008 and 2013. The case reports and data were obtained from the Clinical Excellence Commission database. The data was then categorized by the root causes identified and the recommendations suggested by the RCA committees. Thirty-two case reports were identified in the RCA database. Breaches in policy account for the majority of root causes identified, for example, delays in transfer of patients or wrong-site surgery, which always involved poor adherence to correct patient and site identification procedures. The RCA committees' recommendations included education for staff, and improvements in rostering and procedural guidelines. RCAs have improved the patient safety profile; however, the RCA committees have no power to enforce any recommendation or ensure compliance. A single RCA may provide little learning beyond the unit and staff involved. However, through aggregation of RCA data and dissemination strategies, health care workers can learn from adverse events and prevent future events from occurring. © 2015 Royal Australasian College of Surgeons.

  17. Impact of including or excluding both-armed zero-event studies on using standard meta-analysis methods for rare event outcome: a simulation study

    PubMed Central

    Cheng, Ji; Pullenayegum, Eleanor; Marshall, John K; Thabane, Lehana

    2016-01-01

    Objectives There is no consensus on whether studies with no observed events in the treatment and control arms, the so-called both-armed zero-event studies, should be included in a meta-analysis of randomised controlled trials (RCTs). Current analytic approaches handled them differently depending on the choice of effect measures and authors' discretion. Our objective is to evaluate the impact of including or excluding both-armed zero-event (BA0E) studies in meta-analysis of RCTs with rare outcome events through a simulation study. Method We simulated 2500 data sets for different scenarios varying the parameters of baseline event rate, treatment effect and number of patients in each trial, and between-study variance. We evaluated the performance of commonly used pooling methods in classical meta-analysis—namely, Peto, Mantel-Haenszel with fixed-effects and random-effects models, and inverse variance method with fixed-effects and random-effects models—using bias, root mean square error, length of 95% CI and coverage. Results The overall performance of the approaches of including or excluding BA0E studies in meta-analysis varied according to the magnitude of true treatment effect. Including BA0E studies introduced very little bias, decreased mean square error, narrowed the 95% CI and increased the coverage when no true treatment effect existed. However, when a true treatment effect existed, the estimates from the approach of excluding BA0E studies led to smaller bias than including them. Among all evaluated methods, the Peto method excluding BA0E studies gave the least biased results when a true treatment effect existed. Conclusions We recommend including BA0E studies when treatment effects are unlikely, but excluding them when there is a decisive treatment effect. Providing results of including and excluding BA0E studies to assess the robustness of the pooled estimated effect is a sensible way to communicate the results of a meta-analysis when the treatment

  18. Survey of critical failure events in on-chip interconnect by fault tree analysis

    NASA Astrophysics Data System (ADS)

    Yokogawa, Shinji; Kunii, Kyousuke

    2018-07-01

    In this paper, a framework based on reliability physics is proposed for adopting fault tree analysis (FTA) to the on-chip interconnect system of a semiconductor. By integrating expert knowledge and experience regarding the possibilities of failure on basic events, critical issues of on-chip interconnect reliability will be evaluated by FTA. In particular, FTA is used to identify the minimal cut sets with high risk priority. Critical events affecting the on-chip interconnect reliability are identified and discussed from the viewpoint of long-term reliability assessment. The moisture impact is evaluated as an external event.

  19. Systematic identification and analysis of frequent gene fusion events in metabolic pathways

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Henry, Christopher S.; Lerma-Ortiz, Claudia; Gerdes, Svetlana Y.

    Here, gene fusions are the most powerful type of in silico-derived functional associations. However, many fusion compilations were made when <100 genomes were available, and algorithms for identifying fusions need updating to handle the current avalanche of sequenced genomes. The availability of a large fusion dataset would help probe functional associations and enable systematic analysis of where and why fusion events occur. As a result, here we present a systematic analysis of fusions in prokaryotes. We manually generated two training sets: (i) 121 fusions in the model organism Escherichia coli; (ii) 131 fusions found in B vitamin metabolism. These setsmore » were used to develop a fusion prediction algorithm that captured the training set fusions with only 7 % false negatives and 50 % false positives, a substantial improvement over existing approaches. This algorithm was then applied to identify 3.8 million potential fusions across 11,473 genomes. The results of the analysis are available in a searchable database. A functional analysis identified 3,000 reactions associated with frequent fusion events and revealed areas of metabolism where fusions are particularly prevalent. In conclusion, customary definitions of fusions were shown to be ambiguous, and a stricter one was proposed. Exploring the genes participating in fusion events showed that they most commonly encode transporters, regulators, and metabolic enzymes. The major rationales for fusions between metabolic genes appear to be overcoming pathway bottlenecks, avoiding toxicity, controlling competing pathways, and facilitating expression and assembly of protein complexes. Finally, our fusion dataset provides powerful clues to decipher the biological activities of domains of unknown function.« less

  20. Systematic identification and analysis of frequent gene fusion events in metabolic pathways

    DOE PAGES

    Henry, Christopher S.; Lerma-Ortiz, Claudia; Gerdes, Svetlana Y.; ...

    2016-06-24

    Here, gene fusions are the most powerful type of in silico-derived functional associations. However, many fusion compilations were made when <100 genomes were available, and algorithms for identifying fusions need updating to handle the current avalanche of sequenced genomes. The availability of a large fusion dataset would help probe functional associations and enable systematic analysis of where and why fusion events occur. As a result, here we present a systematic analysis of fusions in prokaryotes. We manually generated two training sets: (i) 121 fusions in the model organism Escherichia coli; (ii) 131 fusions found in B vitamin metabolism. These setsmore » were used to develop a fusion prediction algorithm that captured the training set fusions with only 7 % false negatives and 50 % false positives, a substantial improvement over existing approaches. This algorithm was then applied to identify 3.8 million potential fusions across 11,473 genomes. The results of the analysis are available in a searchable database. A functional analysis identified 3,000 reactions associated with frequent fusion events and revealed areas of metabolism where fusions are particularly prevalent. In conclusion, customary definitions of fusions were shown to be ambiguous, and a stricter one was proposed. Exploring the genes participating in fusion events showed that they most commonly encode transporters, regulators, and metabolic enzymes. The major rationales for fusions between metabolic genes appear to be overcoming pathway bottlenecks, avoiding toxicity, controlling competing pathways, and facilitating expression and assembly of protein complexes. Finally, our fusion dataset provides powerful clues to decipher the biological activities of domains of unknown function.« less

  1. Diagnostic evaluation of distributed physically based model at the REW scale (THREW) using rainfall-runoff event analysis

    NASA Astrophysics Data System (ADS)

    Tian, F.; Sivapalan, M.; Li, H.; Hu, H.

    2007-12-01

    The importance of diagnostic analysis of hydrological models is increasingly recognized by the scientific community (M. Sivapalan, et al., 2003; H. V. Gupta, et al., 2007). Model diagnosis refers to model structures and parameters being identified not only by statistical comparison of system state variables and outputs but also by process understanding in a specific watershed. Process understanding can be gained by the analysis of observational data and model results at the specific watershed as well as through regionalization. Although remote sensing technology can provide valuable data about the inputs, state variables, and outputs of the hydrological system, observational rainfall-runoff data still constitute the most accurate, reliable, direct, and thus a basic component of hydrology related database. One critical question in model diagnostic analysis is, therefore, what signature characteristic can we extract from rainfall and runoff data. To this date only a few studies have focused on this question, such as Merz et al. (2006) and Lana-Renault et al. (2007), still none of these studies related event analysis with model diagnosis in an explicit, rigorous, and systematic manner. Our work focuses on the identification of the dominant runoff generation mechanisms from event analysis of rainfall-runoff data, including correlation analysis and analysis of timing pattern. The correlation analysis involves the identification of the complex relationship among rainfall depth, intensity, runoff coefficient, and antecedent conditions, and the timing pattern analysis aims to identify the clustering pattern of runoff events in relation to the patterns of rainfall events. Our diagnostic analysis illustrates the changing pattern of runoff generation mechanisms in the DMIP2 test watersheds located in Oklahoma region, which is also well recognized by numerical simulations based on TsingHua Representative Elementary Watershed (THREW) model. The result suggests the usefulness of

  2. Assessing and quantifying changes in precipitation patterns using event-driven analysis

    USDA-ARS?s Scientific Manuscript database

    Studies have claimed that climate change may adversely affect precipitation patterns by increasing the occurrence of extreme events. The effects of climate change on precipitation is expected to take place over a long period of time and will require long-term data to demonstrate. Frequency analysis ...

  3. A Content-Adaptive Analysis and Representation Framework for Audio Event Discovery from "Unscripted" Multimedia

    NASA Astrophysics Data System (ADS)

    Radhakrishnan, Regunathan; Divakaran, Ajay; Xiong, Ziyou; Otsuka, Isao

    2006-12-01

    We propose a content-adaptive analysis and representation framework to discover events using audio features from "unscripted" multimedia such as sports and surveillance for summarization. The proposed analysis framework performs an inlier/outlier-based temporal segmentation of the content. It is motivated by the observation that "interesting" events in unscripted multimedia occur sparsely in a background of usual or "uninteresting" events. We treat the sequence of low/mid-level features extracted from the audio as a time series and identify subsequences that are outliers. The outlier detection is based on eigenvector analysis of the affinity matrix constructed from statistical models estimated from the subsequences of the time series. We define the confidence measure on each of the detected outliers as the probability that it is an outlier. Then, we establish a relationship between the parameters of the proposed framework and the confidence measure. Furthermore, we use the confidence measure to rank the detected outliers in terms of their departures from the background process. Our experimental results with sequences of low- and mid-level audio features extracted from sports video show that "highlight" events can be extracted effectively as outliers from a background process using the proposed framework. We proceed to show the effectiveness of the proposed framework in bringing out suspicious events from surveillance videos without any a priori knowledge. We show that such temporal segmentation into background and outliers, along with the ranking based on the departure from the background, can be used to generate content summaries of any desired length. Finally, we also show that the proposed framework can be used to systematically select "key audio classes" that are indicative of events of interest in the chosen domain.

  4. Sources of Infrasound events listed in IDC Reviewed Event Bulletin

    NASA Astrophysics Data System (ADS)

    Bittner, Paulina; Polich, Paul; Gore, Jane; Ali, Sherif; Medinskaya, Tatiana; Mialle, Pierrick

    2017-04-01

    Until 2003 two waveform technologies, i.e. seismic and hydroacoustic were used to detect and locate events included in the International Data Centre (IDC) Reviewed Event Bulletin (REB). The first atmospheric event was published in the REB in 2003, however automatic processing required significant improvements to reduce the number of false events. In the beginning of 2010 the infrasound technology was reintroduced to the IDC operations and has contributed to both automatic and reviewed IDC bulletins. The primary contribution of infrasound technology is to detect atmospheric events. These events may also be observed at seismic stations, which will significantly improve event location. Examples sources of REB events, which were detected by the International Monitoring System (IMS) infrasound network were fireballs (e.g. Bangkok fireball, 2015), volcanic eruptions (e.g. Calbuco, Chile 2015) and large surface explosions (e.g. Tjanjin, China 2015). Query blasts (e.g. Zheleznogorsk) and large earthquakes (e.g. Italy 2016) belong to events primarily recorded at seismic stations of the IMS network but often detected at the infrasound stations. In case of earthquakes analysis of infrasound signals may help to estimate the area affected by ground vibration. Infrasound associations to query blast events may help to obtain better source location. The role of IDC analysts is to verify and improve location of events detected by the automatic system and to add events which were missed in the automatic process. Open source materials may help to identify nature of some events. Well recorded examples may be added to the Reference Infrasound Event Database to help in analysis process. This presentation will provide examples of events generated by different sources which were included in the IDC bulletins.

  5. Life stress events and alcohol misuse: distinguishing contributing stress events from consequential stress events.

    PubMed

    Hart, Kenneth E; Fazaa, Norman

    2004-07-01

    This study examined the relationship between life stress events and level of alcohol misuse using two stress indices. The first index consisted of stress events that are not likely to be caused by alcohol misuse (i.e., alcohol uncontaminated stress events). The second stress index consisted of items that were judged as being likely consequences of alcohol misuse (i.e., alcohol contaminated stress events). Results based on a questionnaire study of 378 undergraduates in 2000 showed that level of alcohol misuse was much more strongly related to alcohol contaminated life stress events than alcohol uncontaminated life events. Comparative analysis of the coefficients of determination indicated the effect size of the association to alcohol contaminated life stress events was 240% larger than the corresponding effect size for the association to alcohol uncontaminated life events. Results suggest that studies, which are tests of the tension reduction hypothesis, should employ greater methodological rigor to ensure measures of life stress events are not inadvertently assessing the consequences of alcohol misuse. The results highlight the need to distinguish between stressful life events that contribute to alcohol misuse and stressful life events that are consequential to alcohol misuse.

  6. Traumatic events and depressive symptoms among youth in Southwest Nigeria: a qualitative analysis.

    PubMed

    Omigbodun, Olayinka; Bakare, Kofoworola; Yusuf, Bidemi

    2008-01-01

    Traumatic experiences have dire consequences for the mental health of young persons. Despite high rates of traumatic experiences in some African cities, there are no reports for Nigerian youth. To investigate the pattern of traumatic events and their association with depressive symptoms among youth in Southwest Nigeria. This is a descriptive cross-sectional study of randomly selected youth in urban and rural schools in Southwest Nigeria. They completed self-reports on traumatic events and depressive symptoms using the Street Children's Project Questionnaire and the Youth DISC Predictive Scale (DPS). Of the 1,768 responses (88.4% response rate) entered into the analysis, 34% reported experiencing a traumatic situation. Following interpretative phenomenological analysis, 13 themes emerged. Frequently occurring traumatic events were 'road traffic accidents' (33.0%), 'sickness' (17.1%), 'lost or trapped' (11.2%) and 'armed robbery attack' (9.7%). A bad dream was described by 3.7%. Traumatic experiences were commoner in males (36.2%) than in females (31.6%) (x2 = 4.2; p = .041). Experiencing a traumatic event was associated with depressive symptoms (X2 = 37.98; p < .001), especially when the event directly affected the youth as in sexual assault or physical abuse. One-third of youth in Southwest Nigeria have described an experienced traumatic event. Road traffic accidents, armed robbery attacks, and communal disturbances depict the prevailing social environment, whereas 'bad dreams' revealed the influence of cultural beliefs. Policy makers must be aware of the social issues making an impact on the health of youth. Multi-agency interventions to improve the social environment and provide mental health services for traumatized young people are essential.

  7. Efficacy and adverse events of cold vs hot polypectomy: A meta-analysis.

    PubMed

    Fujiya, Mikihiro; Sato, Hiroki; Ueno, Nobuhiro; Sakatani, Aki; Tanaka, Kazuyuki; Dokoshi, Tatsuya; Fujibayashi, Shugo; Nomura, Yoshiki; Kashima, Shin; Gotoh, Takuma; Sasajima, Junpei; Moriichi, Kentaro; Watari, Jiro; Kohgo, Yutaka

    2016-06-21

    To compare previously reported randomized controlled studies (RCTs) of cold and hot polypectomy, we systematically reviewed and clarify the utility of cold polypectomy over hot with respect to efficacy and adverse events. A meta-analysis was conducted to evaluate the predominance of cold and hot polypectomy for removing colon polyps. Published articles and abstracts from worldwide conferences were searched using the keywords "cold polypectomy". RCTs that compared either or both the effects or adverse events of cold polypectomy with those of hot polypectomy were collected. The patients' demographics, endoscopic procedures, No. of examined lesions, lesion size, macroscopic and histologic findings, rates of incomplete resection, bleeding amount, perforation, and length of procedure were extracted from each study. A forest plot analysis was used to verify the relative strength of the effects and adverse events of each procedure. A funnel plot was generated to assess the possibility of publication bias. Ultimately, six RCTs were selected. No significant differences were noted in the average lesion size (less than 10 mm) between the cold and hot polypectomy groups in each study. Further, the rates of complete resection and adverse events, including delayed bleeding, did not differ markedly between cold and hot polypectomy. The average procedural time in the cold polypectomy group was significantly shorter than in the hot polypectomy group. Cold polypectomy is a time-saving procedure for removing small polyps with markedly similar curability and safety to hot polypectomy.

  8. Superposed epoch analysis of O+ auroral outflow during sawtooth events and substorms

    NASA Astrophysics Data System (ADS)

    Nowrouzi, N.; Kistler, L. M.; Lund, E. J.; Cai, X.

    2017-12-01

    Sawtooth events are repeated injection of energetic particles at geosynchronous orbit. Studies have shown that 94% of sawtooth events occurred during magnetic storm times. The main factor that causes a sawtooth event is still an open question. Simulations have suggested that heavy ions like O+ may play a role in triggering the injections. One of the sources of the O+ in the Earth's magnetosphere is the nightside aurora. O+ ions coming from the nightside auroral region have direct access to the near-earth magnetotail. A model (Brambles et al. 2013) for interplanetary coronal mass ejection driven sawtooth events found that nightside O+ outflow caused the subsequent teeth of the sawtooth event through a feedback mechanism. This work is a superposed epoch analysis to test whether the observed auroral outflow supports this model. Using FAST spacecraft data from 1997-2007, we examine the auroral O+ outflow as a function of time relative to an injection onset. Then we determine whether the profile of outflow flux of O+ during sawtooth events is different from the outflow observed during isolated substorms. The auroral region boundaries are estimated using the method of (Andersson et al. 2004). Subsequently the O+ outflow flux inside these boundaries are calculated and binned as a function of superposed epoch time for substorms and sawtooth "teeth". In this way, we will determine if sawtooth events do in fact have greater O+ outflow, and if that outflow is predominantly from the nightside, as suggested by the model results.

  9. Low Probability Tail Event Analysis and Mitigation in BPA Control Area: Task 2 Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lu, Shuai; Makarov, Yuri V.; McKinstry, Craig A.

    Task report detailing low probability tail event analysis and mitigation in BPA control area. Tail event refers to the situation in a power system when unfavorable forecast errors of load and wind are superposed onto fast load and wind ramps, or non-wind generators falling short of scheduled output, causing the imbalance between generation and load to become very significant.

  10. Sources of Error and the Statistical Formulation of M S: m b Seismic Event Screening Analysis

    NASA Astrophysics Data System (ADS)

    Anderson, D. N.; Patton, H. J.; Taylor, S. R.; Bonner, J. L.; Selby, N. D.

    2014-03-01

    The Comprehensive Nuclear-Test-Ban Treaty (CTBT), a global ban on nuclear explosions, is currently in a ratification phase. Under the CTBT, an International Monitoring System (IMS) of seismic, hydroacoustic, infrasonic and radionuclide sensors is operational, and the data from the IMS is analysed by the International Data Centre (IDC). The IDC provides CTBT signatories basic seismic event parameters and a screening analysis indicating whether an event exhibits explosion characteristics (for example, shallow depth). An important component of the screening analysis is a statistical test of the null hypothesis H 0: explosion characteristics using empirical measurements of seismic energy (magnitudes). The established magnitude used for event size is the body-wave magnitude (denoted m b) computed from the initial segment of a seismic waveform. IDC screening analysis is applied to events with m b greater than 3.5. The Rayleigh wave magnitude (denoted M S) is a measure of later arriving surface wave energy. Magnitudes are measurements of seismic energy that include adjustments (physical correction model) for path and distance effects between event and station. Relative to m b, earthquakes generally have a larger M S magnitude than explosions. This article proposes a hypothesis test (screening analysis) using M S and m b that expressly accounts for physical correction model inadequacy in the standard error of the test statistic. With this hypothesis test formulation, the 2009 Democratic Peoples Republic of Korea announced nuclear weapon test fails to reject the null hypothesis H 0: explosion characteristics.

  11. Climate network analysis of regional precipitation extremes: The true story told by event synchronization

    NASA Astrophysics Data System (ADS)

    Odenweller, Adrian; Donner, Reik V.

    2017-04-01

    Over the last decade, complex network methods have been frequently used for characterizing spatio-temporal patterns of climate variability from a complex systems perspective, yielding new insights into time-dependent teleconnectivity patterns and couplings between different components of the Earth climate. Among the foremost results reported, network analyses of the synchronicity of extreme events as captured by the so-called event synchronization have been proposed to be powerful tools for disentangling the spatio-temporal organization of particularly extreme rainfall events and anticipating the timing of monsoon onsets or extreme floodings. Rooted in the analysis of spike train synchrony analysis in the neurosciences, event synchronization has the great advantage of automatically classifying pairs of events arising at two distinct spatial locations as temporally close (and, thus, possibly statistically - or even dynamically - interrelated) or not without the necessity of selecting an additional parameter in terms of a maximally tolerable delay between these events. This consideration is conceptually justified in case of the original application to spike trains in electroencephalogram (EEG) recordings, where the inter-spike intervals show relatively narrow distributions at high temporal sampling rates. However, in case of climate studies, precipitation extremes defined by daily precipitation sums exceeding a certain empirical percentile of their local distribution exhibit a distinctively different type of distribution of waiting times between subsequent events. This raises conceptual concerns if event synchronization is still appropriate for detecting interlinkages between spatially distributed precipitation extremes. In order to study this problem in more detail, we employ event synchronization together with an alternative similarity measure for event sequences, event coincidence rates, which requires a manual setting of the tolerable maximum delay between two

  12. A case-crossover analysis of forest fire haze events and mortality in Malaysia

    NASA Astrophysics Data System (ADS)

    Sahani, Mazrura; Zainon, Nurul Ashikin; Wan Mahiyuddin, Wan Rozita; Latif, Mohd Talib; Hod, Rozita; Khan, Md Firoz; Tahir, Norhayati Mohd; Chan, Chang-Chuan

    2014-10-01

    The Southeast Asian (SEA) haze events due to forest fires are recurrent and affect Malaysia, particularly the Klang Valley region. The aim of this study is to examine the risk of haze days due to biomass burning in Southeast Asia on daily mortality in the Klang Valley region between 2000 and 2007. We used a case-crossover study design to model the effect of haze based on PM10 concentration to the daily mortality. The time-stratified control sampling approach was used, adjusted for particulate matter (PM10) concentrations, time trends and meteorological influences. Based on time series analysis of PM10 and backward trajectory analysis, haze days were defined when daily PM10 concentration exceeded 100 μg/m3. The results showed a total of 88 haze days were identified in the Klang Valley region during the study period. A total of 126,822 cases of death were recorded for natural mortality where respiratory mortality represented 8.56% (N = 10,854). Haze events were found to be significantly associated with natural and respiratory mortality at various lags. For natural mortality, haze events at lagged 2 showed significant association with children less than 14 years old (Odd Ratio (OR) = 1.41; 95% Confidence Interval (CI) = 1.01-1.99). Respiratory mortality was significantly associated with haze events for all ages at lagged 0 (OR = 1.19; 95% CI = 1.02-1.40). Age-and-gender-specific analysis showed an incremental risk of respiratory mortality among all males and elderly males above 60 years old at lagged 0 (OR = 1.34; 95% CI = 1.09-1.64 and OR = 1.41; 95% CI = 1.09-1.84 respectively). Adult females aged 15-59 years old were found to be at highest risk of respiratory mortality at lagged 5 (OR = 1.66; 95% CI = 1.03-1.99). This study clearly indicates that exposure to haze events showed immediate and delayed effects on mortality.

  13. Using Pattern Recognition and Discriminance Analysis to Predict Critical Events in Large Signal Databases

    NASA Astrophysics Data System (ADS)

    Feller, Jens; Feller, Sebastian; Mauersberg, Bernhard; Mergenthaler, Wolfgang

    2009-09-01

    Many applications in plant management require close monitoring of equipment performance, in particular with the objective to prevent certain critical events. At each point in time, the information available to classify the criticality of the process, is represented through the historic signal database as well as the actual measurement. This paper presents an approach to detect and predict critical events, based on pattern recognition and discriminance analysis.

  14. Gait Event Detection in Real-World Environment for Long-Term Applications: Incorporating Domain Knowledge Into Time-Frequency Analysis.

    PubMed

    Khandelwal, Siddhartha; Wickstrom, Nicholas

    2016-12-01

    Detecting gait events is the key to many gait analysis applications that would benefit from continuous monitoring or long-term analysis. Most gait event detection algorithms using wearable sensors that offer a potential for use in daily living have been developed from data collected in controlled indoor experiments. However, for real-word applications, it is essential that the analysis is carried out in humans' natural environment; that involves different gait speeds, changing walking terrains, varying surface inclinations and regular turns among other factors. Existing domain knowledge in the form of principles or underlying fundamental gait relationships can be utilized to drive and support the data analysis in order to develop robust algorithms that can tackle real-world challenges in gait analysis. This paper presents a novel approach that exhibits how domain knowledge about human gait can be incorporated into time-frequency analysis to detect gait events from long-term accelerometer signals. The accuracy and robustness of the proposed algorithm are validated by experiments done in indoor and outdoor environments with approximately 93 600 gait events in total. The proposed algorithm exhibits consistently high performance scores across all datasets in both, indoor and outdoor environments.

  15. Dynamic event tree analysis with the SAS4A/SASSYS-1 safety analysis code

    DOE PAGES

    Jankovsky, Zachary K.; Denman, Matthew R.; Aldemir, Tunc

    2018-02-02

    The consequences of a transient in an advanced sodium-cooled fast reactor are difficult to capture with the traditional approach to probabilistic risk assessment (PRA). Numerous safety-relevant systems are passive and may have operational states that cannot be represented by binary success or failure. In addition, the specific order and timing of events may be crucial which necessitates the use of dynamic PRA tools such as ADAPT. The modifications to the SAS4A/SASSYS-1 sodium-cooled fast reactor safety analysis code for linking it to ADAPT to perform a dynamic PRA are described. A test case is used to demonstrate the linking process andmore » to illustrate the type of insights that may be gained with this process. Finally, newly-developed dynamic importance measures are used to assess the significance of reactor parameters/constituents on calculated consequences of initiating events.« less

  16. Dynamic event tree analysis with the SAS4A/SASSYS-1 safety analysis code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jankovsky, Zachary K.; Denman, Matthew R.; Aldemir, Tunc

    The consequences of a transient in an advanced sodium-cooled fast reactor are difficult to capture with the traditional approach to probabilistic risk assessment (PRA). Numerous safety-relevant systems are passive and may have operational states that cannot be represented by binary success or failure. In addition, the specific order and timing of events may be crucial which necessitates the use of dynamic PRA tools such as ADAPT. The modifications to the SAS4A/SASSYS-1 sodium-cooled fast reactor safety analysis code for linking it to ADAPT to perform a dynamic PRA are described. A test case is used to demonstrate the linking process andmore » to illustrate the type of insights that may be gained with this process. Finally, newly-developed dynamic importance measures are used to assess the significance of reactor parameters/constituents on calculated consequences of initiating events.« less

  17. Video Traffic Analysis for Abnormal Event Detection

    DOT National Transportation Integrated Search

    2010-01-01

    We propose the use of video imaging sensors for the detection and classification of abnormal events to be used primarily for mitigation of traffic congestion. Successful detection of such events will allow for new road guidelines; for rapid deploymen...

  18. Video traffic analysis for abnormal event detection.

    DOT National Transportation Integrated Search

    2010-01-01

    We propose the use of video imaging sensors for the detection and classification of abnormal events to : be used primarily for mitigation of traffic congestion. Successful detection of such events will allow for : new road guidelines; for rapid deplo...

  19. Effects of Interventions on Relapse to Narcotics Addiction: An Event-History Analysis.

    ERIC Educational Resources Information Center

    Hser, Yih-Ing; And Others

    1995-01-01

    Event-history analysis was applied to the life history data of 581 male narcotics addicts to specify the concurrent, postintervention, and durational effects of social interventions on relapse to narcotics use. Results indicate the advisability of supporting methadone maintenance with other prevention strategies. (SLD)

  20. Characterization of Strombolian events by using independent component analysis

    NASA Astrophysics Data System (ADS)

    Ciaramella, A.; de Lauro, E.; de Martino, S.; di Lieto, B.; Falanga, M.; Tagliaferri, R.

    2004-10-01

    We apply Independent Component Analysis (ICA) to seismic signals recorded at Stromboli volcano. Firstly, we show how ICA works considering synthetic signals, which are generated by dynamical systems. We prove that Strombolian signals, both tremor and explosions, in the high frequency band (>0.5 Hz), are similar in time domain. This seems to give some insights to the organ pipe model generation for the source of these events. Moreover, we are able to recognize in the tremor signals a low frequency component (<0.5 Hz), with a well defined peak corresponding to 30s.

  1. Investigating cardiorespiratory interaction by cross-spectral analysis of event series

    NASA Astrophysics Data System (ADS)

    Schäfer, Carsten; Rosenblum, Michael G.; Pikovsky, Arkady S.; Kurths, Jürgen

    2000-02-01

    The human cardiovascular and respiratory systems interact with each other and show effects of modulation and synchronization. Here we present a cross-spectral technique that specifically considers the event-like character of the heartbeat and avoids typical restrictions of other spectral methods. Using models as well as experimental data, we demonstrate how modulation and synchronization can be distinguished. Finally, we compare the method to traditional techniques and to the analysis of instantaneous phases.

  2. Analysis of warm convective rain events in Catalonia

    NASA Astrophysics Data System (ADS)

    Ballart, D.; Figuerola, F.; Aran, M.; Rigo, T.

    2009-09-01

    Between the end of September and November, events with high amounts of rainfall are quite common in Catalonia. The high sea surface temperature of the Mediterranean Sea near to the Catalan Coast is one of the most important factors that help to the development of this type of storms. Some of these events have particular characteristics: elevated rain rate during short time periods, not very deep convection and low lightning activity. Consequently, the use of remote sensing tools for the surveillance is quite useless or limited. With reference to the high rain efficiency, this is caused by internal mechanisms of the clouds, and also by the air mass where the precipitation structure is developed. As aforementioned, the contribution of the sea to the air mass is very relevant, not only by the increase of the big condensation nuclei, but also by high temperature of the low layers of the atmosphere, where are allowed clouds with 5 or 6 km of particles in liquid phase. In fact, the freezing level into these clouds can be detected by -15ºC. Due to these characteristics, this type of rainy structures can produce high quantities of rainfall in a relatively brief period of time, and, in the case to be quasi-stationary, precipitation values at surface could be very important. From the point of view of remote sensing tools, the cloud nature implies that the different tools and methodologies commonly used for the analysis of heavy rain events are not useful. This is caused by the following features: lightning are rarely observed, the top temperatures of clouds are not cold enough to be enhanced in the satellite imagery, and, finally, reflectivity radar values are lower than other heavy rain cases. The third point to take into account is the vulnerability of the affected areas. An elevated percentage of the Catalan population lives in the coastal region. In the central coast of Catalonia, the urban areas are surrounded by a not very high mountain range with small basins and

  3. Analysis of extreme rainfall events using attributes control charts in temporal rainfall processes

    NASA Astrophysics Data System (ADS)

    Villeta, María; Valencia, Jose Luis; Saá-Requejo, Antonio; María Tarquis, Ana

    2015-04-01

    The impacts of most intense rainfall events on agriculture and insurance industry can be very severe. This research focuses in the analysis of extreme rainfall events throughout the use of attributes control charts, which constitutes a usual tool in Statistical Process Control (SPC) but unusual in climate studios. Here, series of daily precipitations for the years 1931-2009 within a Spanish region are analyzed, based on a new type of attributes control chart that takes into account the autocorrelation between the extreme rainfall events. The aim is to conclude if there exist or not evidence of a change in the extreme rainfall model of the considered series. After adjusting seasonally the precipitation series and considering the data of the first 30 years, a frequency-based criterion allowed fixing specification limits in order to discriminate between extreme observed rainfall days and normal observed rainfall days. The autocorrelation amongst maximum precipitation is taken into account by a New Binomial Markov Extended Process obtained for each rainfall series. These modelling of the extreme rainfall processes provide a way to generate the attributes control charts for the annual fraction of rainfall extreme days. The extreme rainfall processes along the rest of the years under study can then be monitored by such attributes control charts. The results of the application of this methodology show evidence of change in the model of extreme rainfall events in some of the analyzed precipitation series. This suggests that the attributes control charts proposed for the analysis of the most intense precipitation events will be of practical interest to agriculture and insurance sectors in next future.

  4. Analysis of the Impact of Climate Change on Extreme Hydrological Events in California

    NASA Astrophysics Data System (ADS)

    Ashraf Vaghefi, Saeid; Abbaspour, Karim C.

    2016-04-01

    Estimating magnitude and occurrence frequency of extreme hydrological events is required for taking preventive remedial actions against the impact of climate change on the management of water resources. Examples include: characterization of extreme rainfall events to predict urban runoff, determination of river flows, and the likely severity of drought events during the design life of a water project. In recent years California has experienced its most severe drought in recorded history, causing water stress, economic loss, and an increase in wildfires. In this paper we describe development of a Climate Change Toolkit (CCT) and demonstrate its use in the analysis of dry and wet periods in California for the years 2020-2050 and compare the results with the historic period 1975-2005. CCT provides four modules to: i) manage big databases such as those of Global Climate Models (GCMs), ii) make bias correction using observed local climate data , iii) interpolate gridded climate data to finer resolution, and iv) calculate continuous dry- and wet-day periods based on rainfall, temperature, and soil moisture for analysis of drought and flooding risks. We used bias-corrected meteorological data of five GCMs for extreme CO2 emission scenario rcp8.5 for California to analyze the trend of extreme hydrological events. The findings indicate that frequency of dry period will increase in center and southern parts of California. The assessment of the number of wet days and the frequency of wet periods suggests an increased risk of flooding in north and north-western part of California, especially in the coastal strip. Keywords: Climate Change Toolkit (CCT), Extreme Hydrological Events, California

  5. Video Analysis Verification of Head Impact Events Measured by Wearable Sensors.

    PubMed

    Cortes, Nelson; Lincoln, Andrew E; Myer, Gregory D; Hepburn, Lisa; Higgins, Michael; Putukian, Margot; Caswell, Shane V

    2017-08-01

    Wearable sensors are increasingly used to quantify the frequency and magnitude of head impact events in multiple sports. There is a paucity of evidence that verifies head impact events recorded by wearable sensors. To utilize video analysis to verify head impact events recorded by wearable sensors and describe the respective frequency and magnitude. Cohort study (diagnosis); Level of evidence, 2. Thirty male (mean age, 16.6 ± 1.2 years; mean height, 1.77 ± 0.06 m; mean weight, 73.4 ± 12.2 kg) and 35 female (mean age, 16.2 ± 1.3 years; mean height, 1.66 ± 0.05 m; mean weight, 61.2 ± 6.4 kg) players volunteered to participate in this study during the 2014 and 2015 lacrosse seasons. Participants were instrumented with GForceTracker (GFT; boys) and X-Patch sensors (girls). Simultaneous game video was recorded by a trained videographer using a single camera located at the highest midfield location. One-third of the field was framed and panned to follow the ball during games. Videographic and accelerometer data were time synchronized. Head impact counts were compared with video recordings and were deemed valid if (1) the linear acceleration was ≥20 g, (2) the player was identified on the field, (3) the player was in camera view, and (4) the head impact mechanism could be clearly identified. Descriptive statistics of peak linear acceleration (PLA) and peak rotational velocity (PRV) for all verified head impacts ≥20 g were calculated. For the boys, a total recorded 1063 impacts (2014: n = 545; 2015: n = 518) were logged by the GFT between game start and end times (mean PLA, 46 ± 31 g; mean PRV, 1093 ± 661 deg/s) during 368 player-games. Of these impacts, 690 were verified via video analysis (65%; mean PLA, 48 ± 34 g; mean PRV, 1242 ± 617 deg/s). The X-Patch sensors, worn by the girls, recorded a total 180 impacts during the course of the games, and 58 (2014: n = 33; 2015: n = 25) were verified via video analysis (32%; mean PLA, 39 ± 21 g; mean PRV, 1664

  6. Statistical analysis of mixed recurrent event data with application to cancer survivor study

    PubMed Central

    Zhu, Liang; Tong, Xingwei; Zhao, Hui; Sun, Jianguo; Srivastava, Deo Kumar; Leisenring, Wendy; Robison, Leslie L.

    2014-01-01

    Event history studies occur in many fields including economics, medical studies and social science. In such studies concerning some recurrent events, two types of data have been extensively discussed in the literature. One is recurrent event data that arise if study subjects are monitored or observed continuously. In this case, the observed information provides the times of all occurrences of the recurrent events of interest. The other is panel count data, which occur if the subjects are monitored or observed only periodically. This can happen if the continuous observation is too expensive or not practical and in this case, only the numbers of occurrences of the events between subsequent observation times are available. In this paper, we discuss a third type of data, which is a mixture of recurrent event and panel count data and for which there exists little literature. For regression analysis of such data, a marginal mean model is presented and we propose an estimating equation-based approach for estimation of regression parameters. A simulation study is conducted to assess the finite sample performance of the proposed methodology and indicates that it works well for practical situations. Finally it is applied to a motivating study on childhood cancer survivors. PMID:23139023

  7. Identification of homogeneous regions for rainfall regional frequency analysis considering typhoon event in South Korea

    NASA Astrophysics Data System (ADS)

    Heo, J. H.; Ahn, H.; Kjeldsen, T. R.

    2017-12-01

    South Korea is prone to large, and often disastrous, rainfall events caused by a mixture of monsoon and typhoon rainfall phenomena. However, traditionally, regional frequency analysis models did not consider this mixture of phenomena when fitting probability distributions, potentially underestimating the risk posed by the more extreme typhoon events. Using long-term observed records of extreme rainfall from 56 sites combined with detailed information on the timing and spatial impact of past typhoons from the Korea Meteorological Administration (KMA), this study developed and tested a new mixture model for frequency analysis of two different phenomena; events occurring regularly every year (monsoon) and events only occurring in some years (typhoon). The available annual maximum 24 hour rainfall data were divided into two sub-samples corresponding to years where the annual maximum is from either (1) a typhoon event, or (2) a non-typhoon event. Then, three-parameter GEV distribution was fitted to each sub-sample along with a weighting parameter characterizing the proportion of historical events associated with typhoon events. Spatial patterns of model parameters were analyzed and showed that typhoon events are less commonly associated with annual maximum rainfall in the North-West part of the country (Seoul area), and more prevalent in the southern and eastern parts of the country, leading to the formation of two distinct typhoon regions: (1) North-West; and (2) Southern and Eastern. Using a leave-one-out procedure, a new regional frequency model was tested and compared to a more traditional index flood method. The results showed that the impact of typhoon on design events might previously have been underestimated in the Seoul area. This suggests that the use of the mixture model should be preferred where the typhoon phenomena is less frequent, and thus can have a significant effect on the rainfall-frequency curve. This research was supported by a grant(2017-MPSS31

  8. Uncertainty analysis in fault tree models with dependent basic events.

    PubMed

    Pedroni, Nicola; Zio, Enrico

    2013-06-01

    In general, two types of dependence need to be considered when estimating the probability of the top event (TE) of a fault tree (FT): "objective" dependence between the (random) occurrences of different basic events (BEs) in the FT and "state-of-knowledge" (epistemic) dependence between estimates of the epistemically uncertain probabilities of some BEs of the FT model. In this article, we study the effects on the TE probability of objective and epistemic dependences. The well-known Frèchet bounds and the distribution envelope determination (DEnv) method are used to model all kinds of (possibly unknown) objective and epistemic dependences, respectively. For exemplification, the analyses are carried out on a FT with six BEs. Results show that both types of dependence significantly affect the TE probability; however, the effects of epistemic dependence are likely to be overwhelmed by those of objective dependence (if present). © 2012 Society for Risk Analysis.

  9. An Unsupervised Anomalous Event Detection and Interactive Analysis Framework for Large-scale Satellite Data

    NASA Astrophysics Data System (ADS)

    LIU, Q.; Lv, Q.; Klucik, R.; Chen, C.; Gallaher, D. W.; Grant, G.; Shang, L.

    2016-12-01

    Due to the high volume and complexity of satellite data, computer-aided tools for fast quality assessments and scientific discovery are indispensable for scientists in the era of Big Data. In this work, we have developed a framework for automated anomalous event detection in massive satellite data. The framework consists of a clustering-based anomaly detection algorithm and a cloud-based tool for interactive analysis of detected anomalies. The algorithm is unsupervised and requires no prior knowledge of the data (e.g., expected normal pattern or known anomalies). As such, it works for diverse data sets, and performs well even in the presence of missing and noisy data. The cloud-based tool provides an intuitive mapping interface that allows users to interactively analyze anomalies using multiple features. As a whole, our framework can (1) identify outliers in a spatio-temporal context, (2) recognize and distinguish meaningful anomalous events from individual outliers, (3) rank those events based on "interestingness" (e.g., rareness or total number of outliers) defined by users, and (4) enable interactively query, exploration, and analysis of those anomalous events. In this presentation, we will demonstrate the effectiveness and efficiency of our framework in the application of detecting data quality issues and unusual natural events using two satellite datasets. The techniques and tools developed in this project are applicable for a diverse set of satellite data and will be made publicly available for scientists in early 2017.

  10. An analysis of high-impact, low-predictive skill severe weather events in the northeast U.S

    NASA Astrophysics Data System (ADS)

    Vaughan, Matthew T.

    An objective evaluation of Storm Prediction Center slight risk convective outlooks, as well as a method to identify high-impact severe weather events with poor-predictive skill are presented in this study. The objectives are to assess severe weather forecast skill over the northeast U.S. relative to the continental U.S., build a climatology of high-impact, low-predictive skill events between 1980--2013, and investigate the dynamic and thermodynamic differences between severe weather events with low-predictive skill and high-predictive skill over the northeast U.S. Severe storm reports of hail, wind, and tornadoes are used to calculate skill scores including probability of detection (POD), false alarm ratio (FAR) and threat scores (TS) for each convective outlook. Low predictive skill events are binned into low POD (type 1) and high FAR (type 2) categories to assess temporal variability of low-predictive skill events. Type 1 events were found to occur in every year of the dataset with an average of 6 events per year. Type 2 events occur less frequently and are more common in the earlier half of the study period. An event-centered composite analysis is performed on the low-predictive skill database using the National Centers for Environmental Prediction Climate Forecast System Reanalysis 0.5° gridded dataset to analyze the dynamic and thermodynamic conditions prior to high-impact severe weather events with varying predictive skill. Deep-layer vertical shear between 1000--500 hPa is found to be a significant discriminator in slight risk forecast skill where high-impact events with less than 31-kt shear have lower threat scores than high-impact events with higher shear values. Case study analysis of type 1 events suggests the environment over which severe weather occurs is characterized by high downdraft convective available potential energy, steep low-level lapse rates, and high lifting condensation level heights that contribute to an elevated risk of severe wind.

  11. A formal framework of scenario creation and analysis of extreme hydrological events

    NASA Astrophysics Data System (ADS)

    Lohmann, D.

    2007-12-01

    We are presenting a formal framework for a hydrological risk analysis. Different measures of risk will be introduced, such as average annual loss or occurrence exceedance probability. These are important measures for e.g. insurance companies to determine the cost of insurance. One key aspect of investigating the potential consequences of extreme hydrological events (floods and draughts) is the creation of meteorological scenarios that reflect realistic spatial and temporal patterns of precipitation that also have correct local statistics. 100,000 years of these meteorological scenarios are used in a calibrated rainfall-runoff-flood-loss-risk model to produce flood and draught events that have never been observed. The results of this hazard model are statistically analyzed and linked to socio-economic data and vulnerability functions to show the impact of severe flood events. We are showing results from the Risk Management Solutions (RMS) Europe Flood Model to introduce this formal framework.

  12. "That in your hands". A comprehensive process analysis of a significant event in psychotherapy.

    PubMed

    Elliott, R

    1983-05-01

    This article illustrates a new approach to the study of change processes in psychotherapy. The approach involves selecting significant change events and analyzing them according to the Comprehensive Process Model. In this model, client and therapist behaviors are analyzed for content, interpersonal action, style and response quality by using information derived from Interpersonal Process Recall, client and therapist objective process ratings and qualitative analyses. The event selected for analysis in this paper was rated by client and therapist as significantly helpful. The focal therapist response was a reflective-interpretive intervention in which the therapist collaboratively and evocatively expanded the client's implicit meanings. The event involved working through an earlier insight and realization of progress by the client. The event suggests an association between subjective "felt shifts" and public "process shifts" in client in-therapy behaviors. A model, consistent with Gendlin's experiential psychotherapy (1970), is offered to describe the change process which occurred in this event.

  13. Formal Analysis of BPMN Models Using Event-B

    NASA Astrophysics Data System (ADS)

    Bryans, Jeremy W.; Wei, Wei

    The use of business process models has gone far beyond documentation purposes. In the development of business applications, they can play the role of an artifact on which high level properties can be verified and design errors can be revealed in an effort to reduce overhead at later software development and diagnosis stages. This paper demonstrates how formal verification may add value to the specification, design and development of business process models in an industrial setting. The analysis of these models is achieved via an algorithmic translation from the de-facto standard business process modeling language BPMN to Event-B, a widely used formal language supported by the Rodin platform which offers a range of simulation and verification technologies.

  14. Observations on Rupture Behaviour of Fluid Induced Events at the Basel EGS Based on Empirical Green's Function Analysis

    NASA Astrophysics Data System (ADS)

    Folesky, J.; Kummerow, J.; Shapiro, S. A.; Asanuma, H.; Häring, M. O.

    2015-12-01

    The Emprirical Green's Function (EGF) method uses pairs of events of high wave form similarity and adjacent hypocenters to decompose the influences of source time function, ray path, instrument site, and instrument response. The seismogram of the smaller event is considered as the Green's Function which then can be deconvolved from the other seismogram. The result provides a reconstructed relative source time function (RSTF) of the larger event of that event pair. The comparison of the RSTFs at different stations of the observation systems produces information on the rupture process of the larger event based on the observation of the directivity effect and on changing RSTFs complexities.The Basel EGS dataset of 2006-2007 consists of about 2800 localized events of magnitudes between 0.0event pairs of adequate magnitude difference for EGF analysis. The data has sufficient quality to analyse events with magnitudes down to ML=0.5 for an apparent directivity effect although the approximate rupture duration for those events is of only a few milliseconds. The dataset shows a number of multiplets and repeating earthquakes known from earlier studies. The larger events seem to appear close to the rim of the microseismic cloud. We are interested in their rupture behaviour. Using the EGF method we compute rupture orientations for about 190 event pairs and relate them to the event location, the known fault system, and stress regime. For the majority of events we observe a similar rupture direction which seems to correlate with the over all shape of the microseismic cloud. The large events, however, point back to the injection source. Additionally the rupture direction fitting yields estimates for projections of the rupture velocity on the horizontal plane. They seem to vary between the multiplets in the reservoir from 0.3 to 0.7 times the S-wave velocity.To our knowledge source characterization by EGF analysis has not yet been introduced to microseismic reservoirs

  15. Toxicogenomics and cancer risk assessment: a framework for key event analysis and dose-response assessment for nongenotoxic carcinogens.

    PubMed

    Bercu, Joel P; Jolly, Robert A; Flagella, Kelly M; Baker, Thomas K; Romero, Pedro; Stevens, James L

    2010-12-01

    In order to determine a threshold for nongenotoxic carcinogens, the traditional risk assessment approach has been to identify a mode of action (MOA) with a nonlinear dose-response. The dose-response for one or more key event(s) linked to the MOA for carcinogenicity allows a point of departure (POD) to be selected from the most sensitive effect dose or no-effect dose. However, this can be challenging because multiple MOAs and key events may exist for carcinogenicity and oftentimes extensive research is required to elucidate the MOA. In the present study, a microarray analysis was conducted to determine if a POD could be identified following short-term oral rat exposure with two nongenotoxic rodent carcinogens, fenofibrate and methapyrilene, using a benchmark dose analysis of genes aggregated in Kyoto Encyclopedia of Genes and Genomes (KEGG) pathways and Gene Ontology (GO) biological processes, which likely encompass key event(s) for carcinogenicity. The gene expression response for fenofibrate given to rats for 2days was consistent with its MOA and known key events linked to PPARα activation. The temporal response from daily dosing with methapyrilene demonstrated biological complexity with waves of pathways/biological processes occurring over 1, 3, and 7days; nonetheless, the benchmark dose values were consistent over time. When comparing the dose-response of toxicogenomic data to tumorigenesis or precursor events, the toxicogenomics POD was slightly below any effect level. Our results suggest that toxicogenomic analysis using short-term studies can be used to identify a threshold for nongenotoxic carcinogens based on evaluation of potential key event(s) which then can be used within a risk assessment framework. Copyright © 2010 Elsevier Inc. All rights reserved.

  16. FASEA: A FPGA Acquisition System and Software Event Analysis for liquid scintillation counting

    NASA Astrophysics Data System (ADS)

    Steele, T.; Mo, L.; Bignell, L.; Smith, M.; Alexiev, D.

    2009-10-01

    The FASEA (FPGA based Acquisition and Software Event Analysis) system has been developed to replace the MAC3 for coincidence pulse processing. The system uses a National Instruments Virtex 5 FPGA card (PXI-7842R) for data acquisition and a purpose developed data analysis software for data analysis. Initial comparisons to the MAC3 unit are included based on measurements of 89Sr and 3H, confirming that the system is able to accurately emulate the behaviour of the MAC3 unit.

  17. The Frasnian-Famennian mass killing event(s), methods of identification and evaluation

    NASA Technical Reports Server (NTRS)

    Geldsetzer, H. H. J.

    1988-01-01

    The absence of an abnormally high number of earlier Devonian taxa from Famennian sediments was repeatedly documented and can hardly be questioned. Primary recognition of the event(s) was based on paleontological data, especially common macrofossils. Most paleontologists place the disappearance of these common forms at the gigas/triangularis contact and this boundary was recently proposed as the Frasnian-Famennian (F-F) boundary. Not unexpectedly, alternate F-F positions were suggested caused by temporary Frasnian survivors or sudden post-event radiations of new forms. Secondary supporting evidence for mass killing event(s) is supplied by trace element and stable isotope geochemistry but not with the same success as for the K/T boundary, probably due to additional 300 ma of tectonic and diagenetic overprinting. Another tool is microfacies analysis which is surprisingly rarely used even though it can explain geochemical anomalies or paleontological overlap not detectable by conventional macrofacies analysis. The combination of microfacies analysis and geochemistry was applied at two F-F sections in western Canada and showed how interdependent the two methods are. Additional F-F sections from western Canada, western United States, France, Germany and Australia were sampled or re-sampled and await geochemical/microfacies evaluation.

  18. Adverse events with bismuth salts for Helicobacter pylori eradication: Systematic review and meta-analysis

    PubMed Central

    Ford, Alexander C; Malfertheiner, Peter; Giguère, Monique; Santana, José; Khan, Mostafizur; Moayyedi, Paul

    2008-01-01

    AIM: To assess the safety of bismuth used in Helicobacter pylori (H pylori) eradication therapy regimens. METHODS: We conducted a systematic review and meta-analysis. MEDLINE and EMBASE were searched (up to October 2007) to identify randomised controlled trials comparing bismuth with placebo or no treatment, or bismuth salts in combination with antibiotics as part of eradication therapy with the same dose and duration of antibiotics alone or, in combination, with acid suppression. Total numbers of adverse events were recorded. Data were pooled and expressed as relative risks with 95% confidence intervals (CI). RESULTS: We identified 35 randomised controlled trials containing 4763 patients. There were no serious adverse events occurring with bismuth therapy. There was no statistically significant difference detected in total adverse events with bismuth [relative risk (RR) = 1.01; 95% CI: 0.87-1.16], specific individual adverse events, with the exception of dark stools (RR = 5.06; 95% CI: 1.59-16.12), or adverse events leading to withdrawal of therapy (RR = 0.86; 95% CI: 0.54-1.37). CONCLUSION: Bismuth for the treatment of H pylori is safe and well-tolerated. The only adverse event occurring significantly more commonly was dark stools. PMID:19109870

  19. Bayesian analysis of caustic-crossing microlensing events

    NASA Astrophysics Data System (ADS)

    Cassan, A.; Horne, K.; Kains, N.; Tsapras, Y.; Browne, P.

    2010-06-01

    Aims: Caustic-crossing binary-lens microlensing events are important anomalous events because they are capable of detecting an extrasolar planet companion orbiting the lens star. Fast and robust modelling methods are thus of prime interest in helping to decide whether a planet is detected by an event. Cassan introduced a new set of parameters to model binary-lens events, which are closely related to properties of the light curve. In this work, we explain how Bayesian priors can be added to this framework, and investigate on interesting options. Methods: We develop a mathematical formulation that allows us to compute analytically the priors on the new parameters, given some previous knowledge about other physical quantities. We explicitly compute the priors for a number of interesting cases, and show how this can be implemented in a fully Bayesian, Markov chain Monte Carlo algorithm. Results: Using Bayesian priors can accelerate microlens fitting codes by reducing the time spent considering physically implausible models, and helps us to discriminate between alternative models based on the physical plausibility of their parameters.

  20. Element analysis: a wavelet-based method for analysing time-localized events in noisy time series

    PubMed Central

    2017-01-01

    A method is derived for the quantitative analysis of signals that are composed of superpositions of isolated, time-localized ‘events’. Here, these events are taken to be well represented as rescaled and phase-rotated versions of generalized Morse wavelets, a broad family of continuous analytic functions. Analysing a signal composed of replicates of such a function using another Morse wavelet allows one to directly estimate the properties of events from the values of the wavelet transform at its own maxima. The distribution of events in general power-law noise is determined in order to establish significance based on an expected false detection rate. Finally, an expression for an event’s ‘region of influence’ within the wavelet transform permits the formation of a criterion for rejecting spurious maxima due to numerical artefacts or other unsuitable events. Signals can then be reconstructed based on a small number of isolated points on the time/scale plane. This method, termed element analysis, is applied to the identification of long-lived eddy structures in ocean currents as observed by along-track measurements of sea surface elevation from satellite altimetry. PMID:28484325

  1. Low Probability Tail Event Analysis and Mitigation in BPA Control Area: Task One Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lu, Shuai; Makarov, Yuri V.

    This is a report for task one of the tail event analysis project for BPA. Tail event refers to the situation in a power system when unfavorable forecast errors of load and wind are superposed onto fast load and wind ramps, or non-wind generators falling short of scheduled output, the imbalance between generation and load becomes very significant. This type of events occurs infrequently and appears on the tails of the distribution of system power imbalance; therefore, is referred to as tail events. This report analyzes what happened during the Electric Reliability Council of Texas (ERCOT) reliability event on Februarymore » 26, 2008, which was widely reported because of the involvement of wind generation. The objective is to identify sources of the problem, solutions to it and potential improvements that can be made to the system. Lessons learned from the analysis include the following: (1) Large mismatch between generation and load can be caused by load forecast error, wind forecast error and generation scheduling control error on traditional generators, or a combination of all of the above; (2) The capability of system balancing resources should be evaluated both in capacity (MW) and in ramp rate (MW/min), and be procured accordingly to meet both requirements. The resources need to be able to cover a range corresponding to the variability of load and wind in the system, additional to other uncertainties; (3) Unexpected ramps caused by load and wind can both become the cause leading to serious issues; (4) A look-ahead tool evaluating system balancing requirement during real-time operations and comparing that with available system resources should be very helpful to system operators in predicting the forthcoming of similar events and planning ahead; and (5) Demand response (only load reduction in ERCOT event) can effectively reduce load-generation mismatch and terminate frequency deviation in an emergency situation.« less

  2. A PDA-based system for online recording and analysis of concurrent events in complex behavioral processes.

    PubMed

    Held, Jürgen; Manser, Tanja

    2005-02-01

    This article outlines how a Palm- or Newton-based PDA (personal digital assistant) system for online event recording was used to record and analyze concurrent events. We describe the features of this PDA-based system, called the FIT-System (flexible interface technique), and its application to the analysis of concurrent events in complex behavioral processes--in this case, anesthesia work processes. The patented FIT-System has a unique user interface design allowing the user to design an interface template with a pencil and paper or using a transparency film. The template usually consists of a drawing or sketch that includes icons or symbols that depict the observer's representation of the situation to be observed. In this study, the FIT-System allowed us to create a design for fast, intuitive online recording of concurrent events using a set of 41 observation codes. An analysis of concurrent events leads to a description of action density, and our results revealed a characteristic distribution of action density during the administration of anesthesia in the operating room. This distribution indicated the central role of the overlapping operations in the action sequences of medical professionals as they deal with the varying requirements of this complex task. We believe that the FIT-System for online recording of concurrent events in complex behavioral processes has the potential to be useful across a broad spectrum of research areas.

  3. Propensity for Violence among Homeless and Runaway Adolescents: An Event History Analysis

    ERIC Educational Resources Information Center

    Crawford, Devan M.; Whitbeck, Les B.; Hoyt, Dan R.

    2011-01-01

    Little is known about the prevalence of violent behaviors among homeless and runaway adolescents or the specific behavioral factors that influence violent behaviors across time. In this longitudinal study of 300 homeless and runaway adolescents aged 16 to 19 at baseline, the authors use event history analysis to assess the factors associated with…

  4. Performance Analysis of Cloud Computing Architectures Using Discrete Event Simulation

    NASA Technical Reports Server (NTRS)

    Stocker, John C.; Golomb, Andrew M.

    2011-01-01

    Cloud computing offers the economic benefit of on-demand resource allocation to meet changing enterprise computing needs. However, the flexibility of cloud computing is disadvantaged when compared to traditional hosting in providing predictable application and service performance. Cloud computing relies on resource scheduling in a virtualized network-centric server environment, which makes static performance analysis infeasible. We developed a discrete event simulation model to evaluate the overall effectiveness of organizations in executing their workflow in traditional and cloud computing architectures. The two part model framework characterizes both the demand using a probability distribution for each type of service request as well as enterprise computing resource constraints. Our simulations provide quantitative analysis to design and provision computing architectures that maximize overall mission effectiveness. We share our analysis of key resource constraints in cloud computing architectures and findings on the appropriateness of cloud computing in various applications.

  5. Composite Analysis of Cold Season Atmospheric River Events: Extreme Precipitation and Flooding over the Western United States

    NASA Astrophysics Data System (ADS)

    Eldardiry, H.; Hossain, F.

    2017-12-01

    Atmospheric Rivers (ARs) are narrow elongated corridors with horizontal water vapor transport located within the warm sector of extratropical cyclones. While it is widely known that most of heavy rainfall events across the western United States (US) are driven by ARs, the connection between atmospheric conditions and precipitation during an AR event has not been fully documented. In this study, we present a statistical analysis of the connection between precipitation, temperature, wind, and snowpack during the cold season AR events hitting the coastal regions of the western US. For each AR event, the precipitation and other atmospheric variables are retrieved through the dynamic downscaling of NCEP/NCAR Reanalysis product using the Advanced Research Weather Research and Forecasting Model (ARW-WRF). The results show a low frequency of precipitation (below 0.3) during AR events that reflects the connection of AR with extreme precipitation. Examining the horizontal wind speed during AR events indicates a high correlation (above 0.7) with precipitation. In addition, high levels of snow water equivalence (SWE) are also noticed along the mountainous regions, e.g., Cascade Range and Sierra-Nevada mountain range, during most of AR events. Addressing the impact of duration on the frequency of precipitation, we develop Intensity-Duration-Frequency (IDF) curves during AR events that can potentially describe the future predictability of precipitation along the north and south coast. To complement our analysis, we further investigate the flooding events recorded in the National Centers for Environmental Information (NCEI) storm events database. While some flooding events are attributed to heavy rainfall associated with an AR event, other flooding events are significantly connected to the increase in the snowmelt before the flooding date. Thus, we introduce an index that describes the contribution of rainfall vs snowmelt and categorizes the flooding events during an AR event

  6. Prehospital Interventions During Mass-Casualty Events in Afghanistan: A Case Analysis.

    PubMed

    Schauer, Steven G; April, Michael D; Simon, Erica; Maddry, Joseph K; Carter, Robert; Delorenzo, Robert A

    2017-08-01

    Mass-casualty (MASCAL) events are known to occur in the combat setting. There are very limited data at this time from the Joint Theater (Iraq and Afghanistan) wars specific to MASCAL events. The purpose of this report was to provide preliminary data for the development of prehospital planning and guidelines. Cases were identified using the Department of Defense (DoD; Virginia USA) Trauma Registry (DoDTR) and the Prehospital Trauma Registry (PHTR). These cases were identified as part of a research study evaluating Tactical Combat Casualty Care (TCCC) guidelines. Cases that were designated as or associated with denoted MASCAL events were included. Data Fifty subjects were identified during the course of this project. Explosives were the most common cause of injuries. There was a wide range of vital signs. Tourniquet placement and pressure dressings were the most common interventions, followed by analgesia administration. Oral transmucosal fentanyl citrate (OTFC) was the most common parenteral analgesic drug administered. Most were evacuated as "routine." Follow-up data were available for 36 of the subjects and 97% were discharged alive. The most common prehospital interventions were tourniquet and pressure dressing hemorrhage control, along with pain medication administration. Larger data sets are needed to guide development of MASCAL in-theater clinical practice guidelines. Schauer SG , April MD , Simon E , Maddry JK , Carter R III , Delorenzo RA . Prehospital interventions during mass-casualty events in Afghanistan: a case analysis. Prehosp Disaster Med. 2017;32(4):465-468.

  7. The added predictive value of biphasic events in ST analysis of the fetal electrocardiogram for intrapartum fetal monitoring.

    PubMed

    Becker, Jeroen H; Krikhaar, Anniek; Schuit, Ewoud; Mårtendal, Annika; Maršál, Karel; Kwee, Anneke; Visser, Gerard H A; Amer-Wåhlin, Isis

    2015-02-01

    To study the predictive value of biphasic ST-events for interventions for suspected fetal distress and adverse neonatal outcome, when using ST-analysis of the fetal electrocardiogram (FECG) for intrapartum fetal monitoring. Prospective cohort study. Three academic hospitals in Sweden. Women in labor with a high-risk singleton fetus in cephalic position beyond 36 weeks of gestation. In women in labor who were monitored with conventional cardiotocography, ST-waveform analysis was recorded and concealed. Traces with biphasic ST-events of the FECG (index) were compared with traces without biphasic events of the FECG. The ability of biphasic events to predict interventions for suspected fetal distress and adverse outcome was assessed using univariable and multivariable logistic regression analyses. Interventions for suspected fetal distress and adverse outcome (defined as presence of metabolic acidosis (i.e. umbilical cord pH <7.05 and base deficit in extracellular fluid >12 mmol), umbilical cord pH <7.00, 5-min Apgar score <7, admittance to neonatal intensive care unit or perinatal death). Although the presence of biphasic events of the FECG was associated with more interventions for fetal distress and an increased risk of adverse outcome compared with cases with no biphasic events, the presence of significant (i.e. intervention advised according to cardiotocography interpretation) biphasic events showed no independent association with interventions for fetal distress [odds ratio (OR) 1.71, 95% confidence interval (CI) 0.65-4.50] or adverse outcome (OR 1.96, 95% CI 0.74-5.24). The presence of significant biphasic events did not discriminate in the prediction of interventions for fetal distress or adverse outcome. Therefore, biphasic events in relation to ST-analysis monitoring during birth should be omitted if future studies confirm our findings. © 2014 Nordic Federation of Societies of Obstetrics and Gynecology.

  8. The time course of symbolic number adaptation: oscillatory EEG activity and event-related potential analysis.

    PubMed

    Hsu, Yi-Fang; Szűcs, Dénes

    2012-02-15

    Several functional magnetic resonance imaging (fMRI) studies have used neural adaptation paradigms to detect anatomical locations of brain activity related to number processing. However, currently not much is known about the temporal structure of number adaptation. In the present study, we used electroencephalography (EEG) to elucidate the time course of neural events in symbolic number adaptation. The numerical distance of deviants relative to standards was manipulated. In order to avoid perceptual confounds, all levels of deviants consisted of perceptually identical stimuli. Multiple successive numerical distance effects were detected in event-related potentials (ERPs). Analysis of oscillatory activity further showed at least two distinct stages of neural processes involved in the automatic analysis of numerical magnitude, with the earlier effect emerging at around 200ms and the later effect appearing at around 400ms. The findings support for the hypothesis that numerical magnitude processing involves a succession of cognitive events. Crown Copyright © 2011. Published by Elsevier Inc. All rights reserved.

  9. A cross-sectional analysis of pharmaceutical industry-funded events for health professionals in Australia

    PubMed Central

    Fabbri, Alice; Grundy, Quinn; Mintzes, Barbara; Swandari, Swestika; Moynihan, Ray; Walkom, Emily; Bero, Lisa A

    2017-01-01

    Objectives To analyse patterns and characteristics of pharmaceutical industry sponsorship of events for Australian health professionals and to understand the implications of recent changes in transparency provisions that no longer require reporting of payments for food and beverages. Design Cross-sectional analysis. Participants and setting 301 publicly available company transparency reports downloaded from the website of Medicines Australia, the pharmaceutical industry trade association, covering the period from October 2011 to September 2015. Results Forty-two companies sponsored 116 845 events for health professionals, on average 608 per week with 30 attendees per event. Events typically included a broad range of health professionals: 82.0% included medical doctors, including specialists and primary care doctors, and 38.3% trainees. Oncology, surgery and endocrinology were the most frequent clinical areas of focus. Most events (64.2%) were held in a clinical setting. The median cost per event was $A263 (IQR $A153–1195) and over 90% included food and beverages. Conclusions Over this 4-year period, industry-sponsored events were widespread and pharmaceutical companies maintained a high frequency of contact with health professionals. Most events were held in clinical settings, suggesting a pervasive commercial presence in everyday clinical practice. Food and beverages, known to be associated with changes to prescribing practice, were almost always provided. New Australian transparency provisions explicitly exclude meals from the reporting requirements; thus, a large proportion of potentially influential payments from pharmaceutical companies to health professionals will disappear from public view. PMID:28667226

  10. Time compression of soil erosion by the effect of largest daily event. A regional analysis of USLE database.

    NASA Astrophysics Data System (ADS)

    Gonzalez-Hidalgo, J. C.; Batalla, R.; Cerda, A.; de Luis, M.

    2009-04-01

    When Thornes and Brunsden wrote in 1977 "How often one hears the researcher (and no less the undergraduate) complain that after weeks of observation "nothing happened" only to learn that, the day after his departure, a flood caused unprecedent erosion and channel changes!" (Thornes and Brunsden, 1977, p. 57), they focussed on two different problems in geomorphological research: the effects of extreme events and the temporal compression of geomorphological processes. The time compression is one of the main characteristic of erosion processes. It means that an important amount of the total soil eroded is produced in very short temporal intervals, i.e. few events mostly related to extreme events. From magnitude-frequency analysis we know that few events, not necessarily extreme by magnitude, produce high amount of geomorphological work. Last but not least, extreme isolated events are a classical issue in geomorphology by their specific effects, and they are receiving permanent attention, increased at present because of scenarios of global change. Notwithstanding, the time compression of geomorphological processes could be focused not only on the analysis of extreme events and the traditional magnitude-frequency approach, but on new complementary approach based on the effects of largest events. The classical approach define extreme event as a rare event (identified by its magnitude and quantified by some deviation from central value), while we define largest events by the rank, whatever their magnitude. In a previous research on time compression of soil erosion, using USLE soil erosion database (Gonzalez-Hidalgo et al., EGU 2007), we described a relationship between the total amount of daily erosive events recorded by plot and the percentage contribution to total soil erosion of n-largest aggregated daily events. Now we offer a further refined analysis comparing different agricultural regions in USA. To do that we have analyzed data from 594 erosion plots from USLE

  11. Statistical Analysis of Solar Events Associated with SSC over Year of Solar Maximum during Cycle 23: 1. Identification of Related Sun-Earth Events

    NASA Astrophysics Data System (ADS)

    Grison, B.; Bocchialini, K.; Menvielle, M.; Chambodut, A.; Cornilleau-Wehrlin, N.; Fontaine, D.; Marchaudon, A.; Pick, M.; Pitout, F.; Schmieder, B.; Regnier, S.; Zouganelis, Y.

    2017-12-01

    Taking the 32 sudden storm commencements (SSC) listed by the observatory de l'Ebre / ISGI over the year 2002 (maximal solar activity) as a starting point, we performed a statistical analysis of the related solar sources, solar wind signatures, and terrestrial responses. For each event, we characterized and identified, as far as possible, (i) the sources on the Sun (Coronal Mass Ejections -CME-), with the help of a series of herafter detailed criteria (velocities, drag coefficient, radio waves, polarity), as well as (ii) the structure and properties in the interplanetary medium, at L1, of the event associated to the SSC: magnetic clouds -MC-, non-MC interplanetary coronal mass ejections -ICME-, co-rotating/stream interaction regions -SIR/CIR-, shocks only and unclear events that we call "miscellaneous" events. The categorization of the events at L1 is made on published catalogues. For each potential CME/L1 event association we compare the velocity observed at L1 with the one observed at the Sun and the estimated balistic velocity. Observations of radio emissions (Type II, Type IV detected from the ground and /or by WIND) associated to the CMEs make the solar source more probable. We also compare the polarity of the magnetic clouds with the hemisphere of the solar source. The drag coefficient (estimated with the drag-based model) is calculated for each potential association and it is compared to the expected range values. We identified a solar source for 26 SSC related events. 12 of these 26 associations match all criteria. We finally discuss the difficulty to perform such associations.

  12. Surrogate marker analysis in cancer clinical trials through time-to-event mediation techniques.

    PubMed

    Vandenberghe, Sjouke; Duchateau, Luc; Slaets, Leen; Bogaerts, Jan; Vansteelandt, Stijn

    2017-01-01

    The meta-analytic approach is the gold standard for validation of surrogate markers, but has the drawback of requiring data from several trials. We refine modern mediation analysis techniques for time-to-event endpoints and apply them to investigate whether pathological complete response can be used as a surrogate marker for disease-free survival in the EORTC 10994/BIG 1-00 randomised phase 3 trial in which locally advanced breast cancer patients were randomised to either taxane or anthracycline based neoadjuvant chemotherapy. In the mediation analysis, the treatment effect is decomposed into an indirect effect via pathological complete response and the remaining direct effect. It shows that only 4.2% of the treatment effect on disease-free survival after five years is mediated by the treatment effect on pathological complete response. There is thus no evidence from our analysis that pathological complete response is a valuable surrogate marker to evaluate the effect of taxane versus anthracycline based chemotherapies on progression free survival of locally advanced breast cancer patients. The proposed analysis strategy is broadly applicable to mediation analyses of time-to-event endpoints, is easy to apply and outperforms existing strategies in terms of precision as well as robustness against model misspecification.

  13. Ontology-Based Combinatorial Comparative Analysis of Adverse Events Associated with Killed and Live Influenza Vaccines

    PubMed Central

    Sarntivijai, Sirarat; Xiang, Zuoshuang; Shedden, Kerby A.; Markel, Howard; Omenn, Gilbert S.; Athey, Brian D.; He, Yongqun

    2012-01-01

    Vaccine adverse events (VAEs) are adverse bodily changes occurring after vaccination. Understanding the adverse event (AE) profiles is a crucial step to identify serious AEs. Two different types of seasonal influenza vaccines have been used on the market: trivalent (killed) inactivated influenza vaccine (TIV) and trivalent live attenuated influenza vaccine (LAIV). Different adverse event profiles induced by these two groups of seasonal influenza vaccines were studied based on the data drawn from the CDC Vaccine Adverse Event Report System (VAERS). Extracted from VAERS were 37,621 AE reports for four TIVs (Afluria, Fluarix, Fluvirin, and Fluzone) and 3,707 AE reports for the only LAIV (FluMist). The AE report data were analyzed by a novel combinatorial, ontology-based detection of AE method (CODAE). CODAE detects AEs using Proportional Reporting Ratio (PRR), Chi-square significance test, and base level filtration, and groups identified AEs by ontology-based hierarchical classification. In total, 48 TIV-enriched and 68 LAIV-enriched AEs were identified (PRR>2, Chi-square score >4, and the number of cases >0.2% of total reports). These AE terms were classified using the Ontology of Adverse Events (OAE), MedDRA, and SNOMED-CT. The OAE method provided better classification results than the two other methods. Thirteen out of 48 TIV-enriched AEs were related to neurological and muscular processing such as paralysis, movement disorders, and muscular weakness. In contrast, 15 out of 68 LAIV-enriched AEs were associated with inflammatory response and respiratory system disorders. There were evidences of two severe adverse events (Guillain-Barre Syndrome and paralysis) present in TIV. Although these severe adverse events were at low incidence rate, they were found to be more significantly enriched in TIV-vaccinated patients than LAIV-vaccinated patients. Therefore, our novel combinatorial bioinformatics analysis discovered that LAIV had lower chance of inducing these two

  14. Reduced-Order Modeling and Wavelet Analysis of Turbofan Engine Structural Response Due to Foreign Object Damage (FOD) Events

    NASA Technical Reports Server (NTRS)

    Turso, James; Lawrence, Charles; Litt, Jonathan

    2004-01-01

    The development of a wavelet-based feature extraction technique specifically targeting FOD-event induced vibration signal changes in gas turbine engines is described. The technique performs wavelet analysis of accelerometer signals from specified locations on the engine and is shown to be robust in the presence of significant process and sensor noise. It is envisioned that the technique will be combined with Kalman filter thermal/health parameter estimation for FOD-event detection via information fusion from these (and perhaps other) sources. Due to the lack of high-frequency FOD-event test data in the open literature, a reduced-order turbofan structural model (ROM) was synthesized from a finite element model modal analysis to support the investigation. In addition to providing test data for algorithm development, the ROM is used to determine the optimal sensor location for FOD-event detection. In the presence of significant noise, precise location of the FOD event in time was obtained using the developed wavelet-based feature.

  15. Reduced-Order Modeling and Wavelet Analysis of Turbofan Engine Structural Response Due to Foreign Object Damage "FOD" Events

    NASA Technical Reports Server (NTRS)

    Turso, James A.; Lawrence, Charles; Litt, Jonathan S.

    2007-01-01

    The development of a wavelet-based feature extraction technique specifically targeting FOD-event induced vibration signal changes in gas turbine engines is described. The technique performs wavelet analysis of accelerometer signals from specified locations on the engine and is shown to be robust in the presence of significant process and sensor noise. It is envisioned that the technique will be combined with Kalman filter thermal/ health parameter estimation for FOD-event detection via information fusion from these (and perhaps other) sources. Due to the lack of high-frequency FOD-event test data in the open literature, a reduced-order turbofan structural model (ROM) was synthesized from a finite-element model modal analysis to support the investigation. In addition to providing test data for algorithm development, the ROM is used to determine the optimal sensor location for FOD-event detection. In the presence of significant noise, precise location of the FOD event in time was obtained using the developed wavelet-based feature.

  16. Analysis of electrical penetration graph data: what to do with artificially terminated events?

    USDA-ARS?s Scientific Manuscript database

    Observing the durations of hemipteran feeding behaviors via Electrical Penetration Graph (EPG) results in situations where the duration of the last behavior is not ended by the insect under observation, but by the experimenter. These are artificially terminated events. In data analysis, one must ch...

  17. Using Real-time Event Tracking Sensitivity Analysis to Overcome Sensor Measurement Uncertainties of Geo-Information Management in Drilling Disasters

    NASA Astrophysics Data System (ADS)

    Tavakoli, S.; Poslad, S.; Fruhwirth, R.; Winter, M.

    2012-04-01

    This paper introduces an application of a novel EventTracker platform for instantaneous Sensitivity Analysis (SA) of large scale real-time geo-information. Earth disaster management systems demand high quality information to aid a quick and timely response to their evolving environments. The idea behind the proposed EventTracker platform is the assumption that modern information management systems are able to capture data in real-time and have the technological flexibility to adjust their services to work with specific sources of data/information. However, to assure this adaptation in real time, the online data should be collected, interpreted, and translated into corrective actions in a concise and timely manner. This can hardly be handled by existing sensitivity analysis methods because they rely on historical data and lazy processing algorithms. In event-driven systems, the effect of system inputs on its state is of value, as events could cause this state to change. This 'event triggering' situation underpins the logic of the proposed approach. Event tracking sensitivity analysis method describes the system variables and states as a collection of events. The higher the occurrence of an input variable during the trigger of event, the greater its potential impact will be on the final analysis of the system state. Experiments were designed to compare the proposed event tracking sensitivity analysis with existing Entropy-based sensitivity analysis methods. The results have shown a 10% improvement in a computational efficiency with no compromise for accuracy. It has also shown that the computational time to perform the sensitivity analysis is 0.5% of the time required compared to using the Entropy-based method. The proposed method has been applied to real world data in the context of preventing emerging crises at drilling rigs. One of the major purposes of such rigs is to drill boreholes to explore oil or gas reservoirs with the final scope of recovering the content

  18. Total Thrombus-formation Analysis System Predicts Periprocedural Bleeding Events in Patients With Coronary Artery Disease Undergoing Percutaneous Coronary Intervention.

    PubMed

    Oimatsu, Yu; Kaikita, Koichi; Ishii, Masanobu; Mitsuse, Tatsuro; Ito, Miwa; Arima, Yuichiro; Sueta, Daisuke; Takahashi, Aya; Iwashita, Satomi; Yamamoto, Eiichiro; Kojima, Sunao; Hokimoto, Seiji; Tsujita, Kenichi

    2017-04-24

    Periprocedural bleeding events are common after percutaneous coronary intervention. We evaluated the association of periprocedural bleeding events with thrombogenicity, which was measured quantitatively by the Total Thrombus-formation Analysis System equipped with microchips and thrombogenic surfaces (collagen, platelet chip [PL]; collagen plus tissue factor, atheroma chip [AR]). Between August 2013 and March 2016, 313 consecutive patients with coronary artery disease undergoing elective percutaneous coronary intervention were enrolled. They were divided into those with or without periprocedural bleeding events. We determined the bleeding events as composites of major bleeding events defined by the International Society on Thrombosis and Hemostasis and minor bleeding events (eg, minor hematoma, arteriovenous shunt and pseudoaneurysm). Blood samples obtained at percutaneous coronary intervention were analyzed for thrombus formation area under the curve (PL 24 -AUC 10 for PL chip; AR 10 -AUC 30 for AR chip) by the Total Thrombus-formation Analysis System and P2Y12 reaction unit by the VerifyNow system. Periprocedural bleeding events occurred in 37 patients. PL 24 -AUC 10 levels were significantly lower in patients with such events than those without ( P =0.002). Multiple logistic regression analyses showed association between low PL 24 -AUC 10 levels and periprocedural bleeding events (odds ratio, 2.71 [1.22-5.99]; P =0.01) and association between PL 24 -AUC 10 and periprocedural bleeding events in 176 patients of the femoral approach group (odds ratio, 2.88 [1.11-7.49]; P =0.03). However, PL 24 -AUC 10 levels in 127 patients of the radial approach group were not significantly different in patients with or without periprocedural bleeding events. PL 24 -AUC 10 measured by the Total Thrombus-formation Analysis System is a potentially useful predictor of periprocedural bleeding events in coronary artery disease patients undergoing elective percutaneous coronary

  19. Automatic Single Event Effects Sensitivity Analysis of a 13-Bit Successive Approximation ADC

    NASA Astrophysics Data System (ADS)

    Márquez, F.; Muñoz, F.; Palomo, F. R.; Sanz, L.; López-Morillo, E.; Aguirre, M. A.; Jiménez, A.

    2015-08-01

    This paper presents Analog Fault Tolerant University of Seville Debugging System (AFTU), a tool to evaluate the Single-Event Effect (SEE) sensitivity of analog/mixed signal microelectronic circuits at transistor level. As analog cells can behave in an unpredictable way when critical areas interact with the particle hitting, there is a need for designers to have a software tool that allows an automatic and exhaustive analysis of Single-Event Effects influence. AFTU takes the test-bench SPECTRE design, emulates radiation conditions and automatically evaluates vulnerabilities using user-defined heuristics. To illustrate the utility of the tool, the SEE sensitivity of a 13-bits Successive Approximation Analog-to-Digital Converter (ADC) has been analysed. This circuit was selected not only because it was designed for space applications, but also due to the fact that a manual SEE sensitivity analysis would be too time-consuming. After a user-defined test campaign, it was detected that some voltage transients were propagated to a node where a parasitic diode was activated, affecting the offset cancelation, and therefore the whole resolution of the ADC. A simple modification of the scheme solved the problem, as it was verified with another automatic SEE sensitivity analysis.

  20. Extreme flood event analysis in Indonesia based on rainfall intensity and recharge capacity

    NASA Astrophysics Data System (ADS)

    Narulita, Ida; Ningrum, Widya

    2018-02-01

    Indonesia is very vulnerable to flood disaster because it has high rainfall events throughout the year. Flood is categorized as the most important hazard disaster because it is causing social, economic and human losses. The purpose of this study is to analyze extreme flood event based on satellite rainfall dataset to understand the rainfall characteristic (rainfall intensity, rainfall pattern, etc.) that happened before flood disaster in the area for monsoonal, equatorial and local rainfall types. Recharge capacity will be analyzed using land cover and soil distribution. The data used in this study are CHIRPS rainfall satellite data on 0.05 ° spatial resolution and daily temporal resolution, and GSMap satellite rainfall dataset operated by JAXA on 1-hour temporal resolution and 0.1 ° spatial resolution, land use and soil distribution map for recharge capacity analysis. The rainfall characteristic before flooding, and recharge capacity analysis are expected to become the important information for flood mitigation in Indonesia.

  1. Mines Systems Safety Improvement Using an Integrated Event Tree and Fault Tree Analysis

    NASA Astrophysics Data System (ADS)

    Kumar, Ranjan; Ghosh, Achyuta Krishna

    2017-04-01

    Mines systems such as ventilation system, strata support system, flame proof safety equipment, are exposed to dynamic operational conditions such as stress, humidity, dust, temperature, etc., and safety improvement of such systems can be done preferably during planning and design stage. However, the existing safety analysis methods do not handle the accident initiation and progression of mine systems explicitly. To bridge this gap, this paper presents an integrated Event Tree (ET) and Fault Tree (FT) approach for safety analysis and improvement of mine systems design. This approach includes ET and FT modeling coupled with redundancy allocation technique. In this method, a concept of top hazard probability is introduced for identifying system failure probability and redundancy is allocated to the system either at component or system level. A case study on mine methane explosion safety with two initiating events is performed. The results demonstrate that the presented method can reveal the accident scenarios and improve the safety of complex mine systems simultaneously.

  2. Final Report for Dynamic Models for Causal Analysis of Panel Data. Approaches to the Censoring Problem in Analysis of Event Histories. Part III, Chapter 2.

    ERIC Educational Resources Information Center

    Tuma, Nancy Brandon; Hannan, Michael T.

    The document, part of a series of chapters described in SO 011 759, considers the problem of censoring in the analysis of event-histories (data on dated events, including dates of change from one qualitative state to another). Censoring refers to the lack of information on events that occur before or after the period for which data are available.…

  3. A Systematic Review and Meta-analysis of Thrombotic Events Following Endovenous Thermal Ablation of the Great Saphenous Vein.

    PubMed

    Healy, Donagh A; Kimura, Shiori; Power, David; Elhaj, Abubaker; Abdeldaim, Yasser; Cross, Keith S; McGreal, Gerard T; Burke, Paul E; Moloney, Tony; Manning, Brian J; Kavanagh, Eamon G

    2018-06-09

    A systematic review and meta-analysis was performed to determine the incidence of thrombotic events following great saphenous vein (GSV) endovenous thermal ablation (EVTA). MEDLINE, Embase and conference abstracts were searched. Eligible studies were randomised controlled trials and case series that included at least 100 patients who underwent GSV EVTA (laser ablation or radiofrequency ablation [RFA]) with duplex ultrasound (DUS) within 30 days. The systematic review focused on the complications of endovenous heat induced thrombosis (EHIT), deep venous thrombosis (DVT), and pulmonary embolism (PE). The primary outcome for the meta-analysis was deep venous thrombotic events which were defined as DVT or EHIT Type 2, 3, or 4. Secondary outcomes for the meta-analysis were EHIT Type 2, 3, or 4, DVT and PE. Subgroup analyses were performed for both the RFA and EVLA groups. Pooled proportions were calculated using random effects modelling. Fifty-two studies (16,398 patients) were included. Thrombotic complications occurred infrequently. Deep venous thrombotic events occurred in 1.7% of cases (95% CI 0.9-2.7%) (25 studies; 10,012 patients; 274 events). EHIT Type 2, 3, or 4 occurred in 1.4% of cases (95% CI 0.8-2.3%) (26 studies; 10,225 patients; 249 events). DVT occurred in 0.3% of cases (95% CI = 0.2%-0.5%) (49 studies; 15,676 patients; 48 events). PE occurred in 0.1% of cases (95% CI = 0.1-0.2%) (29 studies; 8223 patients; 3 events). Similar results were found when the RFA and EVLA groups were analysed separately. Thrombotic events occur infrequently following GSV EVTA. Given the large numbers of procedures worldwide and the potential for serious consequences, further research is needed on the burden of these complications and their management. Copyright © 2018 European Society for Vascular Surgery. Published by Elsevier B.V. All rights reserved.

  4. Incidence and risk factors of intraoperative adverse events during donor lobectomy for living-donor liver transplantation: a retrospective analysis.

    PubMed

    Araz, Coskun; Pirat, Arash; Unlukaplan, Aytekin; Torgay, Adnan; Karakayali, Hamdi; Arslan, Gulnaz; Moray, Gokhan; Haberal, Mehmet

    2012-04-01

    To evaluate the frequency, type, and predictors of intraoperative adverse events during donor hepatectomy for living-donor liver transplant. Retrospective analyses of the data from 182 consecutive living-donor liver transplant donors between May 2002 and September 2008. Ninety-one patients (50%) had at least 1 intraoperative adverse event including hypothermia (39%), hypotension (26%), need for transfusions (17%), and hypertension (7%). Patients with an adverse event were older (P = .001), had a larger graft weight (P = .023), more frequently underwent a right hepatectomy (P = .019), and were more frequently classified as American Society of Anesthesiologists physical status class II (P = .027) than those who did not have these adverse events. Logistic regression analysis revealed that only age (95% confidence interval 1.018-1.099; P = .001) was a risk factor for intraoperative adverse events. Patients with these adverse events more frequently required admission to the intensive care unit and were hospitalized longer postoperatively. A before and after analysis showed that after introduction of in-line fluid warmers and more frequent use of acute normovolemic hemodilution, the frequency of intraoperative adverse events was significantly lower (80% vs 29%; P < .001). Intraoperative adverse events such as hypothermia and hypotension were common in living-donor liver transplant donors, and older age was associated with an increased risk of these adverse events. However, the effect of these adverse events on postoperative recovery is not clear.

  5. Event selection services in ATLAS

    NASA Astrophysics Data System (ADS)

    Cranshaw, J.; Cuhadar-Donszelmann, T.; Gallas, E.; Hrivnac, J.; Kenyon, M.; McGlone, H.; Malon, D.; Mambelli, M.; Nowak, M.; Viegas, F.; Vinek, E.; Zhang, Q.

    2010-04-01

    ATLAS has developed and deployed event-level selection services based upon event metadata records ("TAGS") and supporting file and database technology. These services allow physicists to extract events that satisfy their selection predicates from any stage of data processing and use them as input to later analyses. One component of these services is a web-based Event-Level Selection Service Interface (ELSSI). ELSSI supports event selection by integrating run-level metadata, luminosity-block-level metadata (e.g., detector status and quality information), and event-by-event information (e.g., triggers passed and physics content). The list of events that survive after some selection criterion is returned in a form that can be used directly as input to local or distributed analysis; indeed, it is possible to submit a skimming job directly from the ELSSI interface using grid proxy credential delegation. ELSSI allows physicists to explore ATLAS event metadata as a means to understand, qualitatively and quantitatively, the distributional characteristics of ATLAS data. In fact, the ELSSI service provides an easy interface to see the highest missing ET events or the events with the most leptons, to count how many events passed a given set of triggers, or to find events that failed a given trigger but nonetheless look relevant to an analysis based upon the results of offline reconstruction, and more. This work provides an overview of ATLAS event-level selection services, with an emphasis upon the interactive Event-Level Selection Service Interface.

  6. Low frequency events on Montserrat

    NASA Astrophysics Data System (ADS)

    Visser, K.; Neuberg, J.

    2003-04-01

    Earthquake swarms observed on volcanoes consist generally of low frequency events. The low frequency content of these events indicates the presence of interface waves at the boundary of the magma filled conduit and the surrounding country rock. The observed seismic signal at the surface shows therefore a complicated interference pattern of waves originating at various parts of the magma filled conduit, interacting with the free surface and interfaces in the volcanic edifice. This research investigates the applicability of conventional seismic tools on these low frequency events, focusing on hypocenter location analysis using arrival times and particle motion analysis for the Soufrière Hills Volcano on Montserrat. Both single low frequency events and swarms are observed on this volcano. Synthetic low frequency events are used for comparison. Results show that reliable hypocenter locations and particle motions can only be obtained if the low frequency events are single events with an identifiable P wave onset, for example the single events preceding swarms on Montserrat or the first low frequency event of a swarm. Consecutive events of the same swarm are dominated by interface waves which are converted at the top of the conduit into weak secondary P waves and surface waves. Conventional seismic tools fail to correctly analyse these events.

  7. Superposed ruptile deformational events revealed by field and VOM structural analysis

    NASA Astrophysics Data System (ADS)

    Kumaira, Sissa; Guadagnin, Felipe; Keller Lautert, Maiara

    2017-04-01

    Virtual outcrop models (VOM) is becoming an important application in the analysis of geological structures due to the possibility of obtaining the geometry and in some cases kinematic aspects of analyzed structures in a tridimensional photorealistic space. These data are used to gain quantitative information on the deformational features which coupled with numeric models can assist in understands deformational processes. Old basement units commonly register superposed deformational events either ductile or ruptile along its evolution. The Porongos Belt, located at southern Brazil, have a complex deformational history registering at least five ductile and ruptile deformational events. In this study, we presents a structural analysis of a quarry in the Porongos Belt, coupling field and VOM structural information to understand process involved in the last two deformational events. Field information was acquired using traditional structural methods for analysis of ruptile structures, such as the descriptions, drawings, acquisition of orientation vectors and kinematic analysis. VOM was created from the image-based modeling method through photogrammetric data acquisition and orthorectification. Photogrammetric data acquisition was acquired using Sony a3500 camera and a total of 128 photographs were taken from ca. 10-20 m from the outcrop in different orientations. Thirty two control point coordinates were acquired using a combination of RTK dGPS surveying and total station work, providing a precision of few millimeters for x, y and z. Photographs were imported into the Photo Scan software to create a 3D dense point cloud from structure from-motion algorithm, which were triangulated and textured to generate the VOM. VOM was georreferenced (oriented and scaled) using the ground control points, and later analyzed in OpenPlot software to extract structural information. Data was imported in Wintensor software to obtain tensor orientations, and Move software to process and

  8. Climate Central World Weather Attribution (WWA) project: Real-time extreme weather event attribution analysis

    NASA Astrophysics Data System (ADS)

    Haustein, Karsten; Otto, Friederike; Uhe, Peter; Allen, Myles; Cullen, Heidi

    2015-04-01

    Extreme weather detection and attribution analysis has emerged as a core theme in climate science over the last decade or so. By using a combination of observational data and climate models it is possible to identify the role of climate change in certain types of extreme weather events such as sea level rise and its contribution to storm surges, extreme heat events and droughts or heavy rainfall and flood events. These analyses are usually carried out after an extreme event has occurred when reanalysis and observational data become available. The Climate Central WWA project will exploit the increasing forecast skill of seasonal forecast prediction systems such as the UK MetOffice GloSea5 (Global seasonal forecasting system) ensemble forecasting method. This way, the current weather can be fed into climate models to simulate large ensembles of possible weather scenarios before an event has fully emerged yet. This effort runs along parallel and intersecting tracks of science and communications that involve research, message development and testing, staged socialization of attribution science with key audiences, and dissemination. The method we employ uses a very large ensemble of simulations of regional climate models to run two different analyses: one to represent the current climate as it was observed, and one to represent the same events in the world that might have been without human-induced climate change. For the weather "as observed" experiment, the atmospheric model uses observed sea surface temperature (SST) data from GloSea5 (currently) and present-day atmospheric gas concentrations to simulate weather events that are possible given the observed climate conditions. The weather in the "world that might have been" experiments is obtained by removing the anthropogenic forcing from the observed SSTs, thereby simulating a counterfactual world without human activity. The anthropogenic forcing is obtained by comparing the CMIP5 historical and natural simulations

  9. Reinvestigation and analysis a landslide dam event in 2012 using UAV

    NASA Astrophysics Data System (ADS)

    Wang, Kuo-Lung; Huang, Zji-Jie; Lin, Jun-Tin

    2015-04-01

    Geological condition of Taiwan is fracture with locating on Pacific Rim seismic area. Typhoons usually attack during summer and steep mountains are highly weathered, which induces landslide in mountain area. The situation happens more frequently recent years due to weather change effect. Most landslides are very far away from residence area. Field investigation is time consuming, high budget, limited data collected and dangerous. Investigation with satellite images has disadvantages such as less of the actual situation and poor resolution. Thus the possibility for slope investigation with UAV will be proposed and discussed in this research. Hazard investigation and monitoring is adopted UAV in recent years. UAV has advantages such as light weight, small volume, high mobility, safe, easy maintenance and low cost. Investigation can be executed in high risk area. Use the mature aero photogrammetry , combines aero photos with control point. Digital surface model (DSM) and Ortho photos can be produced with control points aligned. The resolution can be less than 5cm thus can be used as temporal creeping monitoring before landslide happens. A large landslide site at 75k of road No. 14 was investigated in this research. Landslide happened in June, 2012 with heavy rainfall and landslide dam was formed quickly after that. Analysis of this landslide failure and mechanism were discussed in this research using DEMs produced prior this event with aero photos and after this event with UAV. Residual slope stability analysis is thus carried out after strength parameters obtain from analysis described above. Thus advice for following potential landslide conditions can be provided.

  10. Psychiatric adverse events during treatment with brodalumab: Analysis of psoriasis clinical trials.

    PubMed

    Lebwohl, Mark G; Papp, Kim A; Marangell, Lauren B; Koo, John; Blauvelt, Andrew; Gooderham, Melinda; Wu, Jashin J; Rastogi, Shipra; Harris, Susan; Pillai, Radhakrishnan; Israel, Robert J

    2018-01-01

    Individuals with psoriasis are at increased risk for psychiatric comorbidities, including suicidal ideation and behavior (SIB). To distinguish between the underlying risk and potential for treatment-induced psychiatric adverse events in patients with psoriasis being treated with brodalumab, a fully human anti-interleukin 17 receptor A monoclonal antibody. Data were evaluated from a placebo-controlled, phase 2 clinical trial; the open-label, long-term extension of the phase 2 clinical trial; and three phase 3, randomized, double-blind, controlled clinical trials (AMAGINE-1, AMAGINE-2, and AMAGINE-3) and their open-label, long-term extensions of patients with moderate-to-severe psoriasis. The analysis included 4464 patients with 9161.8 patient-years of brodalumab exposure. The follow-up time-adjusted incidence rates of SIB events were comparable between the brodalumab and ustekinumab groups throughout the 52-week controlled phases (0.20 vs 0.60 per 100 patient-years). In the brodalumab group, 4 completed suicides were reported, 1 of which was later adjudicated as indeterminate; all patients had underlying psychiatric disorders or stressors. There was no comparator arm past week 52. Controlled study periods were not powered to detect differences in rare events such as suicide. Comparison with controls and the timing of events do not indicate a causal relationship between SIB and brodalumab treatment. Copyright © 2017 American Academy of Dermatology, Inc. Published by Elsevier Inc. All rights reserved.

  11. Grieving experiences amongst adolescents orphaned by AIDS: Analysis from event history calendars.

    PubMed

    Thupayagale-Tshweneagae, Gloria

    2012-09-07

    Mental health is an essential component of adolescent health and wellbeing. Mental health practitioners assess adolescents' mental health status to identify possible issues that may lead to mental health problems. However, very few of the tools used to assess the mental health status of adolescents include assessment for grieving and coping patterns. The current tools used for assessing an individual's mental health are lengthy and not comprehensive. The purpose of this study was to assess grieving patterns of adolescents orphaned by AIDS and to appraise the usefulness of an event history calendar as an assessment tool for identifying grieving experiences, in order to guide and support these adolescents through the grieving process. One hundred and two adolescents aged 14-18 years, who had been orphaned by AIDS, completed an event history calendar, reviewed it with the researcher and reported their perceptions of it. Thematic analysis of the event history calendar content revealed that it is an effective, time-efficient, adolescent-friendly tool that facilitated identification and discussion of the orphaned adolescents' grieving patterns. Crying, isolation, silence and violent outbursts were the main grieving patterns reported by adolescents orphaned by AIDS. The researcher recommends use of the event history calendar for identification of orphaned adolescents' grieving experiences. Early identification would enable mental health practitioners to support them in order to prevent the occurrence of mental illness due to maladaptive grieving.

  12. Accounting for unintended binding events in the analysis of quartz crystal microbalance kinetic data.

    PubMed

    Heller, Gabriella T; Zwang, Theodore J; Sarapata, Elizabeth A; Haber, Michael A; Sazinsky, Matthew H; Radunskaya, Ami E; Johal, Malkiat S

    2014-05-01

    Previous methods for analyzing protein-ligand binding events using the quartz crystal microbalance with dissipation monitoring (QCM-D) fail to account for unintended binding that inevitably occurs during surface measurements and obscure kinetic information. In this article, we present a system of differential equations that accounts for both reversible and irreversible unintended interactions. This model is tested on three protein-ligand systems, each of which has different features, to establish the feasibility of using the QCM-D for protein binding analysis. Based on this analysis, we were able to obtain kinetic information for the intended interaction that is consistent with those obtained in literature via bulk-phase methods. In the appendix, we include a method for decoupling these from the intended binding events and extracting relevant affinity information. Copyright © 2014 Elsevier B.V. All rights reserved.

  13. Projecting Event-Based Analysis Dates in Clinical Trials: An Illustration Based on the International Duration Evaluation of Adjuvant Chemotherapy (IDEA) Collaboration. Projecting analysis dates for the IDEA collaboration.

    PubMed

    Renfro, Lindsay A; Grothey, Axel M; Paul, James; Floriani, Irene; Bonnetain, Franck; Niedzwiecki, Donna; Yamanaka, Takeharu; Souglakos, Ioannis; Yothers, Greg; Sargent, Daniel J

    2014-12-01

    Clinical trials are expensive and lengthy, where success of a given trial depends on observing a prospectively defined number of patient events required to answer the clinical question. The point at which this analysis time occurs depends on both patient accrual and primary event rates, which typically vary throughout the trial's duration. We demonstrate real-time analysis date projections using data from a collection of six clinical trials that are part of the IDEA collaboration, an international preplanned pooling of data from six trials testing the duration of adjuvant chemotherapy in stage III colon cancer, and we additionally consider the hypothetical impact of one trial's early termination of follow-up. In the absence of outcome data from IDEA, monthly accrual rates for each of the six IDEA trials were used to project subsequent trial-specific accrual, while historical data from similar Adjuvant Colon Cancer Endpoints (ACCENT) Group trials were used to construct a parametric model for IDEA's primary endpoint, disease-free survival, under the same treatment regimen. With this information and using the planned total accrual from each IDEA trial protocol, individual patient accrual and event dates were simulated and the overall IDEA interim and final analysis times projected. Projections were then compared with actual (previously undisclosed) trial-specific event totals at a recent census time for validation. The change in projected final analysis date assuming early termination of follow-up for one IDEA trial was also calculated. Trial-specific predicted event totals were close to the actual number of events per trial for the recent census date at which the number of events per trial was known, with the overall IDEA projected number of events only off by eight patients. Potential early termination of follow-up by one IDEA trial was estimated to postpone the overall IDEA final analysis date by 9 months. Real-time projection of the final analysis time during a

  14. Projecting Event-Based Analysis Dates in Clinical Trials: An Illustration Based on the International Duration Evaluation of Adjuvant Chemotherapy (IDEA) Collaboration. Projecting analysis dates for the IDEA collaboration

    PubMed Central

    Renfro, Lindsay A.; Grothey, Axel M.; Paul, James; Floriani, Irene; Bonnetain, Franck; Niedzwiecki, Donna; Yamanaka, Takeharu; Souglakos, Ioannis; Yothers, Greg; Sargent, Daniel J.

    2015-01-01

    Purpose Clinical trials are expensive and lengthy, where success of a given trial depends on observing a prospectively defined number of patient events required to answer the clinical question. The point at which this analysis time occurs depends on both patient accrual and primary event rates, which typically vary throughout the trial's duration. We demonstrate real-time analysis date projections using data from a collection of six clinical trials that are part of the IDEA collaboration, an international preplanned pooling of data from six trials testing the duration of adjuvant chemotherapy in stage III colon cancer, and we additionally consider the hypothetical impact of one trial's early termination of follow-up. Patients and Methods In the absence of outcome data from IDEA, monthly accrual rates for each of the six IDEA trials were used to project subsequent trial-specific accrual, while historical data from similar Adjuvant Colon Cancer Endpoints (ACCENT) Group trials were used to construct a parametric model for IDEA's primary endpoint, disease-free survival, under the same treatment regimen. With this information and using the planned total accrual from each IDEA trial protocol, individual patient accrual and event dates were simulated and the overall IDEA interim and final analysis times projected. Projections were then compared with actual (previously undisclosed) trial-specific event totals at a recent census time for validation. The change in projected final analysis date assuming early termination of follow-up for one IDEA trial was also calculated. Results Trial-specific predicted event totals were close to the actual number of events per trial for the recent census date at which the number of events per trial was known, with the overall IDEA projected number of events only off by eight patients. Potential early termination of follow-up by one IDEA trial was estimated to postpone the overall IDEA final analysis date by 9 months. Conclusions Real

  15. Detection and analysis of high-temperature events in the BIRD mission

    NASA Astrophysics Data System (ADS)

    Zhukov, Boris; Briess, Klaus; Lorenz, Eckehard; Oertel, Dieter; Skrbek, Wolfgang

    2005-01-01

    The primary mission objective of a new small Bi-spectral InfraRed Detection (BIRD) satellite is detection and quantitative analysis of high-temperature events like fires and volcanoes. An absence of saturation in the BIRD infrared channels makes it possible to improve false alarm rejection as well as to retrieve quantitative characteristics of hot targets, including their effective fire temperature, area and the radiative energy release. Examples are given of detection and analysis of wild and coal seam fires, of volcanic activity as well as of oil fires in Iraq. The smallest fires detected by BIRD, which were verified on ground, had an area of 12m2 at daytime and 4m2 at night.

  16. Inner core rotation from event-pair analysis

    NASA Astrophysics Data System (ADS)

    Song, Xiaodong; Poupinet, Georges

    2007-09-01

    The last decade has witnessed an animated debate on whether the inner core rotation is a fact or an artifact. Here we examine the temporal change of inner core waves using a technique that compares differential travel times at the same station but between two events. The method does not require precise knowledge of earthquake locations and earth models. The pairing of the events creates a large data set for the application of statistical tools. Using measurements from 87 events in the South Sandwich Islands recorded at College, Alaska station, we conclude the temporal change is robust. The estimates of the temporal change range from about 0.07 to 0.10 s/decade over the past 50 yr. If we used only pairs with small inter-event distances, which reduce the influence of mantle heterogeneity, the rates range from 0.084 to 0.098 s/decade, nearly identical to the rate inferred by Zhang et al. [Zhang, J., Song, X.D., Li, Y.C., Richards, P.G., Sun, X.L., Waldhauser, F., Inner core differential motion confirmed by earthquake waveform doublets, Science 309 (5739) (2005) 1357-1360.] from waveform doublets. The rate of the DF change seems to change with time, which may be explained by lateral variation of the inner core structure or the change in rotation rate on decadal time scale.

  17. Early blood pressure lowering treatment in acute stroke. Ordinal analysis of vascular events in the Scandinavian Candesartan Acute Stroke Trial (SCAST).

    PubMed

    Jusufovic, Mirza; Sandset, Else Charlotte; Bath, Philip M; Berge, Eivind

    2016-08-01

    Early blood pressure-lowering treatment appears to be beneficial in patients with acute intracerebral haemorrhage and potentially in ischaemic stroke. We used a new method for analysis of vascular events in the Scandinavian Candesartan Acute Stroke Trial to see if the effect was dependent on the timing of treatment. Scandinavian Candesartan Acute Stroke Trial was a randomized controlled and placebo-controlled trial of candesartan within 30 h of ischaemic or haemorrhagic stroke. Of 2029 patients, 231 (11.4%) had a vascular event (vascular death, nonfatal stroke or nonfatal myocardial infarction) during the first 6 months. The modified Rankin Scale (mRS) score following a vascular event was used to categorize vascular events in order of severity: no event (n = 1798), minor (mRS 0-2, n = 59), moderately severe (mRS 3-4, n = 57) and major event (mRS 5-6, n = 115). We used ordinal logistic regression for analysis and adjusted for predefined prognostic variables. Candesartan had no overall effect on vascular events (adjusted common odds ratio 1.11, 95% confidence interval 0.84-1.47, P = 0.48), and the effects were the same in ischaemic and haemorrhagic stroke. Among the patients treated within 6 h, the adjusted common odds ratio for vascular events was 0.37, 95% confidence interval 0.16-0.84, P = 0.02, and there was no heterogeneity of effect between ischaemic and haemorrhagic strokes. Ordinal analysis of vascular events showed no overall effect of candesartan in the subacute phase of stroke. The effect of treatment given within 6 h of stroke onset appears promising, and will be addressed in ongoing trials. Ordinal analysis of vascular events is feasible and can be used in future trials.

  18. Making adjustments to event annotations for improved biological event extraction.

    PubMed

    Baek, Seung-Cheol; Park, Jong C

    2016-09-16

    Current state-of-the-art approaches to biological event extraction train statistical models in a supervised manner on corpora annotated with event triggers and event-argument relations. Inspecting such corpora, we observe that there is ambiguity in the span of event triggers (e.g., "transcriptional activity" vs. 'transcriptional'), leading to inconsistencies across event trigger annotations. Such inconsistencies make it quite likely that similar phrases are annotated with different spans of event triggers, suggesting the possibility that a statistical learning algorithm misses an opportunity for generalizing from such event triggers. We anticipate that adjustments to the span of event triggers to reduce these inconsistencies would meaningfully improve the present performance of event extraction systems. In this study, we look into this possibility with the corpora provided by the 2009 BioNLP shared task as a proof of concept. We propose an Informed Expectation-Maximization (EM) algorithm, which trains models using the EM algorithm with a posterior regularization technique, which consults the gold-standard event trigger annotations in a form of constraints. We further propose four constraints on the possible event trigger annotations to be explored by the EM algorithm. The algorithm is shown to outperform the state-of-the-art algorithm on the development corpus in a statistically significant manner and on the test corpus by a narrow margin. The analysis of the annotations generated by the algorithm shows that there are various types of ambiguity in event annotations, even though they could be small in number.

  19. Cardiovascular Events Following Smoke-Free Legislations: An Updated Systematic Review and Meta-Analysis

    PubMed Central

    Jones, Miranda R.; Barnoya, Joaquin; Stranges, Saverio; Losonczy, Lia; Navas-Acien, Ana

    2014-01-01

    Background Legislations banning smoking in indoor public places and workplaces are being implemented worldwide to protect the population from secondhand smoke exposure. Several studies have reported reductions in hospitalizations for acute coronary events following the enactment of smoke-free laws. Objective We set out to conduct a systematic review and meta-analysis of epidemiologic studies examining how legislations that ban smoking in indoor public places impact the risk of acute coronary events. Methods We searched MEDLINE, EMBASE, and relevant bibliographies including previous systematic reviews for studies that evaluated changes in acute coronary events, following implementation of smoke-free legislations. Studies were identified through December 2013. We pooled relative risk (RR) estimates for acute coronary events comparing post- vs. pre-legislation using inverse-variance weighted random-effects models. Results Thirty-one studies providing estimates for 47 locations were included. The legislations were implemented between 1991 and 2010. Following the enactment of smoke-free legislations, there was a 12 % reduction in hospitalizations for acute coronary events (pooled RR: 0.88, 95 % CI: 0.85–0.90). Reductions were 14 % in locations that implemented comprehensive legislations compared to an 8 % reduction in locations that only had partial restrictions. In locations with reductions in smoking prevalence post-legislation above the mean (2.1 % reduction) there was a 14 % reduction in events compared to 10 % in locations below the mean. The RRs for acute coronary events associated with enacting smoke-free legislation were 0.87 vs. 0.89 in locations with smoking prevalence pre-legislation above and below the mean (23.1 %), and 0.87 vs. 0.89 in studies from the Americas vs. other regions. Conclusion The implementation of smoke-free legislations was related to reductions in acute coronary event hospitalizations in most populations evaluated. Benefits are greater

  20. A lengthy look at the daily grind: time series analysis of events, mood, stress, and satisfaction.

    PubMed

    Fuller, Julie A; Stanton, Jeffrey M; Fisher, Gwenith G; Spitzmuller, Christiane; Russell, Steven S; Smith, Patricia C

    2003-12-01

    The present study investigated processes by which job stress and satisfaction unfold over time by examining the relations between daily stressful events, mood, and these variables. Using a Web-based daily survey of stressor events, perceived strain, mood, and job satisfaction completed by 14 university workers, 1,060 occasions of data were collected. Transfer function analysis, a multivariate version of time series analysis, was used to examine the data for relationships among the measured variables after factoring out the contaminating influences of serial dependency. Results revealed a contrast effect in which a stressful event associated positively with higher strain on the same day and associated negatively with strain on the following day. Perceived strain increased over the course of a semester for a majority of participants, suggesting that effects of stress build over time. Finally, the data were consistent with the notion that job satisfaction is a distal outcome that is mediated by perceived strain. ((c) 2003 APA, all rights reserved)

  1. Interaction between the serotonin transporter gene (5-HTTLPR), stressful life events, and risk of depression: a meta-analysis.

    PubMed

    Risch, Neil; Herrell, Richard; Lehner, Thomas; Liang, Kung-Yee; Eaves, Lindon; Hoh, Josephine; Griem, Andrea; Kovacs, Maria; Ott, Jurg; Merikangas, Kathleen Ries

    2009-06-17

    Substantial resources are being devoted to identify candidate genes for complex mental and behavioral disorders through inclusion of environmental exposures following the report of an interaction between the serotonin transporter linked polymorphic region (5-HTTLPR) and stressful life events on an increased risk of major depression. To conduct a meta-analysis of the interaction between the serotonin transporter gene and stressful life events on depression using both published data and individual-level original data. Search of PubMed, EMBASE, and PsycINFO databases through March 2009 yielded 26 studies of which 14 met criteria for the meta-analysis. Criteria for studies for the meta-analyses included published data on the association between 5-HTTLPR genotype (SS, SL, or LL), number of stressful life events (0, 1, 2, > or = 3) or equivalent, and a categorical measure of depression defined by the Diagnostic and Statistical Manual of Mental Disorders (Fourth Edition) or the International Statistical Classification of Diseases, 10th Revision (ICD-10) or use of a cut point to define depression from standardized rating scales. To maximize our ability to use a common framework for variable definition, we also requested original data from all studies published prior to 2008 that met inclusion criteria. Of the 14 studies included in the meta-analysis, 10 were also included in a second sex-specific meta-analysis of original individual-level data. Logistic regression was used to estimate the effects of the number of short alleles at 5-HTTLPR, the number of stressful life events, and their interaction on depression. Odds ratios (ORs) and 95% confidence intervals (CIs) were calculated separately for each study and then weighted averages of the individual estimates were obtained using random-effects meta-analysis. Both sex-combined and sex-specific meta-analyses were conducted. Of a total of 14,250 participants, 1769 were classified as having depression; 12,481 as not having

  2. Multivariate hydrological frequency analysis for extreme events using Archimedean copula. Case study: Lower Tunjuelo River basin (Colombia)

    NASA Astrophysics Data System (ADS)

    Gómez, Wilmar

    2017-04-01

    By analyzing the spatial and temporal variability of extreme precipitation events we can prevent or reduce the threat and risk. Many water resources projects require joint probability distributions of random variables such as precipitation intensity and duration, which can not be independent with each other. The problem of defining a probability model for observations of several dependent variables is greatly simplified by the joint distribution in terms of their marginal by taking copulas. This document presents a general framework set frequency analysis bivariate and multivariate using Archimedean copulas for extreme events of hydroclimatological nature such as severe storms. This analysis was conducted in the lower Tunjuelo River basin in Colombia for precipitation events. The results obtained show that for a joint study of the intensity-duration-frequency, IDF curves can be obtained through copulas and thus establish more accurate and reliable information from design storms and associated risks. It shows how the use of copulas greatly simplifies the study of multivariate distributions that introduce the concept of joint return period used to represent the needs of hydrological designs properly in frequency analysis.

  3. Toward Joint Hypothesis-Tests Seismic Event Screening Analysis: Ms|mb and Event Depth

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Anderson, Dale; Selby, Neil

    2012-08-14

    Well established theory can be used to combine single-phenomenology hypothesis tests into a multi-phenomenology event screening hypothesis test (Fisher's and Tippett's tests). Commonly used standard error in Ms:mb event screening hypothesis test is not fully consistent with physical basis. Improved standard error - Better agreement with physical basis, and correctly partitions error to include Model Error as a component of variance, correctly reduces station noise variance through network averaging. For 2009 DPRK test - Commonly used standard error 'rejects' H0 even with better scaling slope ({beta} = 1, Selby et al.), improved standard error 'fails to rejects' H0.

  4. Analysis and modeling of a hail event consequences on a building portfolio

    NASA Astrophysics Data System (ADS)

    Nicolet, Pierrick; Voumard, Jérémie; Choffet, Marc; Demierre, Jonathan; Imhof, Markus; Jaboyedoff, Michel

    2014-05-01

    North-West Switzerland has been affected by a severe Hail Storm in July 2011, which was especially intense in the Canton of Aargau. The damage cost of this event is around EUR 105 Million only for the Canton of Aargau, which corresponds to half of the mean annual consolidated damage cost of the last 20 years for the 19 Cantons (over 26) with a public insurance. The aim of this project is to benefit from the collected insurance data to better understand and estimate the risk of such event. In a first step, a simple hail event simulator, which has been developed for a previous hail episode, is modified. The geometric properties of the storm is derived from the maximum intensity radar image by means of a set of 2D Gaussians instead of using 1D Gaussians on profiles, as it was the case in the previous version. The tool is then tested on this new event in order to establish its ability to give a fast damage estimation based on the radar image and buildings value and location. The geometrical properties are used in a further step to generate random outcomes with similar characteristics, which are combined with a vulnerability curve and an event frequency to estimate the risk. The vulnerability curve comes from a 2009 event and is improved with data from this event, whereas the frequency for the Canton is estimated from insurance records. In addition to this regional risk analysis, this contribution aims at studying the relation of the buildings orientation with the damage rate. Indeed, it is expected that the orientation of the roof influences the aging of the material by controlling the frequency and amplitude of thaw-freeze cycles, changing then the vulnerability over time. This part is established by calculating the hours of sunshine, which are used to derive the material temperatures. This information is then compared with insurance claims. A last part proposes a model to study the hail impact on a building, by modeling the different equipment on each facade of the

  5. Microbial-based evaluation of foaming events in full-scale wastewater treatment plants by microscopy survey and quantitative image analysis.

    PubMed

    Leal, Cristiano; Amaral, António Luís; Costa, Maria de Lourdes

    2016-08-01

    Activated sludge systems are prone to be affected by foaming occurrences causing the sludge to rise in the reactor and affecting the wastewater treatment plant (WWTP) performance. Nonetheless, there is currently a knowledge gap hindering the development of foaming events prediction tools that may be fulfilled by the quantitative monitoring of AS systems biota and sludge characteristics. As such, the present study focuses on the assessment of foaming events in full-scale WWTPs, by quantitative protozoa, metazoa, filamentous bacteria, and sludge characteristics analysis, further used to enlighten the inner relationships between these parameters. In the current study, a conventional activated sludge system (CAS) and an oxidation ditch (OD) were surveyed throughout a period of 2 and 3 months, respectively, regarding their biota and sludge characteristics. The biota community was monitored by microscopic observation, and a new filamentous bacteria index was developed to quantify their occurrence. Sludge characteristics (aggregated and filamentous biomass contents and aggregate size) were determined by quantitative image analysis (QIA). The obtained data was then processed by principal components analysis (PCA), cross-correlation analysis, and decision trees to assess the foaming occurrences, and enlighten the inner relationships. It was found that such events were best assessed by the combined use of the relative abundance of testate amoeba and nocardioform filamentous index, presenting a 92.9 % success rate for overall foaming events, and 87.5 and 100 %, respectively, for persistent and mild events.

  6. Two damaging hydrogeological events in Calabria, September 2000 and November 2015. Comparative analysis of causes and effects

    NASA Astrophysics Data System (ADS)

    Petrucci, Olga; Caloiero, Tommaso; Aurora Pasqua, Angela

    2016-04-01

    Each year, especially during winter season, some episode of intense rain affects Calabria, the southernmost Italian peninsular region, triggering flash floods and mass movements that cause damage and fatalities. This work presents a comparative analysis between two events that affected the southeast sector of the region, in 2000 and 2014, respectively. The event occurred between 9th and 10th of September 2000 is known in Italy as Soverato event, after the name of the municipality where it reached the highest damage severity. In the Soverato area, more than 200 mm of rain that fell in 24 hours caused a disastrous flood that swept away a campsite at about 4 a.m., killing 13 people and hurting 45. Besides, the rain affected a larger area, causing damage in 89 (out of 409) municipalities of the region. Flooding was the most common process, which damaged housing and trading. Landslide mostly affected the road network, housing and cultivations. The most recent event affected the same regional sector between 30th October and 2nd November 2015. The daily rain recorded at some of the rain gauges of the area almost reached 400 mm. Out of the 409 municipalities of Calabria, 109 suffered damage. The most frequent types of processes were both flash floods and landslides. The most heavily damaged element was the road network: the representative picture of the event is a railway bridge destroyed by the river flow. Housing was damaged too, and 486 people were temporarily evacuated from home. The event also caused a victim killed by a flood. The event-centred study approach aims to highlight differences and similarities in both the causes and the effects of the two events that occurred at a temporal distance of 14 years. The comparative analysis focus on three main aspects: the intensity of triggering rain, the modifications of urbanised areas, and the evolution of emergency management. The comparative analysis of rain is made by comparing the return period of both daily and

  7. Do climate extreme events foster violent civil conflicts? A coincidence analysis

    NASA Astrophysics Data System (ADS)

    Schleussner, Carl-Friedrich; Donges, Jonathan F.; Donner, Reik V.

    2014-05-01

    Civil conflicts promoted by adverse environmental conditions represent one of the most important potential feedbacks in the global socio-environmental nexus. While the role of climate extremes as a triggering factor is often discussed, no consensus is yet reached about the cause-and-effect relation in the observed data record. Here we present results of a rigorous statistical coincidence analysis based on the Munich Re Inc. extreme events database and the Uppsala conflict data program. We report evidence for statistically significant synchronicity between climate extremes with high economic impact and violent conflicts for various regions, although no coherent global signal emerges from our analysis. Our results indicate the importance of regional vulnerability and might aid to identify hot-spot regions for potential climate-triggered violent social conflicts.

  8. OAE: The Ontology of Adverse Events.

    PubMed

    He, Yongqun; Sarntivijai, Sirarat; Lin, Yu; Xiang, Zuoshuang; Guo, Abra; Zhang, Shelley; Jagannathan, Desikan; Toldo, Luca; Tao, Cui; Smith, Barry

    2014-01-01

    A medical intervention is a medical procedure or application intended to relieve or prevent illness or injury. Examples of medical interventions include vaccination and drug administration. After a medical intervention, adverse events (AEs) may occur which lie outside the intended consequences of the intervention. The representation and analysis of AEs are critical to the improvement of public health. The Ontology of Adverse Events (OAE), previously named Adverse Event Ontology (AEO), is a community-driven ontology developed to standardize and integrate data relating to AEs arising subsequent to medical interventions, as well as to support computer-assisted reasoning. OAE has over 3,000 terms with unique identifiers, including terms imported from existing ontologies and more than 1,800 OAE-specific terms. In OAE, the term 'adverse event' denotes a pathological bodily process in a patient that occurs after a medical intervention. Causal adverse events are defined by OAE as those events that are causal consequences of a medical intervention. OAE represents various adverse events based on patient anatomic regions and clinical outcomes, including symptoms, signs, and abnormal processes. OAE has been used in the analysis of several different sorts of vaccine and drug adverse event data. For example, using the data extracted from the Vaccine Adverse Event Reporting System (VAERS), OAE was used to analyse vaccine adverse events associated with the administrations of different types of influenza vaccines. OAE has also been used to represent and classify the vaccine adverse events cited in package inserts of FDA-licensed human vaccines in the USA. OAE is a biomedical ontology that logically defines and classifies various adverse events occurring after medical interventions. OAE has successfully been applied in several adverse event studies. The OAE ontological framework provides a platform for systematic representation and analysis of adverse events and of the factors (e

  9. Forensic hydro-meteorological analysis of an extreme flash flood: The 2016-05-29 event in Braunsbach, SW Germany.

    PubMed

    Bronstert, Axel; Agarwal, Ankit; Boessenkool, Berry; Crisologo, Irene; Fischer, Madlen; Heistermann, Maik; Köhn-Reich, Lisei; López-Tarazón, José Andrés; Moran, Thomas; Ozturk, Ugur; Reinhardt-Imjela, Christian; Wendi, Dadiyorto

    2018-07-15

    The flash-flood in Braunsbach in the north-eastern part of Baden-Wuerttemberg/Germany was a particularly strong and concise event which took place during the floods in southern Germany at the end of May/early June 2016. This article presents a detailed analysis of the hydro-meteorological forcing and the hydrological consequences of this event. A specific approach, the "forensic hydrological analysis" was followed in order to include and combine retrospectively a variety of data from different disciplines. Such an approach investigates the origins, mechanisms and course of such natural events if possible in a "near real time" mode, in order to follow the most recent traces of the event. The results show that it was a very rare rainfall event with extreme intensities which, in combination with catchment properties, led to extreme runoff plus severe geomorphological hazards, i.e. great debris flows, which together resulted in immense damage in this small rural town Braunsbach. It was definitely a record-breaking event and greatly exceeded existing design guidelines for extreme flood discharge for this region, i.e. by a factor of about 10. Being such a rare or even unique event, it is not reliably feasible to put it into a crisp probabilistic context. However, one can conclude that a return period clearly above 100years can be assigned for all event components: rainfall, peak discharge and sediment transport. Due to the complex and interacting processes, no single flood cause or reason for the very high damage can be identified, since only the interplay and the cascading characteristics of those led to such an event. The roles of different human activities on the origin and/or intensification of such an extreme event are finally discussed. Copyright © 2018. Published by Elsevier B.V.

  10. A meta-analysis of the risk of total cardiovascular events of isosmolar iodixanol compared with low-osmolar contrast media.

    PubMed

    Zhang, Bu-Chun; Wu, Qiang; Wang, Cheng; Li, Dong-Ye; Wang, Zhi-Rong

    2014-04-01

    The iso-osmolar contrast agent iodixanol may be associated with a lower incidence of cardiac events than low-osmolar contrast media (LOCM), but previous trials have yielded mixed results. To compare the risk of total cardiovascular events of the iso-osmolar contrast medium, iodixanol, to LOCM. Medical literature databases were searched to identify comparisons between iodixanol and LOCM with cardiovascular events as a primary endpoint. A random-effects model was used to obtain pooled odds ratio (OR) for within-hospital and 30-day events. A total of 2 prospective cross-sectional studies and 11 randomized controlled trials (RCTs) (covering 6859 subjects) met our criteria. There was no significant difference in the incidence of within-hospital and 30-day cardiovascular events when iodixanol was compared with LOCM, with pooled OR of 0.72 (95%CI 0.49-1.06, p=0.09) and 1.19 (95%CI 0.70-2.02, p=0.53), respectively. Subgroup analysis showed no relative difference when iodixanol was compared with ioxaglate (OR=0.92, 95%CI 0.50-1.70, p=0.80) and iohexol (OR=0.75, 95%CI 0.48-1.17, p=0.21). However, a reduction in the within-hospital cardiovascular events was observed when iodixanol was compared with LOCM in the RCT subgroup (OR=0.65, 95%CI 0.44-0.96, p=0.03). Sensitivity analyses revealed that three studies had a strong impact on the association of within-hospital cardiovascular events between iodixanol and LOCM. Meta-regression analysis failed to account for heterogeneity. No publication bias was detected. This meta-analysis demonstrates that there is no conclusive evidence that iodixanol is superior to LOCM overall with regard to fewer cardiovascular events. Copyright © 2014. Published by Elsevier Ltd.

  11. Analysis of Cumulus Solar Irradiance Reflectance (CSIR) Events

    NASA Technical Reports Server (NTRS)

    Laird, John L.; Harshvardham

    1996-01-01

    Clouds are extremely important with regard to the transfer of solar radiation at the earth's surface. This study investigates Cumulus Solar Irradiance Reflection (CSIR) using ground-based pyranometers. CSIR events are short-term increases in solar radiation observed at the surface as a result of reflection off the sides of convective clouds. When sun-cloud observer geometry is favorable, these occurrences produce characteristic spikes in the pyranometer traces and solar irradiance values may exceed expected clear-sky values. Ultraviolet CSIR events were investigated during the summer of 1995 using Yankee Environmental Systems UVA-1 and UVB-1 pyranometers. Observed data were compared to clear-sky curves which were generated using a third degree polynomial best-fit line technique. Periods during which the observed data exceeded this clear-sky curve were identified as CSIR events. The magnitude of a CSIR event was determined by two different quantitative calculations. The MAC (magnitude above clear-sky) is an absolute measure of the difference between the observed and clear-sky irradiances. Maximum MAC values of 3.4 Wm(exp -2) and 0.069 Wm(exp -2) were observed at the UV-A and UV-B wavelengths, respectively. The second calculation determined the percentage above clear-sky (PAC) which indicated the relative magnitude of a CSIR event. Maximum UV-A and UV-B PAC magnitudes of 10.1% and 7.8%, respectively, were observed during the study. Also of interest was the duration of the CSIR events which is a function of sun-cloud-sensor geometry and the speed of cloud propagation over the measuring site. In both the UV-A and UV-B wavelengths, significant CSIR durations of up to 30 minutes were observed.

  12. Statistical analysis of hydrodynamic cavitation events

    NASA Astrophysics Data System (ADS)

    Gimenez, G.; Sommer, R.

    1980-10-01

    The frequency (number of events per unit time) of pressure pulses produced by hydrodynamic cavitation bubble collapses is investigated using statistical methods. The results indicate that this frequency is distributed according to a normal law, its parameters not being time-evolving.

  13. Constraining shallow seismic event depth via synthetic modeling for Expert Technical Analysis at the IDC

    NASA Astrophysics Data System (ADS)

    Stachnik, J.; Rozhkov, M.; Baker, B.; Bobrov, D.; Friberg, P. A.

    2015-12-01

    Depth of event is an important criterion of seismic event screening at the International Data Center, CTBTO. However, a thorough determination of the event depth can be conducted mostly through special analysis because the IDC's Event Definition Criteria is based, in particular, on depth estimation uncertainties. This causes a large number of events in the Reviewed Event Bulletin to have depth constrained to the surface. When the true origin depth is greater than that reasonable for a nuclear test (3 km based on existing observations), this may result in a heavier workload to manually distinguish between shallow and deep events. Also, IDC depth criterion is not applicable to the events with the small t(pP-P) travel time difference, which is the case of the nuclear test. Since the shape of the first few seconds of signal of very shallow events is very sensitive to the presence of the depth phase, cross correlation between observed and theoretic seismogram can provide an estimate for the depth of the event, and so provide an expansion to the screening process. We exercised this approach mostly with events at teleseismic and partially regional distances. We found that such approach can be very efficient for the seismic event screening process, with certain caveats related mostly to the poorly defined crustal models at source and receiver which can shift the depth estimate. We used adjustable t* teleseismic attenuation model for synthetics since this characteristic is not determined for most of the rays we studied. We studied a wide set of historical records of nuclear explosions, including so called Peaceful Nuclear Explosions (PNE) with presumably known depths, and recent DPRK nuclear tests. The teleseismic synthetic approach is based on the stationary phase approximation with Robert Herrmann's hudson96 program, and the regional modelling was done with the generalized ray technique by Vlastislav Cerveny modified to the complex source topography.

  14. Multilingual Analysis of Twitter News in Support of Mass Emergency Events

    NASA Astrophysics Data System (ADS)

    Zielinski, A.; Bügel, U.; Middleton, L.; Middleton, S. E.; Tokarchuk, L.; Watson, K.; Chaves, F.

    2012-04-01

    Social media are increasingly becoming an additional source of information for event-based early warning systems in the sense that they can help to detect natural crises and support crisis management during or after disasters. Within the European FP7 TRIDEC project we study the problem of analyzing multilingual twitter feeds for emergency events. Specifically, we consider tsunami and earthquakes, as one possible originating cause of tsunami, and propose to analyze twitter messages for capturing testified information at affected points of interest in order to obtain a better picture of the actual situation. For tsunami, these could be the so called Forecast Points, i.e. agreed-upon points chosen by the Regional Tsunami Warning Centers (RTWC) and the potentially affected countries, which must be considered when calculating expected tsunami arrival times. Generally, local civil protection authorities and the population are likely to respond in their native languages. Therefore, the present work focuses on English as "lingua franca" and on under-resourced Mediterranean languages in endangered zones, particularly in Turkey, Greece, and Romania. We investigated ten earthquake events and defined four language-specific classifiers that can be used to detect natural crisis events by filtering out irrelevant messages that do not relate to the event. Preliminary results indicate that such a filter has the potential to support earthquake detection and could be integrated into seismographic sensor networks. One hindrance in our study is the lack of geo-located data for asserting the geographical origin of the tweets and thus to be able to observe correlations of events across languages. One way to overcome this deficit consists in identifying geographic names contained in tweets that correspond to or which are located in the vicinity of specific points-of-interest such as the forecast points of the tsunami scenario. We also intend to use twitter analysis for situation picture

  15. Subjective Well-Being and Adaptation to Life Events: A Meta-Analysis on Differences Between Cognitive and Affective Well-Being

    PubMed Central

    Luhmann, Maike; Hofmann, Wilhelm; Eid, Michael; Lucas, Richard E.

    2012-01-01

    Previous research has shown that major life events can have short- and long-term effects on subjective well-being (SWB). The present meta-analysis examines (a) whether life events have different effects on cognitive and affective well-being and (b) how the rate of adaptation varies across different life events. Longitudinal data from 188 publications (313 samples, N = 65,911) were integrated to describe the reaction and adaptation to four family events (marriage, divorce, bereavement, child birth) and four work events (unemployment, reemployment, retirement, relocation/migration). The findings show that life events have very different effects on affective and cognitive well-being, and that for most events the effects of life events on cognitive well-being are stronger and more consistent across samples. Different life events differ in their effects on SWB, but these effects are not a function of the alleged desirability of events. The results are discussed with respect to their theoretical implications, and recommendations for future studies on adaptation are given. PMID:22059843

  16. Replica analysis of overfitting in regression models for time-to-event data

    NASA Astrophysics Data System (ADS)

    Coolen, A. C. C.; Barrett, J. E.; Paga, P.; Perez-Vicente, C. J.

    2017-09-01

    Overfitting, which happens when the number of parameters in a model is too large compared to the number of data points available for determining these parameters, is a serious and growing problem in survival analysis. While modern medicine presents us with data of unprecedented dimensionality, these data cannot yet be used effectively for clinical outcome prediction. Standard error measures in maximum likelihood regression, such as p-values and z-scores, are blind to overfitting, and even for Cox’s proportional hazards model (the main tool of medical statisticians), one finds in literature only rules of thumb on the number of samples required to avoid overfitting. In this paper we present a mathematical theory of overfitting in regression models for time-to-event data, which aims to increase our quantitative understanding of the problem and provide practical tools with which to correct regression outcomes for the impact of overfitting. It is based on the replica method, a statistical mechanical technique for the analysis of heterogeneous many-variable systems that has been used successfully for several decades in physics, biology, and computer science, but not yet in medical statistics. We develop the theory initially for arbitrary regression models for time-to-event data, and verify its predictions in detail for the popular Cox model.

  17. Extreme events in total ozone: Spatio-temporal analysis from local to global scale

    NASA Astrophysics Data System (ADS)

    Rieder, Harald E.; Staehelin, Johannes; Maeder, Jörg A.; Ribatet, Mathieu; di Rocco, Stefania; Jancso, Leonhardt M.; Peter, Thomas; Davison, Anthony C.

    2010-05-01

    Recently tools from extreme value theory (e.g. Coles, 2001; Ribatet, 2007) have been applied for the first time in the field of stratospheric ozone research, as statistical analysis showed that previously used concepts assuming a Gaussian distribution (e.g. fixed deviations from mean values) of total ozone data do not address the internal data structure concerning extremes adequately (Rieder et al., 2010a,b). A case study the world's longest total ozone record (Arosa, Switzerland - for details see Staehelin et al., 1998a,b) illustrates that tools based on extreme value theory are appropriate to identify ozone extremes and to describe the tails of the total ozone record. Excursions in the frequency of extreme events reveal "fingerprints" of dynamical factors such as ENSO or NAO, and chemical factors, such as cold Arctic vortex ozone losses, as well as major volcanic eruptions of the 20th century (e.g. Gunung Agung, El Chichón, Mt. Pinatubo). Furthermore, atmospheric loading in ozone depleting substances led to a continuous modification of column ozone in the northern hemisphere also with respect to extreme values (partly again in connection with polar vortex contributions). It is shown that application of extreme value theory allows the identification of many more such fingerprints than conventional time series analysis of annual and seasonal mean values. Especially, the extremal analysis shows the strong influence of dynamics, revealing that even moderate ENSO and NAO events have a discernible effect on total ozone (Rieder et al., 2010b). Overall the extremes concept provides new information on time series properties, variability, trends and the influence of dynamics and chemistry, complementing earlier analyses focusing only on monthly (or annual) mean values. Findings described above could be proven also for the total ozone records of 5 other long-term series (Belsk, Hohenpeissenberg, Hradec Kralove, Potsdam, Uccle) showing that strong influence of atmospheric

  18. Using Web Crawler Technology for Text Analysis of Geo-Events: A Case Study of the Huangyan Island Incident

    NASA Astrophysics Data System (ADS)

    Hu, H.; Ge, Y. J.

    2013-11-01

    With the social networking and network socialisation have brought more text information and social relationships into our daily lives, the question of whether big data can be fully used to study the phenomenon and discipline of natural sciences has prompted many specialists and scholars to innovate their research. Though politics were integrally involved in the hyperlinked word issues since 1990s, automatic assembly of different geospatial web and distributed geospatial information systems utilizing service chaining have explored and built recently, the information collection and data visualisation of geo-events have always faced the bottleneck of traditional manual analysis because of the sensibility, complexity, relativity, timeliness and unexpected characteristics of political events. Based on the framework of Heritrix and the analysis of web-based text, word frequency, sentiment tendency and dissemination path of the Huangyan Island incident is studied here by combining web crawler technology and the text analysis method. The results indicate that tag cloud, frequency map, attitudes pie, individual mention ratios and dissemination flow graph based on the data collection and processing not only highlight the subject and theme vocabularies of related topics but also certain issues and problems behind it. Being able to express the time-space relationship of text information and to disseminate the information regarding geo-events, the text analysis of network information based on focused web crawler technology can be a tool for understanding the formation and diffusion of web-based public opinions in political events.

  19. Transcriptome Bioinformatical Analysis of Vertebrate Stages of Schistosoma japonicum Reveals Alternative Splicing Events

    PubMed Central

    Wang, Xinye; Xu, Xindong; Lu, Xingyu; Zhang, Yuanbin; Pan, Weiqing

    2015-01-01

    Alternative splicing is a molecular process that contributes greatly to the diversification of proteome and to gene functions. Understanding the mechanisms of stage-specific alternative splicing can provide a better understanding of the development of eukaryotes and the functions of different genes. Schistosoma japonicum is an infectious blood-dwelling trematode with a complex lifecycle that causes the tropical disease schistosomiasis. In this study, we analyzed the transcriptome of Schistosoma japonicum to discover alternative splicing events in this parasite, by applying RNA-seq to cDNA library of adults and schistosomula. Results were validated by RT-PCR and sequencing. We found 11,623 alternative splicing events among 7,099 protein encoding genes and average proportion of alternative splicing events per gene was 42.14%. We showed that exon skip is the most common type of alternative splicing events as found in high eukaryotes, whereas intron retention is the least common alternative splicing type. According to intron boundary analysis, the parasite possesses same intron boundaries as other organisms, namely the classic “GT-AG” rule. And in alternative spliced introns or exons, this rule is less strict. And we have attempted to detect alternative splicing events in genes encoding proteins with signal peptides and transmembrane helices, suggesting that alternative splicing could change subcellular locations of specific gene products. Our results indicate that alternative splicing is prevalent in this parasitic worm, and that the worm is close to its hosts. The revealed secretome involved in alternative splicing implies new perspective into understanding interaction between the parasite and its host. PMID:26407301

  20. The Braunsbach Flashflood of May 29, 2016: A forensic analysis of the meteorological origin and the hydrological development an extreme hydro-meteorological event

    NASA Astrophysics Data System (ADS)

    Bronstert, Axel; Ankit, Agarwal; Berry, Boessenkool; Madlen, Fischer; Maik, Heistermann; Lisei, Köhn-Reich; Thomas, Moran; Dadiyorto, Wendi

    2017-04-01

    The flash-flood at 29th May 2016 in the vicinity of the village of Braunsbach in Southwestern Germany, State of Baden-Wuerttemberg, has been a particularly concise event of the floods occurring in southern Germany at the end of May / early June 2016. This extreme event was triggered by a convective high intensity rain storm, causing extreme discharge rates and subsequent debris flow in the local creek. This led to severe flooding of the village with immense damages. Besides its extreme nature, the event is characterized by very local and short term scales, i.e. the catchment of the creek covers an area of only six km2 and the whole event lasted only two hours. This contribution presents a retrospective analysis with regard to meteorology and hydrology to obtain a quantitative assessment of the governing processes and their development. We term this a "forensic analysis" because due to the very local and sudden feature of this flashflood event, the processes cannot be directly measured during the event and/or at the site. Instead, they need to be reconstructed and estimated after the event from a variety of rather different information sources and "soft" data. Using these types of post event observations and analysis, we aim at obtaining a rather comprehensive picture of the event and its consequences. Regarding rainfall, both station data from the surroundings of the catchment and radar data from the German Weather Service were analyzed, including the analysis of different errors types and dynamic features of the convective system. The flood hydrograph, including the maximum discharge rate during the event, was estimated by three different approaches, which were compared to obtain an idea of the associated uncertainty. The overall results of this forensic analysis show that it was a very rare rainfall event with extreme rainfall intensities, e.g. return period exceeding 100 years. Catalyzed by catchment properties, this lead to extreme runoff, severe soil erosion

  1. Single Particle Analysis by Combined Chemical Imaging to Study Episodic Air Pollution Events in Vienna

    NASA Astrophysics Data System (ADS)

    Ofner, Johannes; Eitenberger, Elisabeth; Friedbacher, Gernot; Brenner, Florian; Hutter, Herbert; Schauer, Gerhard; Kistler, Magdalena; Greilinger, Marion; Lohninger, Hans; Lendl, Bernhard; Kasper-Giebl, Anne

    2017-04-01

    The aerosol composition of a city like Vienna is characterized by a complex interaction of local emissions and atmospheric input on a regional and continental scale. The identification of major aerosol constituents for basic source appointment and air quality issues needs a high analytical effort. Exceptional episodic air pollution events strongly change the typical aerosol composition of a city like Vienna on a time-scale of few hours to several days. Analyzing the chemistry of particulate matter from these events is often hampered by the sampling time and related sample amount necessary to apply the full range of bulk analytical methods needed for chemical characterization. Additionally, morphological and single particle features are hardly accessible. Chemical Imaging evolved to a powerful tool for image-based chemical analysis of complex samples. As a complementary technique to bulk analytical methods, chemical imaging can address a new access to study air pollution events by obtaining major aerosol constituents with single particle features at high temporal resolutions and small sample volumes. The analysis of the chemical imaging datasets is assisted by multivariate statistics with the benefit of image-based chemical structure determination for direct aerosol source appointment. A novel approach in chemical imaging is combined chemical imaging or so-called multisensor hyperspectral imaging, involving elemental imaging (electron microscopy-based energy dispersive X-ray imaging), vibrational imaging (Raman micro-spectroscopy) and mass spectrometric imaging (Time-of-Flight Secondary Ion Mass Spectrometry) with subsequent combined multivariate analytics. Combined chemical imaging of precipitated aerosol particles will be demonstrated by the following examples of air pollution events in Vienna: Exceptional episodic events like the transformation of Saharan dust by the impact of the city of Vienna will be discussed and compared to samples obtained at a high alpine

  2. Painful and provocative events scale and fearlessness about death among Veterans: Exploratory factor analysis.

    PubMed

    Poindexter, Erin K; Nazem, Sarra; Forster, Jeri E

    2017-01-15

    The interpersonal theory of suicide suggests three proximal risk factors for suicide: perceived burdensomeness, thwarted belongingness, and acquired capability. Previous literature indicates that repetitive exposure to painful and provocative events is related to increased acquired capability for suicide. Despite this, research related to the assessment of painful and provocative events has been insufficient. Research has inconsistently administered the Painful and Provocative Events Scale (PPES; a painful and provocative events assessment), and no study has examined the factor structure of the English PPES. This study explored the factor structure of the PPES and the relation between factors and fearlessness about death. The sample was a cross-sectional, self-report study comprised of 119 Veterans (Mage = 46.5, SD = 13.5). Findings from an exploratory factor analysis indicated a four-factor solution for the PPES; however, no factor from the PPES significantly related to fearlessness about death (measured by the Acquired Capability for Suicide Scale - Fearlessness About Death Scale; all p >.21). Cross-sectional, small Veteran sample. Findings suggest that the PPES lacks the psychometric properties necessary to reliably investigate painful and provocative factors. Consequently, this measure may not reliably capture and explain how painful and provocative events relate to fearlessness about death, which is a barrier to improving suicide risk assessment and prediction. Recommendations for the construction of a new PPES are offered. Published by Elsevier B.V.

  3. Rare event computation in deterministic chaotic systems using genealogical particle analysis

    NASA Astrophysics Data System (ADS)

    Wouters, J.; Bouchet, F.

    2016-09-01

    In this paper we address the use of rare event computation techniques to estimate small over-threshold probabilities of observables in deterministic dynamical systems. We demonstrate that genealogical particle analysis algorithms can be successfully applied to a toy model of atmospheric dynamics, the Lorenz ’96 model. We furthermore use the Ornstein-Uhlenbeck system to illustrate a number of implementation issues. We also show how a time-dependent objective function based on the fluctuation path to a high threshold can greatly improve the performance of the estimator compared to a fixed-in-time objective function.

  4. Trauma and recent life events in individuals at ultra high risk for psychosis: review and meta-analysis.

    PubMed

    Kraan, Tamar; Velthorst, Eva; Smit, Filip; de Haan, Lieuwe; van der Gaag, Mark

    2015-02-01

    Childhood trauma and recent life-events have been related to psychotic disorders. The aim of the present study was to examine whether childhood trauma and recent life-events are significantly more prevalent in patients at Ultra High Risk (UHR) of developing a psychotic disorder compared to healthy controls. A search of PsychInfo and Embase was conducted, relevant papers were reviewed, and three random-effects meta-analyses were performed. One meta-analysis assessed the prevalence rate of childhood trauma in UHR subjects and two meta-analyses were conducted to compare UHR subjects and healthy control subjects on the experience of childhood trauma and recent life-events. We found 12 studies on the prevalence of (childhood) trauma in UHR populations and 4 studies on recent life-events in UHR populations. We performed a meta-analysis on 6 studies (of which trauma prevalence rates were available) on childhood trauma in UHR populations, yielding a mean prevalence rate of 86.8% (95% CI 77%-93%). Childhood trauma was significantly more prevalent in UHR subjects compared to healthy control groups (Random effects Hedges' g=1.09; Z=4.60, p<.001). In contrast to our hypothesis, life-event rates were significantly lower in UHR subjects compared to healthy controls (Random effects Hedges' g=-0.53; Z=-2.36, p<.02). Our meta-analytic results illustrate that childhood trauma is highly prevalent among UHR subjects and that childhood trauma is related to UHR status. These results are in line with studies on childhood trauma in psychotic populations. In contrast to studies on recent life-events in psychotic populations, our results show that recent life-events are not associated with UHR status. Copyright © 2014 Elsevier B.V. All rights reserved.

  5. Living with extreme weather events - perspectives from climatology, geomorphological analysis, chronicles and opinion polls

    NASA Astrophysics Data System (ADS)

    Auer, I.; Kirchengast, A.; Proske, H.

    2009-09-01

    The ongoing climate change debate focuses more and more on changing extreme events. Information on past events can be derived from a number of sources, such as instrumental data, residual impacts in the landscape, but also chronicles and people's memories. A project called "A Tale of Two Valleys” within the framework of the research program "proVision” allowed to study past extreme events in two inner-alpine valleys from the sources mentioned before. Instrumental climate time series provided information for the past 200 years, however great attention had to be given to the homogeneity of the series. To derive homogenized time series of selected climate change indices methods like HOCLIS and Vincent have been applied. Trend analyses of climate change indices inform about increase or decrease of extreme events. Traces of major geomorphodynamic processes of the past (e.g. rockfalls, landslides, debris flows) which were triggered or affected by extreme weather events are still apparent in the landscape and could be evaluated by geomorphological analysis using remote sensing and field data. Regional chronicles provided additional knowledge and covered longer periods back in time, however compared to meteorological time series they enclose a high degree of subjectivity and intermittent recordings cannot be obviated. Finally, questionnaires and oral history complemented our picture of past extreme weather events. People were differently affected and have different memories of it. The joint analyses of these four data sources showed agreement to some extent, however also showed some reasonable differences: meteorological data are point measurements only with a sometimes too coarse temporal resolution. Due to land-use changes and improved constructional measures the impact of an extreme meteorological event may be different today compared to earlier times.

  6. Root cause analysis of serious adverse events among older patients in the Veterans Health Administration.

    PubMed

    Lee, Alexandra; Mills, Peter D; Neily, Julia; Hemphill, Robin R

    2014-06-01

    Preventable adverse events are more likely to occur among older patients because of the clinical complexity of their care. The Veterans Health Administration (VHA) National Center for Patient Safety (NCPS) stores data about serious adverse events when a root cause analysis (RCA) has been performed. A primary objective of this study was to describe the types of adverse events occurring among older patients (age > or = 65 years) in Department of Veterans Affairs (VA) hospitals. Secondary objectives were to determine the underlying reasons for the occurrence of these events and report on effective action plans that have been implemented in VA hospitals. In a retrospective, cross-sectional review, RCA reports were reviewed and outcomes reported using descriptive statistics for all VA hospitals that conducted an RCA for a serious geriatric adverse event from January 2010 to January 2011 that resulted in sustained injury or death. The search produced 325 RCA reports on VA patients (age > or = 65 years). Falls (34.8%), delays in diagnosis and/or treatment (11.7%), unexpected death (9.9%), and medication errors (9.0%) were the most commonly reported adverse events among older VA patients. Communication was the most common underlying reason for these events, representing 43.9% of reported root causes. Approximately 40% of implemented action plans were judged by local staff to be effective. The RCA process identified falls and communication as important themes in serious adverse events. Concrete actions, such as process standardization and changes to communication, were reported by teams to yield some improvement. However, fewer than half of the action plans were reported to be effective. Further research is needed to guide development and implementation of effective action plans.

  7. Analysis and Prediction of Exon Skipping Events from RNA-Seq with Sequence Information Using Rotation Forest.

    PubMed

    Du, Xiuquan; Hu, Changlin; Yao, Yu; Sun, Shiwei; Zhang, Yanping

    2017-12-12

    In bioinformatics, exon skipping (ES) event prediction is an essential part of alternative splicing (AS) event analysis. Although many methods have been developed to predict ES events, a solution has yet to be found. In this study, given the limitations of machine learning algorithms with RNA-Seq data or genome sequences, a new feature, called RS (RNA-seq and sequence) features, was constructed. These features include RNA-Seq features derived from the RNA-Seq data and sequence features derived from genome sequences. We propose a novel Rotation Forest classifier to predict ES events with the RS features (RotaF-RSES). To validate the efficacy of RotaF-RSES, a dataset from two human tissues was used, and RotaF-RSES achieved an accuracy of 98.4%, a specificity of 99.2%, a sensitivity of 94.1%, and an area under the curve (AUC) of 98.6%. When compared to the other available methods, the results indicate that RotaF-RSES is efficient and can predict ES events with RS features.

  8. Preliminary analysis on faint luminous lightning events recorded by multiple high speed cameras

    NASA Astrophysics Data System (ADS)

    Alves, J.; Saraiva, A. V.; Pinto, O.; Campos, L. Z.; Antunes, L.; Luz, E. S.; Medeiros, C.; Buzato, T. S.

    2013-12-01

    The objective of this work is the study of some faint luminous events produced by lightning flashes that were recorded simultaneously by multiple high-speed cameras during the previous RAMMER (Automated Multi-camera Network for Monitoring and Study of Lightning) campaigns. The RAMMER network is composed by three fixed cameras and one mobile color camera separated by, in average, distances of 13 kilometers. They were located in the Paraiba Valley (in the cities of São José dos Campos and Caçapava), SP, Brazil, arranged in a quadrilateral shape, centered in São José dos Campos region. This configuration allowed RAMMER to see a thunderstorm from different angles, registering the same lightning flashes simultaneously by multiple cameras. Each RAMMER sensor is composed by a triggering system and a Phantom high-speed camera version 9.1, which is set to operate at a frame rate of 2,500 frames per second with a lens Nikkor (model AF-S DX 18-55 mm 1:3.5 - 5.6 G in the stationary sensors, and a lens model AF-S ED 24 mm - 1:1.4 in the mobile sensor). All videos were GPS (Global Positioning System) time stamped. For this work we used a data set collected in four RAMMER manual operation days in the campaign of 2012 and 2013. On Feb. 18th the data set is composed by 15 flashes recorded by two cameras and 4 flashes recorded by three cameras. On Feb. 19th a total of 5 flashes was registered by two cameras and 1 flash registered by three cameras. On Feb. 22th we obtained 4 flashes registered by two cameras. Finally, in March 6th two cameras recorded 2 flashes. The analysis in this study proposes an evaluation methodology for faint luminous lightning events, such as continuing current. Problems in the temporal measurement of the continuing current can generate some imprecisions during the optical analysis, therefore this work aim to evaluate the effects of distance in this parameter with this preliminary data set. In the cases that include the color camera we analyzed the RGB

  9. g-PRIME: A Free, Windows Based Data Acquisition and Event Analysis Software Package for Physiology in Classrooms and Research Labs.

    PubMed

    Lott, Gus K; Johnson, Bruce R; Bonow, Robert H; Land, Bruce R; Hoy, Ronald R

    2009-01-01

    We present g-PRIME, a software based tool for physiology data acquisition, analysis, and stimulus generation in education and research. This software was developed in an undergraduate neurophysiology course and strongly influenced by instructor and student feedback. g-PRIME is a free, stand-alone, windows application coded and "compiled" in Matlab (does not require a Matlab license). g-PRIME supports many data acquisition interfaces from the PC sound card to expensive high throughput calibrated equipment. The program is designed as a software oscilloscope with standard trigger modes, multi-channel visualization controls, and data logging features. Extensive analysis options allow real time and offline filtering of signals, multi-parameter threshold-and-window based event detection, and two-dimensional display of a variety of parameters including event time, energy density, maximum FFT frequency component, max/min amplitudes, and inter-event rate and intervals. The software also correlates detected events with another simultaneously acquired source (event triggered average) in real time or offline. g-PRIME supports parameter histogram production and a variety of elegant publication quality graphics outputs. A major goal of this software is to merge powerful engineering acquisition and analysis tools with a biological approach to studies of nervous system function.

  10. Stressful life events during adolescence and risk for externalizing and internalizing psychopathology: a meta-analysis.

    PubMed

    March-Llanes, Jaume; Marqués-Feixa, Laia; Mezquita, Laura; Fañanás, Lourdes; Moya-Higueras, Jorge

    2017-12-01

    The main objective of the present research was to analyze the relations between stressful life events and the externalizing and internalizing spectra of psychopathology using meta-analytical procedures. After removing the duplicates, a total of 373 papers were found in a literature search using several bibliographic databases, such as the PsycINFO, Medline, Scopus, and Web of Science. Twenty-seven studies were selected for the meta-analytical analysis after applying different inclusion and exclusion criteria in different phases. The statistical procedure was performed using a random/mixed-effects model based on the correlations found in the studies. Significant positive correlations were found in cross-sectional and longitudinal studies. A transactional effect was then found in the present study. Stressful life events could be a cause, but also a consequence, of psychopathological spectra. The level of controllability of the life events did not affect the results. Special attention should be given to the usage of stressful life events in gene-environment interaction and correlation studies, and also for clinical purposes.

  11. Adverse events following yellow fever immunization: Report and analysis of 67 neurological cases in Brazil.

    PubMed

    Martins, Reinaldo de Menezes; Pavão, Ana Luiza Braz; de Oliveira, Patrícia Mouta Nunes; dos Santos, Paulo Roberto Gomes; Carvalho, Sandra Maria D; Mohrdieck, Renate; Fernandes, Alexandre Ribeiro; Sato, Helena Keico; de Figueiredo, Patricia Mandali; von Doellinger, Vanessa Dos Reis; Leal, Maria da Luz Fernandes; Homma, Akira; Maia, Maria de Lourdes S

    2014-11-20

    Neurological adverse events following administration of the 17DD substrain of yellow fever vaccine (YEL-AND) in the Brazilian population are described and analyzed. Based on information obtained from the National Immunization Program through passive surveillance or intensified passive surveillance, from 2007 to 2012, descriptive analysis, national and regional rates of YFV associated neurotropic, neurological autoimmune disease, and reporting rate ratios with their respective 95% confidence intervals were calculated for first time vaccinees stratified on age and year. Sixty-seven neurological cases were found, with the highest rate of neurological adverse events in the age group from 5 to 9 years (2.66 per 100,000 vaccine doses in Rio Grande do Sul state, and 0.83 per 100,000 doses in national analysis). Two cases had a combination of neurotropic and autoimmune features. This is the largest sample of YEL-AND already analyzed. Rates are similar to other recent studies, but on this study the age group from 5 to 9 years of age had the highest risk. As neurological adverse events have in general a good prognosis, they should not contraindicate the use of yellow fever vaccine in face of risk of infection by yellow fever virus. Copyright © 2014 Elsevier Ltd. All rights reserved.

  12. Flow detection via sparse frame analysis for suspicious event recognition in infrared imagery

    NASA Astrophysics Data System (ADS)

    Fernandes, Henrique C.; Batista, Marcos A.; Barcelos, Celia A. Z.; Maldague, Xavier P. V.

    2013-05-01

    It is becoming increasingly evident that intelligent systems are very bene¯cial for society and that the further development of such systems is necessary to continue to improve society's quality of life. One area that has drawn the attention of recent research is the development of automatic surveillance systems. In our work we outline a system capable of monitoring an uncontrolled area (an outside parking lot) using infrared imagery and recognizing suspicious events in this area. The ¯rst step is to identify moving objects and segment them from the scene's background. Our approach is based on a dynamic background-subtraction technique which robustly adapts detection to illumination changes. It is analyzed only regions where movement is occurring, ignoring in°uence of pixels from regions where there is no movement, to segment moving objects. Regions where movement is occurring are identi¯ed using °ow detection via sparse frame analysis. During the tracking process the objects are classi¯ed into two categories: Persons and Vehicles, based on features such as size and velocity. The last step is to recognize suspicious events that may occur in the scene. Since the objects are correctly segmented and classi¯ed it is possible to identify those events using features such as velocity and time spent motionless in one spot. In this paper we recognize the suspicious event suspicion of object(s) theft from inside a parked vehicle at spot X by a person" and results show that the use of °ow detection increases the recognition of this suspicious event from 78:57% to 92:85%.

  13. An Event History Analysis of Teacher Attrition: Salary, Teacher Tracking, and Socially Disadvantaged Schools

    ERIC Educational Resources Information Center

    Kelly, Sean

    2004-01-01

    In this event history analysis of the 1990-1991 Schools and Staffing Survey and the 1992 Teacher Follow-up Survey, a retrospective person-year database was constructed to examine teacher attrition over the course of the teaching career. Consistent with prior research, higher teacher salaries reduced attrition, but only slightly so. Teacher…

  14. Visual search of cyclic spatio-temporal events

    NASA Astrophysics Data System (ADS)

    Gautier, Jacques; Davoine, Paule-Annick; Cunty, Claire

    2018-05-01

    The analysis of spatio-temporal events, and especially of relationships between their different dimensions (space-time-thematic attributes), can be done with geovisualization interfaces. But few geovisualization tools integrate the cyclic dimension of spatio-temporal event series (natural events or social events). Time Coil and Time Wave diagrams represent both the linear time and the cyclic time. By introducing a cyclic temporal scale, these diagrams may highlight the cyclic characteristics of spatio-temporal events. However, the settable cyclic temporal scales are limited to usual durations like days or months. Because of that, these diagrams cannot be used to visualize cyclic events, which reappear with an unusual period, and don't allow to make a visual search of cyclic events. Also, they don't give the possibility to identify the relationships between the cyclic behavior of the events and their spatial features, and more especially to identify localised cyclic events. The lack of possibilities to represent the cyclic time, outside of the temporal diagram of multi-view geovisualization interfaces, limits the analysis of relationships between the cyclic reappearance of events and their other dimensions. In this paper, we propose a method and a geovisualization tool, based on the extension of Time Coil and Time Wave, to provide a visual search of cyclic events, by allowing to set any possible duration to the diagram's cyclic temporal scale. We also propose a symbology approach to push the representation of the cyclic time into the map, in order to improve the analysis of relationships between space and the cyclic behavior of events.

  15. Effect of Intravitreal Anti-Vascular Endothelial Growth Factor Therapy on the Risk of Arterial Thromboembolic Events: A Meta-Analysis

    PubMed Central

    Lu, Guo-Cai; Wei, Rui-Li

    2012-01-01

    Background Intravitreal anti-vascular endothelial growth factor (VEGF) monoclonal antibodies are used in ocular neovascular diseases. A consensus has emerged that intravenous anti-VEGF can increase the risk of arterial thromboembolic events. However, the role of intravitreal anti-VEGF in arterial thromboembolism is controversial. Therefore, we did a systematic review and meta-analysis to investigate the effects of intravitreal anti-VEGF on the risk of arterial thromboembolic events. Methods Electronic databases were searched to identify relevant randomized clinical trials comparing intravitreal anti-VEGF with controls. Criteria for inclusion in our meta-analysis included a study duration of no less than 12 months, the use of a randomized control group not receiving any intravitreal active agent, and the availability of outcome data for arterial thromboembolic events, myocardial infarction, cerebrovascular accidents, and vascular death. The risk ratios and 95% CIs were calculated using a fixed-effects or random-effects model, depending on the heterogeneity of the included studies. Results A total of 4942 patients with a variety of ocular neovascular diseases from 13 randomized controlled trials were identified and included for analysis. There was no significant difference between intravitreal anti-VEGF and control in the risk of all events, with risk ratios of 0.87 (95% CI, 0.64 to 1.19) for arterial thromboembolic events, 0.96 (95% CI, 0.55–1.68) for cerebrovascular accidents, 0.69 (95% CI 0.40–1.21) for myocardial infarctions, and 0.68 (95% CI, 0.37–1.27) for vascular death. Conclusions The strength evidence suggests that the intravitreal use of anti-VEGF antibodies is not associated with an increased risk of arterial thromboembolic events. PMID:22829940

  16. Compression Algorithm Analysis of In-Situ (S)TEM Video: Towards Automatic Event Detection and Characterization

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Teuton, Jeremy R.; Griswold, Richard L.; Mehdi, Beata L.

    Precise analysis of both (S)TEM images and video are time and labor intensive processes. As an example, determining when crystal growth and shrinkage occurs during the dynamic process of Li dendrite deposition and stripping involves manually scanning through each frame in the video to extract a specific set of frames/images. For large numbers of images, this process can be very time consuming, so a fast and accurate automated method is desirable. Given this need, we developed software that uses analysis of video compression statistics for detecting and characterizing events in large data sets. This software works by converting the datamore » into a series of images which it compresses into an MPEG-2 video using the open source “avconv” utility [1]. The software does not use the video itself, but rather analyzes the video statistics from the first pass of the video encoding that avconv records in the log file. This file contains statistics for each frame of the video including the frame quality, intra-texture and predicted texture bits, forward and backward motion vector resolution, among others. In all, avconv records 15 statistics for each frame. By combining different statistics, we have been able to detect events in various types of data. We have developed an interactive tool for exploring the data and the statistics that aids the analyst in selecting useful statistics for each analysis. Going forward, an algorithm for detecting and possibly describing events automatically can be written based on statistic(s) for each data type.« less

  17. Point pattern analysis applied to flood and landslide damage events in Switzerland (1972-2009)

    NASA Astrophysics Data System (ADS)

    Barbería, Laura; Schulte, Lothar; Carvalho, Filipe; Peña, Juan Carlos

    2017-04-01

    Damage caused by meteorological and hydrological extreme events depends on many factors, not only on hazard, but also on exposure and vulnerability. In order to reach a better understanding of the relation of these complex factors, their spatial pattern and underlying processes, the spatial dependency between values of damage recorded at sites of different distances can be investigated by point pattern analysis. For the Swiss flood and landslide damage database (1972-2009) first steps of point pattern analysis have been carried out. The most severe events have been selected (severe, very severe and catastrophic, according to GEES classification, a total number of 784 damage points) and Ripley's K-test and L-test have been performed, amongst others. For this purpose, R's library spatstat has been used. The results confirm that the damage points present a statistically significant clustered pattern, which could be connected to prevalence of damages near watercourses and also to rainfall distribution of each event, together with other factors. On the other hand, bivariate analysis shows there is no segregated pattern depending on process type: flood/debris flow vs landslide. This close relation points to a coupling between slope and fluvial processes, connectivity between small-size and middle-size catchments and the influence of spatial distribution of precipitation, temperature (snow melt and snow line) and other predisposing factors such as soil moisture, land-cover and environmental conditions. Therefore, further studies will investigate the relationship between the spatial pattern and one or more covariates, such as elevation, distance from watercourse or land use. The final goal will be to perform a regression model to the data, so that the adjusted model predicts the intensity of the point process as a function of the above mentioned covariates.

  18. Dealing With Major Life Events and Transitions: A Systematic Literature Review on and Occupational Analysis of Spirituality.

    PubMed

    Maley, Christine M; Pagana, Nicole K; Velenger, Christa A; Humbert, Tamera Keiter

    2016-01-01

    This systematic literature review analyzed the construct of spirituality as perceived by people who have experienced or are experiencing a major life event or transition. The researchers investigated studies that used narrative analysis or a phenomenological methodology related to the topic. Thematic analysis resulted in three major themes: (1) avenues to and through spirituality, (2) the experience of spirituality, and (3) the meaning of spirituality. The results provide insights into the intersection of spirituality, meaning, and occupational engagement as understood by people experiencing a major life event or transition and suggest further research that addresses spirituality in occupational therapy and interdisciplinary intervention. Copyright © 2016 by the American Occupational Therapy Association, Inc.

  19. Adverse drug events related to mood and emotion in paediatric patients treated for ADHD: A meta-analysis.

    PubMed

    Pozzi, Marco; Carnovale, Carla; Peeters, Gabriëlla G A M; Gentili, Marta; Antoniazzi, Stefania; Radice, Sonia; Clementi, Emilio; Nobile, Maria

    2018-05-22

    ADHD is frequently comorbid with anxiety and mood disorders, which may increase the severity of inattention and hyperactivity symptoms. Emotional symptoms (anxiety, irritability, mood lability) also affect patients without comorbidity or emerge as adverse drug events. The influence of ADHD drugs on emotional symptoms demands investigation to improve therapies. Systematic review of trials reporting adverse events in patients pharmacologically treated for ADHD. Meta-analysis of the occurrence of irritability, anxiety, apathy, reduced talk, sadness, crying, emotional lability, biting nails, staring, perseveration, euphoria. Meta-regression analysis. Forty-five trials were meta-analysed. The most frequently reported outcomes were irritability, anxiety, sadness, and apathy. Methylphenidates, especially immediate-release formulations, were most studied; amphetamines were half as studied and were predominantly mixed amphetamine salts. Reports on atomoxetine were scant. Meta-analysis showed that methylphenidates reduced the risk of irritability, anxiety, euphoria, whereas they worsened the risk of apathy and reduced talk; amphetamines worsened the risk of emotional lability. Factors influencing risks were study year and design, patients' sex and age, drug dose and release formulation. Possible discrepancy between adverse events as indicated in clinical trials and as summarised herein. Confounding due to the aggregation of drugs into groups; uninvestigated sources of bias; incomplete lists of adverse events; lack of observations on self-injury. Methylphenidates appeared safer than amphetamines, although younger patients and females may incur higher risks, especially with high-dose, immediate-release methylphenidates. Only atomoxetine holds a black-box warning, but amphetamines and methylphenidates also did not show a safe profile regarding mood and emotional symptoms. Copyright © 2018. Published by Elsevier B.V.

  20. Ontology-Based Vaccine Adverse Event Representation and Analysis.

    PubMed

    Xie, Jiangan; He, Yongqun

    2017-01-01

    Vaccine is the one of the greatest inventions of modern medicine that has contributed most to the relief of human misery and the exciting increase in life expectancy. In 1796, an English country physician, Edward Jenner, discovered that inoculating mankind with cowpox can protect them from smallpox (Riedel S, Edward Jenner and the history of smallpox and vaccination. Proceedings (Baylor University. Medical Center) 18(1):21, 2005). Based on the vaccination worldwide, we finally succeeded in the eradication of smallpox in 1977 (Henderson, Vaccine 29:D7-D9, 2011). Other disabling and lethal diseases, like poliomyelitis and measles, are targeted for eradication (Bonanni, Vaccine 17:S120-S125, 1999).Although vaccine development and administration are tremendously successful and cost-effective practices to human health, no vaccine is 100% safe for everyone because each person reacts to vaccinations differently given different genetic background and health conditions. Although all licensed vaccines are generally safe for the majority of people, vaccinees may still suffer adverse events (AEs) in reaction to various vaccines, some of which can be serious or even fatal (Haber et al., Drug Saf 32(4):309-323, 2009). Hence, the double-edged sword of vaccination remains a concern.To support integrative AE data collection and analysis, it is critical to adopt an AE normalization strategy. In the past decades, different controlled terminologies, including the Medical Dictionary for Regulatory Activities (MedDRA) (Brown EG, Wood L, Wood S, et al., Drug Saf 20(2):109-117, 1999), the Common Terminology Criteria for Adverse Events (CTCAE) (NCI, The Common Terminology Criteria for Adverse Events (CTCAE). Available from: http://evs.nci.nih.gov/ftp1/CTCAE/About.html . Access on 7 Oct 2015), and the World Health Organization (WHO) Adverse Reactions Terminology (WHO-ART) (WHO, The WHO Adverse Reaction Terminology - WHO-ART. Available from: https://www.umc-products.com/graphics/28010.pdf

  1. Analysis of brand personality to involve event involvement and loyalty: A case study of Jakarta Fashion Week 2017

    NASA Astrophysics Data System (ADS)

    Nasution, A. H.; Rachmawan, Y. A.

    2018-04-01

    Fashion trend in the world changed extremely fast. Fashion has become the one of people’s lifestyle in the world. Fashion week events in several areas can be a measurement of fahion trend nowadays. There was a fashion week event in Indonesia called Jakarta Fashion Week (JFW) aims to show fashion trend to people who want to improve their fashion style. People will join some events if the event has involvement to them, hence they will come to that event again and again. Annually and continuously event is really important to create loyalty among people who are involved in it, in order to increase positive development towards the organizer in organizing the next event. Saving a huge amount from the marketing budget, and creating a higher quality event. This study aims to know the effect of 5 brand personality dimension to event involvement and loyalty in Jakarta Fashion Week (JFW). This study use quantitative confirmative method with Structural Equation Model (SEM) analysis technique. The sample of this study is 150 respondents who became a participant of Jakarta Fashion Week 2017. Result show that there was significant effect of 5 brand personality dimension to 3 dimension of event involvement and loyalty. Meanwhile, there was one dimension of event involvement called personal self-expression that has not effect to loyalty.

  2. A cyber-event correlation framework and metrics

    NASA Astrophysics Data System (ADS)

    Kang, Myong H.; Mayfield, Terry

    2003-08-01

    In this paper, we propose a cyber-event fusion, correlation, and situation assessment framework that, when instantiated, will allow cyber defenders to better understand the local, regional, and global cyber-situation. This framework, with associated metrics, can be used to guide assessment of our existing cyber-defense capabilities, and to help evaluate the state of cyber-event correlation research and where we must focus our future cyber-event correlation research. The framework, based on the cyber-event gathering activities and analysis functions, consists of five operational steps, each of which provides a richer set of contextual information to support greater situational understanding. The first three steps are categorically depicted as increasingly richer and broader-scoped contexts achieved through correlation activity, while in the final two steps, these richer contexts are achieved through analytical activities (situation assessment, and threat analysis & prediction). Category 1 Correlation focuses on the detection of suspicious activities and the correlation of events from a single cyber-event source. Category 2 Correlation clusters the same or similar events from multiple detectors that are located at close proximity and prioritizes them. Finally, the events from different time periods and event sources at different location/regions are correlated at Category 3 to recognize the relationship among different events. This is the category that focuses on the detection of large-scale and coordinated attacks. The situation assessment step (Category 4) focuses on the assessment of cyber asset damage and the analysis of the impact on missions. The threat analysis and prediction step (Category 5) analyzes attacks based on attack traces and predicts the next steps. Metrics that can distinguish correlation and cyber-situation assessment tools for each category are also proposed.

  3. Investigation of 2-stage meta-analysis methods for joint longitudinal and time-to-event data through simulation and real data application.

    PubMed

    Sudell, Maria; Tudur Smith, Catrin; Gueyffier, François; Kolamunnage-Dona, Ruwanthi

    2018-04-15

    Joint modelling of longitudinal and time-to-event data is often preferred over separate longitudinal or time-to-event analyses as it can account for study dropout, error in longitudinally measured covariates, and correlation between longitudinal and time-to-event outcomes. The joint modelling literature focuses mainly on the analysis of single studies with no methods currently available for the meta-analysis of joint model estimates from multiple studies. We propose a 2-stage method for meta-analysis of joint model estimates. These methods are applied to the INDANA dataset to combine joint model estimates of systolic blood pressure with time to death, time to myocardial infarction, and time to stroke. Results are compared to meta-analyses of separate longitudinal or time-to-event models. A simulation study is conducted to contrast separate versus joint analyses over a range of scenarios. Using the real dataset, similar results were obtained by using the separate and joint analyses. However, the simulation study indicated a benefit of use of joint rather than separate methods in a meta-analytic setting where association exists between the longitudinal and time-to-event outcomes. Where evidence of association between longitudinal and time-to-event outcomes exists, results from joint models over standalone analyses should be pooled in 2-stage meta-analyses. © 2017 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.

  4. Paleo-event data standards for dendrochronology

    Treesearch

    Elaine Kennedy Sutherland; P. Brewer; W. Gross

    2017-01-01

    Extreme environmental events, such as storm winds, landslides, insect infestations, and wildfire, cause loss of life, resources, and human infrastructure. Disaster riskreduction analysis can be improved with information about past frequency, intensity, and spatial patterns of extreme events. Tree-ring analyses can provide such information: tree rings reflect events as...

  5. Multi-Sensory Aerosol Data and the NRL NAAPS model for Regulatory Exceptional Event Analysis

    NASA Astrophysics Data System (ADS)

    Husar, R. B.; Hoijarvi, K.; Westphal, D. L.; Haynes, J.; Omar, A. H.; Frank, N. H.

    2013-12-01

    Beyond scientific exploration and analysis, multi-sensory observations along with models are finding increasing applications for operational air quality management. EPA's Exceptional Event (EE) Rule allows the exclusion of data strongly influenced by impacts from "exceptional events," such as smoke from wildfires or dust from abnormally high winds. The EE Rule encourages the use of satellite observations and other non-standard data along with models as evidence for formal documentation of EE samples for exclusion. Thus, the implementation of the EE Rule is uniquely suited for the direct application of integrated multi-sensory observations and indirectly through the assimilation into an aerosol simulation model. Here we report the results of a project: NASA and NAAPS Products for Air Quality Decision Making. The project uses of observations from multiple satellite sensors, surface-based aerosol measurements and the NRL Aerosol Analysis and Prediction System (NAAPS) model that assimilates key satellite observations. The satellite sensor data for detecting and documenting smoke and dust events include: MODIS AOD and Images; OMI Aerosol Index, Tropospheric NO2; AIRS, CO. The surface observations include the EPA regulatory PM2.5 network; the IMPROVE/STN aerosol chemical network; AIRNOW PM2.5 mass network, and surface met. data. Within this application, crucial role is assigned to the NAAPS model for estimating the surface concentration of windblown dust and biomass smoke. The operational model assimilates quality-assured daily MODIS data and 2DVAR to adjust the model concentrations and CALIOP-based climatology to adjust the vertical profiles at 6-hour intervals. The assimilation of satellite data from multiple satellites significantly contributes to the usefulness of NAAPS for EE analysis. The NAAPS smoke and dust simulations were evaluated using the IMPROVE/STN chemical data. The multi-sensory observations along with the model simulations are integrated into a web

  6. Cold periods and coronary events: an analysis of populations worldwide

    PubMed Central

    Barnett, A.; Dobson, A.; McElduff, P.; Salomaa, V.; Kuulasmaa, K.; Sans, S.; t for

    2005-01-01

    Study objective: To investigate the association between cold periods and coronary events, and the extent to which climate, sex, age, and previous cardiac history increase risk during cold weather. Design: A hierarchical analyses of populations from the World Health Organisation's MONICA project. Setting: Twenty four populations from the WHO's MONICA project, a 21 country register made between 1980 and 1995. Patients: People aged 35–64 years who had a coronary event. Main results: Daily rates of coronary events were correlated with the average temperature over the current and previous three days. In cold periods, coronary event rates increased more in populations living in warm climates than in populations living in cold climates, where the increases were slight. The increase was greater in women than in men, especially in warm climates. On average, the odds for women having an event in the cold periods were 1.07 higher than the odds for men (95% posterior interval: 1.03 to 1.11). The effects of cold periods were similar in those with and without a history of a previous myocardial infarction. Conclusions: Rates of coronary events increased during comparatively cold periods, especially in warm climates. The smaller increases in colder climates suggest that some events in warmer climates are preventable. It is suggested that people living in warm climates, particularly women, should keep warm on cold days. PMID:15965137

  7. Latent profile analysis of regression-based norms demonstrates relationship of compounding MS symptom burden and negative work events.

    PubMed

    Frndak, Seth E; Smerbeck, Audrey M; Irwin, Lauren N; Drake, Allison S; Kordovski, Victoria M; Kunker, Katrina A; Khan, Anjum L; Benedict, Ralph H B

    2016-10-01

    We endeavored to clarify how distinct co-occurring symptoms relate to the presence of negative work events in employed multiple sclerosis (MS) patients. Latent profile analysis (LPA) was utilized to elucidate common disability patterns by isolating patient subpopulations. Samples of 272 employed MS patients and 209 healthy controls (HC) were administered neuroperformance tests of ambulation, hand dexterity, processing speed, and memory. Regression-based norms were created from the HC sample. LPA identified latent profiles using the regression-based z-scores. Finally, multinomial logistic regression tested for negative work event differences among the latent profiles. Four profiles were identified via LPA: a common profile (55%) characterized by slightly below average performance in all domains, a broadly low-performing profile (18%), a poor motor abilities profile with average cognition (17%), and a generally high-functioning profile (9%). Multinomial regression analysis revealed that the uniformly low-performing profile demonstrated a higher likelihood of reported negative work events. Employed MS patients with co-occurring motor, memory and processing speed impairments were most likely to report a negative work event, classifying them as uniquely at risk for job loss.

  8. Semiparametric Time-to-Event Modeling in the Presence of a Latent Progression Event

    PubMed Central

    Rice, John D.; Tsodikov, Alex

    2017-01-01

    Summary In cancer research, interest frequently centers on factors influencing a latent event that must precede a terminal event. In practice it is often impossible to observe the latent event precisely, making inference about this process difficult. To address this problem, we propose a joint model for the unobserved time to the latent and terminal events, with the two events linked by the baseline hazard. Covariates enter the model parametrically as linear combinations that multiply, respectively, the hazard for the latent event and the hazard for the terminal event conditional on the latent one. We derive the partial likelihood estimators for this problem assuming the latent event is observed, and propose a profile likelihood–based method for estimation when the latent event is unobserved. The baseline hazard in this case is estimated nonparametrically using the EM algorithm, which allows for closed-form Breslow-type estimators at each iteration, bringing improved computational efficiency and stability compared with maximizing the marginal likelihood directly. We present simulation studies to illustrate the finite-sample properties of the method; its use in practice is demonstrated in the analysis of a prostate cancer data set. PMID:27556886

  9. Semiparametric time-to-event modeling in the presence of a latent progression event.

    PubMed

    Rice, John D; Tsodikov, Alex

    2017-06-01

    In cancer research, interest frequently centers on factors influencing a latent event that must precede a terminal event. In practice it is often impossible to observe the latent event precisely, making inference about this process difficult. To address this problem, we propose a joint model for the unobserved time to the latent and terminal events, with the two events linked by the baseline hazard. Covariates enter the model parametrically as linear combinations that multiply, respectively, the hazard for the latent event and the hazard for the terminal event conditional on the latent one. We derive the partial likelihood estimators for this problem assuming the latent event is observed, and propose a profile likelihood-based method for estimation when the latent event is unobserved. The baseline hazard in this case is estimated nonparametrically using the EM algorithm, which allows for closed-form Breslow-type estimators at each iteration, bringing improved computational efficiency and stability compared with maximizing the marginal likelihood directly. We present simulation studies to illustrate the finite-sample properties of the method; its use in practice is demonstrated in the analysis of a prostate cancer data set. © 2016, The International Biometric Society.

  10. Analysis of arrhythmic events is useful to detect lead failure earlier in patients followed by remote monitoring.

    PubMed

    Nishii, Nobuhiro; Miyoshi, Akihito; Kubo, Motoki; Miyamoto, Masakazu; Morimoto, Yoshimasa; Kawada, Satoshi; Nakagawa, Koji; Watanabe, Atsuyuki; Nakamura, Kazufumi; Morita, Hiroshi; Ito, Hiroshi

    2018-03-01

    Remote monitoring (RM) has been advocated as the new standard of care for patients with cardiovascular implantable electronic devices (CIEDs). RM has allowed the early detection of adverse clinical events, such as arrhythmia, lead failure, and battery depletion. However, lead failure was often identified only by arrhythmic events, but not impedance abnormalities. To compare the usefulness of arrhythmic events with conventional impedance abnormalities for identifying lead failure in CIED patients followed by RM. CIED patients in 12 hospitals have been followed by the RM center in Okayama University Hospital. All transmitted data have been analyzed and summarized. From April 2009 to March 2016, 1,873 patients have been followed by the RM center. During the mean follow-up period of 775 days, 42 lead failure events (atrial lead 22, right ventricular pacemaker lead 5, implantable cardioverter defibrillator [ICD] lead 15) were detected. The proportion of lead failures detected only by arrhythmic events, which were not detected by conventional impedance abnormalities, was significantly higher than that detected by impedance abnormalities (arrhythmic event 76.2%, 95% CI: 60.5-87.9%; impedance abnormalities 23.8%, 95% CI: 12.1-39.5%). Twenty-seven events (64.7%) were detected without any alert. Of 15 patients with ICD lead failure, none has experienced inappropriate therapy. RM can detect lead failure earlier, before clinical adverse events. However, CIEDs often diagnose lead failure as just arrhythmic events without any warning. Thus, to detect lead failure earlier, careful human analysis of arrhythmic events is useful. © 2017 Wiley Periodicals, Inc.

  11. Reducing uncertainty in Climate Response Time Scale by Bayesian Analysis of the 8.2 ka event

    NASA Astrophysics Data System (ADS)

    Lorenz, A.; Held, H.; Bauer, E.; Schneider von Deimling, T.

    2009-04-01

    We analyze the possibility of uncertainty reduction in Climate Response Time Scale by utilizing Greenland ice-core data that contain the 8.2 ka event within a Bayesian model-data intercomparison with the Earth system model of intermediate complexity, CLIMBER-2.3. Within a stochastic version of the model it has been possible to mimic the 8.2 ka event within a plausible experimental setting and with relatively good accuracy considering the timing of the event in comparison to other modeling exercises [1]. The simulation of the centennial cold event is effectively determined by the oceanic cooling rate which depends largely on the ocean diffusivity described by diffusion coefficients of relatively wide uncertainty ranges. The idea now is to discriminate between the different values of diffusivities according to their likelihood to rightly represent the duration of the 8.2 ka event and thus to exploit the paleo data to constrain uncertainty in model parameters in analogue to [2]. Implementing this inverse Bayesian Analysis with this model the technical difficulty arises to establish the related likelihood numerically in addition to the uncertain model parameters: While mainstream uncertainty analyses can assume a quasi-Gaussian shape of likelihood, with weather fluctuating around a long term mean, the 8.2 ka event as a highly nonlinear effect precludes such an a priori assumption. As a result of this study [3] the Bayesian Analysis showed a reduction of uncertainty in vertical ocean diffusivity parameters of factor 2 compared to prior knowledge. This learning effect on the model parameters is propagated to other model outputs of interest; e.g. the inverse ocean heat capacity, which is important for the dominant time scale of climate response to anthropogenic forcing which, in combination with climate sensitivity, strongly influences the climate systems reaction for the near- and medium-term future. 1 References [1] E. Bauer, A. Ganopolski, M. Montoya: Simulation of the

  12. Assessment of Critical Events Corridors through Multivariate Cascading Outages Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Makarov, Yuri V.; Samaan, Nader A.; Diao, Ruisheng

    2011-10-17

    Massive blackouts of electrical power systems in North America over the past decade has focused increasing attention upon ways to identify and simulate network events that may potentially lead to widespread network collapse. This paper summarizes a method to simulate power-system vulnerability to cascading failures to a supplied set of initiating events synonymously termed as Extreme Events. The implemented simulation method is currently confined to simulating steady state power-system response to a set of extreme events. The outlined method of simulation is meant to augment and provide a new insight into bulk power transmission network planning that at present remainsmore » mainly confined to maintaining power system security for single and double component outages under a number of projected future network operating conditions. Although one of the aims of this paper is to demonstrate the feasibility of simulating network vulnerability to cascading outages, a more important goal has been to determine vulnerable parts of the network that may potentially be strengthened in practice so as to mitigate system susceptibility to cascading failures. This paper proposes to demonstrate a systematic approach to analyze extreme events and identify vulnerable system elements that may be contributing to cascading outages. The hypothesis of critical events corridors is proposed to represent repeating sequential outages that can occur in the system for multiple initiating events. The new concept helps to identify system reinforcements that planners could engineer in order to 'break' the critical events sequences and therefore lessen the likelihood of cascading outages. This hypothesis has been successfully validated with a California power system model.« less

  13. Incidence and economic burden of suspected adverse events and adverse event monitoring during AF therapy.

    PubMed

    Kim, M H; Lin, J; Hussein, M; Battleman, D

    2009-12-01

    Rhythm- and rate-control therapies are an essential part of atrial fibrillation (AF) management; however, the use of existing agents is often limited by the occurrence of adverse events. The aim of this study was to evaluate suspected adverse events and adverse event monitoring, and associated medical costs, in patients receiving AF rhythm-control and/or rate-control therapy. This retrospective cohort study used claims data from the Integrated Healthcare Information Systems National Managed Care Benchmark Database from 2002-2006. Patients hospitalized for AF (primary diagnosis), and who had at least 365 days' enrollment before and after the initial (index) AF hospitalization, were included in the analysis. Suspected AF therapy-related adverse events and function tests for adverse event monitoring were identified according to pre-specified diagnosis codes/procedures, and examined over the 12 months following discharge from the index hospitalization. Events/function tests had to have occurred within 90 days of a claim for AF therapy to be considered a suspected adverse event/adverse event monitoring. Of 4174 AF patients meeting the study criteria, 3323 received AF drugs; 428 received rhythm-control only (12.9%), 2130 rate-control only (64.1%), and 765 combined rhythm/rate-control therapy (23.0%). Overall, 50.1% of treated patients had a suspected adverse event and/or function test for adverse event monitoring (45.5% with rate-control, 53.5% with rhythm-control, and 61.2% with combined rhythm/rate-control). Suspected cardiovascular adverse events were the most common events (occurring in 36.1% of patients), followed by pulmonary (6.1%), and endocrine events (5.9%). Overall, suspected adverse events/function tests were associated with mean annual per-patient costs of $3089 ($1750 with rhythm-control, $2041 with rate control, and $6755 with combined rhythm/rate-control). As a retrospective analysis, the study is subject to potential selection bias, while its reliance on

  14. Image Analysis Algorithms for Immunohistochemical Assessment of Cell Death Events and Fibrosis in Tissue Sections

    PubMed Central

    Krajewska, Maryla; Smith, Layton H.; Rong, Juan; Huang, Xianshu; Hyer, Marc L.; Zeps, Nikolajs; Iacopetta, Barry; Linke, Steven P.; Olson, Allen H.; Reed, John C.; Krajewski, Stan

    2009-01-01

    Cell death is of broad physiological and pathological importance, making quantification of biochemical events associated with cell demise a high priority for experimental pathology. Fibrosis is a common consequence of tissue injury involving necrotic cell death. Using tissue specimens from experimental mouse models of traumatic brain injury, cardiac fibrosis, and cancer, as well as human tumor specimens assembled in tissue microarray (TMA) format, we undertook computer-assisted quantification of specific immunohistochemical and histological parameters that characterize processes associated with cell death. In this study, we demonstrated the utility of image analysis algorithms for color deconvolution, colocalization, and nuclear morphometry to characterize cell death events in tissue specimens: (a) subjected to immunostaining for detecting cleaved caspase-3, cleaved poly(ADP-ribose)-polymerase, cleaved lamin-A, phosphorylated histone H2AX, and Bcl-2; (b) analyzed by terminal deoxyribonucleotidyl transferase–mediated dUTP nick end labeling assay to detect DNA fragmentation; and (c) evaluated with Masson's trichrome staining. We developed novel algorithm-based scoring methods and validated them using TMAs as a high-throughput format. The proposed computer-assisted scoring methods for digital images by brightfield microscopy permit linear quantification of immunohistochemical and histochemical stainings. Examples are provided of digital image analysis performed in automated or semiautomated fashion for successful quantification of molecular events associated with cell death in tissue sections. (J Histochem Cytochem 57:649–663, 2009) PMID:19289554

  15. FLOCK cluster analysis of mast cell event clustering by high-sensitivity flow cytometry predicts systemic mastocytosis.

    PubMed

    Dorfman, David M; LaPlante, Charlotte D; Pozdnyakova, Olga; Li, Betty

    2015-11-01

    In our high-sensitivity flow cytometric approach for systemic mastocytosis (SM), we identified mast cell event clustering as a new diagnostic criterion for the disease. To objectively characterize mast cell gated event distributions, we performed cluster analysis using FLOCK, a computational approach to identify cell subsets in multidimensional flow cytometry data in an unbiased, automated fashion. FLOCK identified discrete mast cell populations in most cases of SM (56/75 [75%]) but only a minority of non-SM cases (17/124 [14%]). FLOCK-identified mast cell populations accounted for 2.46% of total cells on average in SM cases and 0.09% of total cells on average in non-SM cases (P < .0001) and were predictive of SM, with a sensitivity of 75%, a specificity of 86%, a positive predictive value of 76%, and a negative predictive value of 85%. FLOCK analysis provides useful diagnostic information for evaluating patients with suspected SM, and may be useful for the analysis of other hematopoietic neoplasms. Copyright© by the American Society for Clinical Pathology.

  16. Incidence of cardiovascular events and associated risk factors in kidney transplant patients: a competing risks survival analysis.

    PubMed

    Seoane-Pillado, María Teresa; Pita-Fernández, Salvador; Valdés-Cañedo, Francisco; Seijo-Bestilleiro, Rocio; Pértega-Díaz, Sonia; Fernández-Rivera, Constantino; Alonso-Hernández, Ángel; González-Martín, Cristina; Balboa-Barreiro, Vanesa

    2017-03-07

    The high prevalence of cardiovascular risk factors among the renal transplant population accounts for increased mortality. The aim of this study is to determine the incidence of cardiovascular events and factors associated with cardiovascular events in these patients. An observational ambispective follow-up study of renal transplant recipients (n = 2029) in the health district of A Coruña (Spain) during the period 1981-2011 was completed. Competing risk survival analysis methods were applied to estimate the cumulative incidence of developing cardiovascular events over time and to identify which characteristics were associated with the risk of these events. Post-transplant cardiovascular events are defined as the presence of myocardial infarction, invasive coronary artery therapy, cerebral vascular events, new-onset angina, congestive heart failure, rhythm disturbances, peripheral vascular disease and cardiovascular disease and death. The cause of death was identified through the medical history and death certificate using ICD9 (390-459, except: 427.5, 435, 446, 459.0). The mean age of patients at the time of transplantation was 47.0 ± 14.2 years; 62% were male. 16.5% had suffered some cardiovascular disease prior to transplantation and 9.7% had suffered a cardiovascular event. The mean follow-up period for the patients with cardiovascular event was 3.5 ± 4.3 years. Applying competing risk methodology, it was observed that the accumulated incidence of the event was 5.0% one year after transplantation, 8.1% after five years, and 11.9% after ten years. After applying multivariate models, the variables with an independent effect for predicting cardiovascular events are: male sex, age of recipient, previous cardiovascular disorders, pre-transplant smoking and post-transplant diabetes. This study makes it possible to determine in kidney transplant patients, taking into account competitive events, the incidence of post-transplant cardiovascular events and

  17. 10-Year analysis of adverse event reports to the Food and Drug Administration for phosphodiesterase type-5 inhibitors.

    PubMed

    Lowe, Gregory; Costabile, Raymond A

    2012-01-01

    To ensure public safety all Food and Drug Administration (FDA)-approved medications undergo postapproval safety analysis. Phosphodiesterase type-5 inhibitors (PDE5-i) are generally regarded as safe and effective. We performed a nonindustry-sponsored analysis of FDA reports for sildenafil, tadalafil, and vardenafil to evaluate the reported cardiovascular and mortality events over the past 10 years. Summarized reports of adverse events (AEs) for each PDE5-i were requested from the Center for Drug Evaluation and Research within the FDA. These data are available under the Freedom of Information Act and document industry and nonindustry reports of AEs entered into the computerized system maintained by the Office of Surveillance and Epidemiology. The data were analyzed for the number of AE reports, number of objective cardiovascular events, and reported deaths. Overall, 14,818 AEs were reported for sildenafil. There were 1,824 (12.3%) reported deaths, and reports of cardiovascular AEs numbered 2,406 (16.2%). Tadalafil was associated with 5,548 AEs and 236 deaths were reported. Vardenafil was associated with 6,085 AEs and 121 reports of deaths. The percentage of reported severe cardiovascular disorders has stabilized at 10% to 15% of all AE reports for sildenafil and tadalafil and 5% to 10% for vardenafil. Only 10% of AE reports sent to the FDA for PDE5-i were from pharmaceutical manufacturers. Reports of deaths associated with PDE5-i remain around 5% of total reported events. Despite inherent limitations from evaluating FDA reports of AEs, it is important that these reports be reviewed outside pharmaceutical industry support in order to provide due diligence and transparency. Lowe G and Costabile RA. 10-year analysis of adverse event reports to the Food and Drug Administration for phosphodiesterase type-5 inhibitors. J Sex Med 2012;9:265-270. © 2011 International Society for Sexual Medicine.

  18. Photographic Analysis Technique for Assessing External Tank Foam Loss Events

    NASA Technical Reports Server (NTRS)

    Rieckhoff, T. J.; Covan, M.; OFarrell, J. M.

    2001-01-01

    A video camera and recorder were placed inside the solid rocket booster forward skirt in order to view foam loss events over an area on the external tank (ET) intertank surface. In this Technical Memorandum, a method of processing video images to allow rapid detection of permanent changes indicative of foam loss events on the ET surface was defined and applied to accurately count, categorize, and locate such events.

  19. 76 FR 70768 - Common-Cause Failure Analysis in Event and Condition Assessment: Guidance and Research, Draft...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-11-15

    ... NUCLEAR REGULATORY COMMISSION [NRC-2011-0254] Common-Cause Failure Analysis in Event and Condition Assessment: Guidance and Research, Draft Report for Comment; Correction AGENCY: Nuclear Regulatory Commission. ACTION: Draft NUREG; request for comment; correction. SUMMARY: This document corrects a notice appearing...

  20. Adverse events and treatment failure leading to discontinuation of recently approved antipsychotic drugs in schizophrenia: A network meta-analysis.

    PubMed

    Tonin, Fernanda S; Piazza, Thais; Wiens, Astrid; Fernandez-Llimos, Fernando; Pontarolo, Roberto

    2015-12-01

    Objective:We aimed to gather evidence of the discontinuation rates owing to adverse events or treatment failure for four recently approved antipsychotics (asenapine, blonanserin, iloperidone, and lurasidone).Methods: A systematic review followed by pairwise meta-analysis and mixed treatment comparison meta analysis(MTC) was performed, including randomized controlled trials (RCTs) that compared the use of the above-mentioned drugs versus placebo in patients with schizophrenia. An electronic search was conducted in PubMed, Scopus, Science Direct, Scielo, the Cochrane Library, and International Pharmaceutical Abstracts(January 2015). The included trials were at least single blinded. The main outcome measures extracted were discontinuation owing to adverse events and discontinuation owing to treatment failure.Results: Fifteen RCTs were identified (n = 5400 participants) and 13 of them were amenable for use in our meta-analyses. No significant differences were observed between any of the four drugs and placebo as regards discontinuation owing to adverse events, whether in pairwise meta-analysis or in MTC. All drugs presented a better profile than placebo on discontinuation owing to treatment failure, both in pairwise meta-analysis and MTC. Asenapine was found to be the best therapy in terms of tolerability owing to failure,while lurasidone was the worst treatment in terms of adverse events. The evidence around blonanserin is weak.Conclusion: MTCs allowed the creation of two different rank orders of these four antipsychotic drugs in two outcome measures. This evidence-generating method allows direct and indirect comparisons, supporting approval and pricing decisions when lacking sufficient, direct, head-to-head trials.

  1. Grid Frequency Extreme Event Analysis and Modeling: Preprint

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Florita, Anthony R; Clark, Kara; Gevorgian, Vahan

    Sudden losses of generation or load can lead to instantaneous changes in electric grid frequency and voltage. Extreme frequency events pose a major threat to grid stability. As renewable energy sources supply power to grids in increasing proportions, it becomes increasingly important to examine when and why extreme events occur to prevent destabilization of the grid. To better understand frequency events, including extrema, historic data were analyzed to fit probability distribution functions to various frequency metrics. Results showed that a standard Cauchy distribution fit the difference between the frequency nadir and prefault frequency (f_(C-A)) metric well, a standard Cauchy distributionmore » fit the settling frequency (f_B) metric well, and a standard normal distribution fit the difference between the settling frequency and frequency nadir (f_(B-C)) metric very well. Results were inconclusive for the frequency nadir (f_C) metric, meaning it likely has a more complex distribution than those tested. This probabilistic modeling should facilitate more realistic modeling of grid faults.« less

  2. Predictability of rogue events.

    PubMed

    Birkholz, Simon; Brée, Carsten; Demircan, Ayhan; Steinmeyer, Günter

    2015-05-29

    Using experimental data from three different rogue wave supporting systems, determinism, and predictability of the underlying dynamics are evaluated with methods of nonlinear time series analysis. We included original records from the Draupner platform in the North Sea as well as time series from two optical systems in our analysis. One of the latter was measured in the infrared tail of optical fiber supercontinua, the other in the fluence profiles of multifilaments. All three data sets exhibit extreme-value statistics and exceed the significant wave height in the respective system by a factor larger than 2. Nonlinear time series analysis indicates a different degree of determinism in the systems. The optical fiber scenario is found to be driven by quantum noise whereas rogue waves emerge as a consequence of turbulence in the others. With the large number of rogue events observed in the multifilament system, we can systematically explore the predictability of such events in a turbulent system. We observe that rogue events do not necessarily appear without a warning, but are often preceded by a short phase of relative order. This surprising finding sheds some new light on the fascinating phenomenon of rogue waves.

  3. “Smooth” Semiparametric Regression Analysis for Arbitrarily Censored Time-to-Event Data

    PubMed Central

    Zhang, Min; Davidian, Marie

    2008-01-01

    Summary A general framework for regression analysis of time-to-event data subject to arbitrary patterns of censoring is proposed. The approach is relevant when the analyst is willing to assume that distributions governing model components that are ordinarily left unspecified in popular semiparametric regression models, such as the baseline hazard function in the proportional hazards model, have densities satisfying mild “smoothness” conditions. Densities are approximated by a truncated series expansion that, for fixed degree of truncation, results in a “parametric” representation, which makes likelihood-based inference coupled with adaptive choice of the degree of truncation, and hence flexibility of the model, computationally and conceptually straightforward with data subject to any pattern of censoring. The formulation allows popular models, such as the proportional hazards, proportional odds, and accelerated failure time models, to be placed in a common framework; provides a principled basis for choosing among them; and renders useful extensions of the models straightforward. The utility and performance of the methods are demonstrated via simulations and by application to data from time-to-event studies. PMID:17970813

  4. Developing future precipitation events from historic events: An Amsterdam case study.

    NASA Astrophysics Data System (ADS)

    Manola, Iris; van den Hurk, Bart; de Moel, Hans; Aerts, Jeroen

    2016-04-01

    Due to climate change, the frequency and intensity of extreme precipitation events is expected to increase. It is therefore of high importance to develop climate change scenarios tailored towards the local and regional needs of policy makers in order to develop efficient adaptation strategies to reduce the risks from extreme weather events. Current approaches to tailor climate scenarios are often not well adopted in hazard management, since average changes in climate are not a main concern to policy makers, and tailoring climate scenarios to simulate future extremes can be complex. Therefore, a new concept has been introduced recently that uses known historic extreme events as a basis, and modifies the observed data for these events so that the outcome shows how the same event would occur in a warmer climate. This concept is introduced as 'Future Weather', and appeals to the experience of stakeholders and users. This research presents a novel method of projecting a future extreme precipitation event, based on a historic event. The selected precipitation event took place over the broader area of Amsterdam, the Netherlands in the summer of 2014, which resulted in blocked highways, disruption of air transportation, flooded buildings and public facilities. An analysis of rain monitoring stations showed that an event of such intensity has a 5 to 15 years return period. The method of projecting a future event follows a non-linear delta transformation that is applied directly on the observed event assuming a warmer climate to produce an "up-scaled" future precipitation event. The delta transformation is based on the observed behaviour of the precipitation intensity as a function of the dew point temperature during summers. The outcome is then compared to a benchmark method using the HARMONIE numerical weather prediction model, where the boundary conditions of the event from the Ensemble Prediction System of ECMWF (ENS) are perturbed to indicate a warmer climate. The two

  5. Broadband Array Analysis of the 2005 Episodic Tremor and Slip Event in Northern Cascadia

    NASA Astrophysics Data System (ADS)

    Wech, A.; Creager, K.; McCausland, W.; Frassetto, A.; Qamar, A.; Derosier, S.; Carmichael, J.; Malone, S.; Johnson, D.

    2005-12-01

    The region of Cascadia from the Olympic Mountains through southern Vancouver Island and down-dip of the subduction megathrust has repeatedly experienced episodes of slow slip. This episodic slip, which has been observed to take place over a period of two to several weeks, is accompanied by a seismic tremor signal. Based on the average recurrence interval of 14 months, the next episodic tremor and slip (ETS) event should occur within six weeks of mid-September, 2005. Indeed, it appears to have begun on September 3, as this abstract was being written. In order to record this anticipated event, we deployed an array of 11 three-component seismometers on the northern side of the Olympic Peninsula augmenting Pacific Northwest Seismographic Network stations as well as the first few EarthScope BigFoot stations and Plate Boundary Observatory borehole seismometers. This seismic array was comprised of six short-period and five broadband instruments with spacings of 500 m and 2200 m respectively. In conjunction with this Earthscope seismic deployment, we also installed a dense network of 29 temporary, continuous GPS stations across the entire Olympic Peninsula to integrate seismic and geodetic observations. One of the primary goals of this research is to utilize the broadband instrumentation in the array to investigate the possible correlation of low frequency energy with the rest of the tremor activity. ETS has been carefully investigated at high-frequency (seismic tremor at 2-6 Hz) and very low-frequency (slip occurring over weeks, observed by GPS). An important goal of this experiment is to investigate the possibility that the tremor generates intermediate, low-frequency signals. Preliminary analysis of short-period array recordings of the July, 2004 ETS event suggests that the tremor displays signs of lower-frequency energy (~0.5 Hz) correlated with its higher frequency activity. Our array should enable us to distinguish low- frequency signals originating in the direction

  6. Use of the Hadoop structured storage tools for the ATLAS EventIndex event catalogue

    NASA Astrophysics Data System (ADS)

    Favareto, A.

    2016-09-01

    The ATLAS experiment at the LHC collects billions of events each data-taking year, and processes them to make them available for physics analysis in several different formats. An even larger amount of events is in addition simulated according to physics and detector models and then reconstructed and analysed to be compared to real events. The EventIndex is a catalogue of all events in each production stage; it includes for each event a few identification parameters, some basic non-mutable information coming from the online system, and the references to the files that contain the event in each format (plus the internal pointers to the event within each file for quick retrieval). Each EventIndex record is logically simple but the system has to hold many tens of billions of records, all equally important. The Hadoop technology was selected at the start of the EventIndex project development in 2012 and proved to be robust and flexible to accommodate this kind of information; both the insertion and query response times are acceptable for the continuous and automatic operation that started in Spring 2015. This paper describes the EventIndex data input and organisation in Hadoop and explains the operational challenges that were overcome in order to achieve the expected performance.

  7. Characterization of high-intensity, long-duration continuous auroral activity (HILDCAA) events using recurrence quantification analysis

    NASA Astrophysics Data System (ADS)

    Mendes, Odim; Oliveira Domingues, Margarete; Echer, Ezequiel; Hajra, Rajkumar; Everton Menconi, Varlei

    2017-08-01

    Considering the magnetic reconnection and the viscous interaction as the fundamental mechanisms for transfer particles and energy into the magnetosphere, we study the dynamical characteristics of auroral electrojet (AE) index during high-intensity, long-duration continuous auroral activity (HILDCAA) events, using a long-term geomagnetic database (1975-2012), and other distinct interplanetary conditions (geomagnetically quiet intervals, co-rotating interaction regions (CIRs)/high-speed streams (HSSs) not followed by HILDCAAs, and events of AE comprised in global intense geomagnetic disturbances). It is worth noting that we also study active but non-HILDCAA intervals. Examining the geomagnetic AE index, we apply a dynamics analysis composed of the phase space, the recurrence plot (RP), and the recurrence quantification analysis (RQA) methods. As a result, the quantification finds two distinct clusterings of the dynamical behaviours occurring in the interplanetary medium: one regarding a geomagnetically quiet condition regime and the other regarding an interplanetary activity regime. Furthermore, the HILDCAAs seem unique events regarding a visible, intense manifestations of interplanetary Alfvénic waves; however, they are similar to the other kinds of conditions regarding a dynamical signature (based on RQA), because it is involved in the same complex mechanism of generating geomagnetic disturbances. Also, by characterizing the proper conditions of transitions from quiescent conditions to weaker geomagnetic disturbances inside the magnetosphere and ionosphere system, the RQA method indicates clearly the two fundamental dynamics (geomagnetically quiet intervals and HILDCAA events) to be evaluated with magneto-hydrodynamics simulations to understand better the critical processes related to energy and particle transfer into the magnetosphere-ionosphere system. Finally, with this work, we have also reinforced the potential applicability of the RQA method for

  8. Investigation of 2‐stage meta‐analysis methods for joint longitudinal and time‐to‐event data through simulation and real data application

    PubMed Central

    Tudur Smith, Catrin; Gueyffier, François; Kolamunnage‐Dona, Ruwanthi

    2017-01-01

    Background Joint modelling of longitudinal and time‐to‐event data is often preferred over separate longitudinal or time‐to‐event analyses as it can account for study dropout, error in longitudinally measured covariates, and correlation between longitudinal and time‐to‐event outcomes. The joint modelling literature focuses mainly on the analysis of single studies with no methods currently available for the meta‐analysis of joint model estimates from multiple studies. Methods We propose a 2‐stage method for meta‐analysis of joint model estimates. These methods are applied to the INDANA dataset to combine joint model estimates of systolic blood pressure with time to death, time to myocardial infarction, and time to stroke. Results are compared to meta‐analyses of separate longitudinal or time‐to‐event models. A simulation study is conducted to contrast separate versus joint analyses over a range of scenarios. Results Using the real dataset, similar results were obtained by using the separate and joint analyses. However, the simulation study indicated a benefit of use of joint rather than separate methods in a meta‐analytic setting where association exists between the longitudinal and time‐to‐event outcomes. Conclusions Where evidence of association between longitudinal and time‐to‐event outcomes exists, results from joint models over standalone analyses should be pooled in 2‐stage meta‐analyses. PMID:29250814

  9. Geostationary Coastal and Air Pollution Events (GEO-CAPE) Sensitivity Analysis Experiment

    NASA Technical Reports Server (NTRS)

    Lee, Meemong; Bowman, Kevin

    2014-01-01

    Geostationary Coastal and Air pollution Events (GEO-CAPE) is a NASA decadal survey mission to be designed to provide surface reflectance at high spectral, spatial, and temporal resolutions from a geostationary orbit necessary for studying regional-scale air quality issues and their impact on global atmospheric composition processes. GEO-CAPE's Atmospheric Science Questions explore the influence of both gases and particles on air quality, atmospheric composition, and climate. The objective of the GEO-CAPE Observing System Simulation Experiment (OSSE) is to analyze the sensitivity of ozone to the global and regional NOx emissions and improve the science impact of GEO-CAPE with respect to the global air quality. The GEO-CAPE OSSE team at Jet propulsion Laboratory has developed a comprehensive OSSE framework that can perform adjoint-sensitivity analysis for a wide range of observation scenarios and measurement qualities. This report discusses the OSSE framework and presents the sensitivity analysis results obtained from the GEO-CAPE OSSE framework for seven observation scenarios and three instrument systems.

  10. The May 17, 2012 Solar Event: Back-Tracing Analysis and Flux Reconstruction with PAMELA

    NASA Technical Reports Server (NTRS)

    Bruno, A.; Adriani, O.; Barbarino, G. C.; Bazilevskaya, G. A.; Bellotti, R.; Boezio, M.; Bogomolov, E. A.; Bongi, M.; Bonvicini, V.; Bottai, S.; hide

    2016-01-01

    The PAMELA space experiment is providing first direct observations of Solar Energetic Particles (SEPs) with energies from about 80 MeV to several GeV in near-Earth orbit, bridging the low energy measurements by other spacecrafts and the Ground Level Enhancement (GLE) data by the worldwide network of neutron monitors. Its unique observational capabilities include the possibility of measuring the flux angular distribution and thus investigating possible anisotropies associated to SEP events. The analysis is supported by an accurate back-tracing simulation based on a realistic description of the Earth's magnetosphere, which is exploited to estimate the SEP energy spectra as a function of the asymptotic direction of arrival with respect to the Interplanetary Magnetic Field (IMF). In this work we report the results for the May 17, 2012 event.

  11. Event reweighting with the NuWro neutrino interaction generator

    NASA Astrophysics Data System (ADS)

    Pickering, Luke; Stowell, Patrick; Sobczyk, Jan

    2017-09-01

    Event reweighting has been implemented in the NuWro neutrino event generator for a number of free theory parameters in the interaction model. Event reweighting is a key analysis technique, used to efficiently study the effect of neutrino interaction model uncertainties. This opens up the possibility for NuWro to be used as a primary event generator by experimental analysis groups. A preliminary model tuning to ANL and BNL data of quasi-elastic and single pion production events was performed to validate the reweighting engine.

  12. Automated Detection of Events of Scientific Interest

    NASA Technical Reports Server (NTRS)

    James, Mark

    2007-01-01

    A report presents a slightly different perspective of the subject matter of Fusing Symbolic and Numerical Diagnostic Computations (NPO-42512), which appears elsewhere in this issue of NASA Tech Briefs. Briefly, the subject matter is the X-2000 Anomaly Detection Language, which is a developmental computing language for fusing two diagnostic computer programs one implementing a numerical analysis method, the other implementing a symbolic analysis method into a unified event-based decision analysis software system for real-time detection of events. In the case of the cited companion NASA Tech Briefs article, the contemplated events that one seeks to detect would be primarily failures or other changes that could adversely affect the safety or success of a spacecraft mission. In the case of the instant report, the events to be detected could also include natural phenomena that could be of scientific interest. Hence, the use of X- 2000 Anomaly Detection Language could contribute to a capability for automated, coordinated use of multiple sensors and sensor-output-data-processing hardware and software to effect opportunistic collection and analysis of scientific data.

  13. Analysis of Infrequent (Quasi-Decadal) Large Groundwater Recharge Events: A Case Study for Northern Utah, United States

    NASA Astrophysics Data System (ADS)

    Masbruch, M.; Rumsey, C.; Gangopadhyay, S.; Susong, D.; Pruitt, T.

    2015-12-01

    There has been a considerable amount of research linking climatic variability to hydrologic responses in arid and semi-arid regions such as the western United States. Although much effort has been spent to assess and predict changes in surface-water resources, little has been done to understand how climatic events and changes affect groundwater resources. This study focuses on quantifying the effects of large quasi-decadal groundwater recharge events on groundwater in the northern Utah portion of the Great Basin for the period 1960 to 2013. Groundwater-level monitoring data were analyzed with climatic data to characterize climatic conditions and frequency of these large recharge events. Using observed water-level changes and multivariate analysis, five large groundwater recharge events were identified within the study area and period, with a frequency of about 11 to 13 years. These events were generally characterized as having above-average annual precipitation and snow water equivalent and below-average seasonal temperatures, especially during the spring (April through June). Existing groundwater flow models for several basins within the study area were used to quantify changes in groundwater storage from these events. Simulated groundwater storage increases per basin from a single event ranged from about 115 Mm3 (93,000 acre-feet) to 205 Mm3 (166,000 acre-ft). Extrapolating these amounts over the entire northern Great Basin indicates that even a single large quasi-decadal recharge event could result in billions of cubic meters (millions of acre-feet) of groundwater recharge. Understanding the role of these large quasi-decadal recharge events in replenishing aquifers and sustaining water supplies is crucial for making informed water management decisions.

  14. Kickoff to Conflict: A Sequence Analysis of Intra-State Conflict-Preceding Event Structures

    PubMed Central

    D'Orazio, Vito; Yonamine, James E.

    2015-01-01

    While many studies have suggested or assumed that the periods preceding the onset of intra-state conflict are similar across time and space, few have empirically tested this proposition. Using the Integrated Crisis Early Warning System's domestic event data in Asia from 1998–2010, we subject this proposition to empirical analysis. We code the similarity of government-rebel interactions in sequences preceding the onset of intra-state conflict to those preceding further periods of peace using three different metrics: Euclidean, Levenshtein, and mutual information. These scores are then used as predictors in a bivariate logistic regression to forecast whether we are likely to observe conflict in neither, one, or both of the states. We find that our model accurately classifies cases where both sequences precede peace, but struggles to distinguish between cases in which one sequence escalates to conflict and where both sequences escalate to conflict. These findings empirically suggest that generalizable patterns exist between event sequences that precede peace. PMID:25951105

  15. Numerical analysis of seismic events distributions on the planetary scale and celestial bodies astrometrical parameters

    NASA Astrophysics Data System (ADS)

    Bulatova, Dr.

    2012-04-01

    Modern research in the domains of Earth sciences is developing from the descriptions of each individual natural phenomena to the systematic complex research in interdisciplinary areas. For studies of its kind in the form numerical analysis of three-dimensional (3D) systems, the author proposes space-time Technology (STT), based on a Ptolemaic geocentric system, consist of two modules, each with its own coordinate system: (1) - 3D model of a Earth, the coordinates of which provides databases of the Earth's events (here seismic), and (2) - a compact model of the relative motion of celestial bodies in space - time on Earth known as the "Method of a moving source" (MDS), which was developed in MDS (Bulatova, 1998-2000) for the 3D space. Module (2) was developed as a continuation of the geocentric Ptolemaic system of the world, built on the astronomical parameters heavenly bodies. Based on the aggregation data of Space and Earth Sciences, systematization, and cooperative analysis, this is an attempt to establish a cause-effect relationship between the position of celestial bodies (Moon, Sun) and Earth's seismic events.

  16. Biological event composition

    PubMed Central

    2012-01-01

    Background In recent years, biological event extraction has emerged as a key natural language processing task, aiming to address the information overload problem in accessing the molecular biology literature. The BioNLP shared task competitions have contributed to this recent interest considerably. The first competition (BioNLP'09) focused on extracting biological events from Medline abstracts from a narrow domain, while the theme of the latest competition (BioNLP-ST'11) was generalization and a wider range of text types, event types, and subject domains were considered. We view event extraction as a building block in larger discourse interpretation and propose a two-phase, linguistically-grounded, rule-based methodology. In the first phase, a general, underspecified semantic interpretation is composed from syntactic dependency relations in a bottom-up manner. The notion of embedding underpins this phase and it is informed by a trigger dictionary and argument identification rules. Coreference resolution is also performed at this step, allowing extraction of inter-sentential relations. The second phase is concerned with constraining the resulting semantic interpretation by shared task specifications. We evaluated our general methodology on core biological event extraction and speculation/negation tasks in three main tracks of BioNLP-ST'11 (GENIA, EPI, and ID). Results We achieved competitive results in GENIA and ID tracks, while our results in the EPI track leave room for improvement. One notable feature of our system is that its performance across abstracts and articles bodies is stable. Coreference resolution results in minor improvement in system performance. Due to our interest in discourse-level elements, such as speculation/negation and coreference, we provide a more detailed analysis of our system performance in these subtasks. Conclusions The results demonstrate the viability of a robust, linguistically-oriented methodology, which clearly distinguishes

  17. Pediatric emergency department census during major sporting events.

    PubMed

    Kim, Tommy Y; Barcega, Besh B; Denmark, T Kent

    2012-11-01

    Our study attempted to evaluate the effects of major sporting events on the census of a pediatric emergency department (ED) in the United States specifically related to the National Football League Super Bowl, National Basketball Association (NBA) Finals, and Major League Baseball World Series. We performed a retrospective data analysis of our pediatric ED census on the number of visits during major sporting events over a 5-year period. Data during the same period 1 week after the major sporting event were collected for comparison as the control. We evaluated the medians of 2-hour increments around the event start time. Subgroup analysis was performed for games involving the local sporting teams. Our results showed no significant difference in ED census during the sporting events, except in the post 6 to 8 hours of the NBA finals. Subgroup analysis of the Los Angeles Lakers showed the same significant findings in the post 6 to 8 hours of the NBA finals. No major difference in pediatric ED census is observed during the most major sporting events in the United States.

  18. Learning from Adverse Events in Obstetrics: Is a Standardized Computer Tool an Effective Strategy for Root Cause Analysis?

    PubMed

    Murray-Davis, Beth; McDonald, Helen; Cross-Sudworth, Fiona; Ahmed, Rashid; Simioni, Julia; Dore, Sharon; Marrin, Michael; DeSantis, Judy; Leyland, Nicholas; Gardosi, Jason; Hutton, Eileen; McDonald, Sarah

    2015-08-01

    Adverse events occur in up to 10% of obstetric cases, and up to one half of these could be prevented. Case reviews and root cause analysis using a structured tool may help health care providers to learn from adverse events and to identify trends and recurring systems issues. We sought to establish the reliability of a root cause analysis computer application called Standardized Clinical Outcome Review (SCOR). We designed a mixed methods study to evaluate the effectiveness of the tool. We conducted qualitative content analysis of five charts reviewed by both the traditional obstetric quality assurance methods and the SCOR tool. We also determined inter-rater reliability by having four health care providers review the same five cases using the SCOR tool. The comparative qualitative review revealed that the traditional quality assurance case review process used inconsistent language and made serious, personalized recommendations for those involved in the case. In contrast, the SCOR review provided a consistent format for recommendations, a list of action points, and highlighted systems issues. The mean percentage agreement between the four reviewers for the five cases was 75%. The different health care providers completed data entry and assessment of the case in a similar way. Missing data from the chart and poor wording of questions were identified as issues affecting percentage agreement. The SCOR tool provides a standardized, objective, obstetric-specific tool for root cause analysis that may improve identification of risk factors and dissemination of action plans to prevent future events.

  19. Spatio-Temporal Information Analysis of Event-Related BOLD Responses

    PubMed Central

    Alpert, Galit Fuhrmann; Handwerker, Dan; Sun, Felice T.; D’Esposito, Mark; Knight, Robert T.

    2009-01-01

    A new approach for analysis of event related fMRI (BOLD) signals is proposed. The technique is based on measures from information theory and is used both for spatial localization of task related activity, as well as for extracting temporal information regarding the task dependent propagation of activation across different brain regions. This approach enables whole brain visualization of voxels (areas) most involved in coding of a specific task condition, the time at which they are most informative about the condition, as well as their average amplitude at that preferred time. The approach does not require prior assumptions about the shape of the hemodynamic response function (HRF), nor about linear relations between BOLD response and presented stimuli (or task conditions). We show that relative delays between different brain regions can also be computed without prior knowledge of the experimental design, suggesting a general method that could be applied for analysis of differential time delays that occur during natural, uncontrolled conditions. Here we analyze BOLD signals recorded during performance of a motor learning task. We show that during motor learning, the BOLD response of unimodal motor cortical areas precedes the response in higher-order multimodal association areas, including posterior parietal cortex. Brain areas found to be associated with reduced activity during motor learning, predominantly in prefrontal brain regions, are informative about the task typically at significantly later times. PMID:17188515

  20. Joint models for longitudinal and time-to-event data: a review of reporting quality with a view to meta-analysis.

    PubMed

    Sudell, Maria; Kolamunnage-Dona, Ruwanthi; Tudur-Smith, Catrin

    2016-12-05

    Joint models for longitudinal and time-to-event data are commonly used to simultaneously analyse correlated data in single study cases. Synthesis of evidence from multiple studies using meta-analysis is a natural next step but its feasibility depends heavily on the standard of reporting of joint models in the medical literature. During this review we aim to assess the current standard of reporting of joint models applied in the literature, and to determine whether current reporting standards would allow or hinder future aggregate data meta-analyses of model results. We undertook a literature review of non-methodological studies that involved joint modelling of longitudinal and time-to-event medical data. Study characteristics were extracted and an assessment of whether separate meta-analyses for longitudinal, time-to-event and association parameters were possible was made. The 65 studies identified used a wide range of joint modelling methods in a selection of software. Identified studies concerned a variety of disease areas. The majority of studies reported adequate information to conduct a meta-analysis (67.7% for longitudinal parameter aggregate data meta-analysis, 69.2% for time-to-event parameter aggregate data meta-analysis, 76.9% for association parameter aggregate data meta-analysis). In some cases model structure was difficult to ascertain from the published reports. Whilst extraction of sufficient information to permit meta-analyses was possible in a majority of cases, the standard of reporting of joint models should be maintained and improved. Recommendations for future practice include clear statement of model structure, of values of estimated parameters, of software used and of statistical methods applied.

  1. Use of Bayesian event trees in semi-quantitative volcano eruption forecasting and hazard analysis

    NASA Astrophysics Data System (ADS)

    Wright, Heather; Pallister, John; Newhall, Chris

    2015-04-01

    Use of Bayesian event trees to forecast eruptive activity during volcano crises is an increasingly common practice for the USGS-USAID Volcano Disaster Assistance Program (VDAP) in collaboration with foreign counterparts. This semi-quantitative approach combines conceptual models of volcanic processes with current monitoring data and patterns of occurrence to reach consensus probabilities. This approach allows a response team to draw upon global datasets, local observations, and expert judgment, where the relative influence of these data depends upon the availability and quality of monitoring data and the degree to which the volcanic history is known. The construction of such event trees additionally relies upon existence and use of relevant global databases and documented past periods of unrest. Because relevant global databases may be underpopulated or nonexistent, uncertainty in probability estimations may be large. Our 'hybrid' approach of combining local and global monitoring data and expert judgment facilitates discussion and constructive debate between disciplines: including seismology, gas geochemistry, geodesy, petrology, physical volcanology and technology/engineering, where difference in opinion between response team members contributes to definition of the uncertainty in the probability estimations. In collaboration with foreign colleagues, we have created event trees for numerous areas experiencing volcanic unrest. Event trees are created for a specified time frame and are updated, revised, or replaced as the crisis proceeds. Creation of an initial tree is often prompted by a change in monitoring data, such that rapid assessment of probability is needed. These trees are intended as a vehicle for discussion and a way to document relevant data and models, where the target audience is the scientists themselves. However, the probabilities derived through the event-tree analysis can also be used to help inform communications with emergency managers and the

  2. Advanced Mechanistic 3D Spatial Modeling and Analysis Methods to Accurately Represent Nuclear Facility External Event Scenarios

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sezen, Halil; Aldemir, Tunc; Denning, R.

    Probabilistic risk assessment of nuclear power plants initially focused on events initiated by internal faults at the plant, rather than external hazards including earthquakes and flooding. Although the importance of external hazards risk analysis is now well recognized, the methods for analyzing low probability external hazards rely heavily on subjective judgment of specialists, often resulting in substantial conservatism. This research developed a framework to integrate the risk of seismic and flooding events using realistic structural models and simulation of response of nuclear structures. The results of four application case studies are presented.

  3. Cost-effectiveness analysis of applying the Cholesterol and Recurrent Events (CARE) study protocol in Hong Kong.

    PubMed

    Chau, J; Cheung, B M; McGhee, S M; Lauder, I J; Lau, C P; Kumana, C R

    2001-12-01

    To determine the cost-effectiveness of secondary prevention with pravastatin in Hong Kong patients with coronary heart disease and average cholesterol levels. Cost-effectiveness analysis based on published results of the CARE study. Men and women post-myocardial infarction with average cholesterol levels. Cost-effectiveness analysis: cost per life saved, cost per fatal or non-fatal coronary event prevented, cost per procedure prevented, and cost per fatal or non-fatal stroke prevented. Cost-utility analysis: gross cost and net cost per quality-adjusted life year gained calculated using two alternative models. Cost per life saved or death prevented was HK$4,442,350 (non-discounted); cost per fatal or non-fatal cardiac event prevented HK$1,146,413; cost per procedure prevented HK$732,759; and cost per fatal or non-fatal stroke prevented HK$2,961,566. Net cost per quality adjusted life year gained was HK$73,218 and HK$65,280 non-discounted, respectively using the two alternative models. The results of this study can assist in prioritising the use of health care resources in Hong Kong but should be considered alongside the benefits and costs of alternative interventions for coronary heart disease.

  4. P2Y12 Polymorphisms and the Risk of Adverse Clinical Events in Patients Treated with Clopidogrel: A Meta-Analysis.

    PubMed

    Zhao, Kun; Yang, Ming; Lu, Yanxia; Sun, Shusen; Li, Wei; Li, Xingang; Zhao, Zhigang

    2018-05-23

    Some studies have reported an association between P2Y12 gene polymorphisms and clopidogrel adverse outcomes with inconsistent results. We aimed to explore the relationship between P2Y12 polymorphisms and the risk of adverse clinical events in patients treated with clopidogrel through a meta-analysis. A systematic search of PubMed, Web of Science and the Cochrane Library was conducted. Retrieved articles were comprehensively reviewed and eligible studies were included, and the relevant data was extracted for this meta-analysis. All statistical tests were performed by the Review Manager 5.3 software. A total of 14 studies involving 8,698 patients were included. In the Han Chinese population, ischemic events were associated with P2Y12 T744C polymorphism in the CC vs TT+CT genetic model (OR=3.32, 95%CI=1.62-6.82, P =0.001), and the events were associated with P2Y12 C34T polymorphism in the TT+TC vs CC genetic model (OR=1.70, 95%CI=1.22-2.36, P =0.002). However, ischemic events were not related to P2Y12 G52T polymorphism (TT+TG vs GG: OR=1.13, 95%CI=0.76-1.68, P =0.56; TT vs GG+TG: OR=2.02, 95%CI=0.65-6.28, P =0.22). The associations between the P2Y12 polymorphism and ischemic events were not significant in T744C, G52T and C34T genotype for another subgroup of the Caucasian population ( P >0.05). Only two studies referring to bleeding events were included in this analysis of C34T polymorphism, and no significant association was found (TT+TC vs CC: OR=1.07, 95%CI=0.37-3.15, P =0.90). In the Caucasian population, P2Y12 gene polymorphisms are not associated with clinical events. However, in the Chinese Han population, P2Y12 T744C and C34T polymorphisms are significantly associated with adverse clinical events. © Georg Thieme Verlag KG Stuttgart · New York.

  5. Meta-analysis and psychophysiology: A tutorial using depression and action-monitoring event-related potentials.

    PubMed

    Moran, Tim P; Schroder, Hans S; Kneip, Chelsea; Moser, Jason S

    2017-01-01

    Meta-analyses are regularly used to quantitatively integrate the findings of a field, assess the consistency of an effect and make decisions based on extant research. The current article presents an overview and step-by-step tutorial of meta-analysis aimed at psychophysiological researchers. We also describe best-practices and steps that researchers can take to facilitate future meta-analysis in their sub-discipline. Lastly, we illustrate each of the steps by presenting a novel meta-analysis on the relationship between depression and action-monitoring event-related potentials - the error-related negativity (ERN) and the feedback negativity (FN). This meta-analysis found that the literature on depression and the ERN is contaminated by publication bias. With respect to the FN, the meta-analysis found that depression does predict the magnitude of the FN; however, this effect was dependent on the type of task used by the study. Copyright © 2016 Elsevier B.V. All rights reserved.

  6. Risk of Death in Infants Who Have Experienced a Brief Resolved Unexplained Event: A Meta-Analysis.

    PubMed

    Brand, Donald A; Fazzari, Melissa J

    2018-06-01

    To estimate an upper bound on the risk of death after a brief resolved unexplained event (BRUE), a sudden alteration in an infant's breathing, color, tone, or responsiveness, previously labeled "apparent life-threatening event" (ALTE). The meta-analysis incorporated observational studies of patients with ALTE that included data on in-hospital and post-discharge deaths with at least 1 week of follow-up after hospital discharge. Pertinent studies were identified from a published review of the literature from 1970 through 2014 and a supplementary PubMed query through February 2017. The 12 included studies (n = 3005) reported 12 deaths, of which 8 occurred within 4 months of the event. Applying a Poisson-normal random effects model to the 8 proximate deaths using a 4-month time horizon yielded a post-ALTE mortality rate of about 1 in 800, which constitutes an upper bound on the risk of death after a BRUE. This risk is about the same as the baseline risk of death during the first year of life. The meta-analysis therefore supports the return-home approach advocated in a recently published clinical practice guideline-not routine hospitalization-for BRUE patients who have been evaluated in the emergency department and determined to be at lower risk. Copyright © 2017 Elsevier Inc. All rights reserved.

  7. A systematic review and meta-analysis on herpes zoster and the risk of cardiac and cerebrovascular events.

    PubMed

    Erskine, Nathaniel; Tran, Hoang; Levin, Leonard; Ulbricht, Christine; Fingeroth, Joyce; Kiefe, Catarina; Goldberg, Robert J; Singh, Sonal

    2017-01-01

    Patients who develop herpes zoster or herpes zoster ophthalmicus may be at risk for cerebrovascular and cardiac complications. We systematically reviewed the published literature to determine the association between herpes zoster and its subtypes with the occurrence of cerebrovascular and cardiac events. Systematic searches of PubMed (MEDLINE), SCOPUS (Embase) and Google Scholar were performed in December 2016. Eligible studies were cohort, case-control, and self-controlled case-series examining the association between herpes zoster or subtypes of herpes zoster with the occurrence of cerebrovascular and cardiac events including stroke, transient ischemic attack, coronary heart disease, and myocardial infarction. Data on the occurrence of the examined events were abstracted. Odds ratios and their accompanying confidence intervals were estimated using random and fixed effects models with statistical heterogeneity estimated with the I2 statistic. Twelve studies examining 7.9 million patients up to 28 years after the onset of herpes zoster met our pre-defined eligibility criteria. Random and fixed effects meta-analyses showed that herpes zoster, type unspecified, and herpes zoster ophthalmicus were associated with a significantly increased risk of cerebrovascular events, without any evidence of statistical heterogeneity. Our meta-analysis also found a significantly increased risk of cardiac events associated with herpes zoster, type unspecified. Our results are consistent with the accumulating body of evidence that herpes zoster and herpes zoster ophthalmicus are significantly associated with cerebrovascular and cardiovascular events.

  8. The electrical heart axis and ST events in fetal monitoring: A post-hoc analysis following a multicentre randomised controlled trial.

    PubMed

    Vullings, Rik; Verdurmen, Kim M J; Hulsenboom, Alexandra D J; Scheffer, Stephanie; de Lau, Hinke; Kwee, Anneke; Wijn, Pieter F F; Amer-Wåhlin, Isis; van Laar, Judith O E H; Oei, S Guid

    2017-01-01

    Reducing perinatal morbidity and mortality is one of the major challenges in modern health care. Analysing the ST segment of the fetal electrocardiogram was thought to be the breakthrough in fetal monitoring during labour. However, its implementation in clinical practice yields many false alarms and ST monitoring is highly dependent on cardiotocogram assessment, limiting its value for the prediction of fetal distress during labour. This study aims to evaluate the relation between physiological variations in the orientation of the fetal electrical heart axis and the occurrence of ST events. A post-hoc analysis was performed following a multicentre randomised controlled trial, including 1097 patients from two participating centres. All women were monitored with ST analysis during labour. Cases of fetal metabolic acidosis, poor signal quality, missing blood gas analysis, and congenital heart disease were excluded. The orientation of the fetal electrical heart axis affects the height of the initial T/QRS baseline, and therefore the incidence of ST events. We grouped tracings with the same initial baseline T/QRS value. We depicted the number of ST events as a function of the initial baseline T/QRS value with a linear regression model. A significant increment of ST events was observed with increasing height of the initial T/QRS baseline, irrespective of the fetal condition; correlation coefficient 0.63, p<0.001. The most frequent T/QRS baseline is 0.12. The orientation of the fetal electrical heart axis and accordingly the height of the initial T/QRS baseline should be taken into account in fetal monitoring with ST analysis.

  9. Searching for Effective Training Solutions for Firefighting: The Analysis of Emergency Responses and Line of Duty Death Reports for Low Frequency, High Risk Events

    DTIC Science & Technology

    2017-09-01

    whether emergency incidents connected to low frequency and high risk events contain sufficient warning signs or indicators of imminent catastrophic... high risk events contain sufficient warning signs or indicators of imminent catastrophic events, if firefighters could identify them, and if there...EFFECTIVE TRAINING SOLUTIONS FOR FIREFIGHTING: THE ANALYSIS OF EMERGENCY RESPONSES AND LINE OF DUTY DEATH REPORTS FOR LOW FREQUENCY, HIGH RISK EVENTS

  10. Prediction of collision events: an EEG coherence analysis.

    PubMed

    Spapé, Michiel M; Serrien, Deborah J

    2011-05-01

    A common daily-life task is the interaction with moving objects for which prediction of collision events is required. To evaluate the sources of information used in this process, this EEG study required participants to judge whether two moving objects would collide with one another or not. In addition, the effect of a distractor object is evaluated. The measurements included the behavioural decision time and accuracy, eye movement fixation times, and the neural dynamics which was determined by means of EEG coherence, expressing functional connectivity between brain areas. Collision judgment involved widespread information processing across both hemispheres. When a distractor object was present, task-related activity was increased whereas distractor activity induced modulation of local sensory processing. Also relevant were the parietal regions communicating with bilateral occipital and midline areas and a left-sided sensorimotor circuit. Besides visual cues, cognitive and strategic strategies are used to establish a decision of events in time. When distracting information is introduced into the collision judgment process, it is managed at different processing levels and supported by distinct neural correlates. These data shed light on the processing mechanisms that support judgment of collision events; an ability that implicates higher-order decision-making. Copyright © 2011 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.

  11. Skylab ATM/S-056 X-ray event analyzer: Instrument description, parameter determination, and analysis example (15 June 1973 1B/M3 flare)

    NASA Technical Reports Server (NTRS)

    Wilson, R. M.

    1976-01-01

    The Skylab ATM/S-056 X-Ray Event Analyzer, part of an X-ray telescope experiment, is described. The techniques employed in the analysis of its data to determine electron temperatures and emission measures are reviewed. The analysis of a sample event - the 15 June 1973 1B/M3 flare - is performed. Comparison of the X-Ray Event Analyzer data with that of the SolRad 9 observations indicates that the X-Ray Event Analyzer accurately monitored the sun's 2.5 to 7.25 A X-ray emission and to a lesser extent the 6.1 to 20 A emission. A mean average peak temperature of 15 million K at 1,412 UT and a mean average peak electron density (assuming a flare volume of 10 to the 13 power cu km) of 27 million/cu mm at 1,416 to 1,417 UT are deduced for the event. The X-Ray Event Analyzer data, having a 2.5 s time resolution, should be invaluable in comparisons with other high-time resolution data (e.g., radio bursts).

  12. Analysis of an ordinary bedload transport event in a mountain torrent (Rio Vanti, Verona, Italy)

    NASA Astrophysics Data System (ADS)

    Pastorello, Roberta; D'Agostino, Vincenzo

    2016-04-01

    The correct simulation of the sediment-transport response of mountain torrents both for extreme and ordinary flood events is a fundamental step to understand the process, but also to drive proper decisions on the protection works. The objective of this research contribution is to reconstruct the 'ordinary' flood event with the associated sediment-graph of a flood that caused on the 14th of October, 2014 the formation of a little debris cone (about 200-210 m3) at the junction between the 'Rio Vanti' torrent catchment and the 'Selva di Progno' torrent (Veneto Region, Prealps, Verona, Italy). To this purpose, it is important to notice that a great part of equations developed for the computation of the bedload transport capacity, like for example that of Schoklitsch (1962) or Smart and Jaeggi (1983), are focused on extraordinary events heavily affecting the river-bed armour. These formulas do not provide reliable results if used on events, like the one under analysis, not too far from the bankfull conditions. The Rio Vanti event was characterized by a total rainfall depth of 36.2 mm and a back-calculated peak discharge of 6.12 m3/s with a return period of 1-2 years. The classical equations to assess the sediment transport capacity overestimate the total volume of the event of several orders of magnitude. By the consequence, the following experimental bedload transport equation has been applied (D'Agostino and Lenzi, 1999), which is valid for ordinary flood events (q: unit water discharge; qc: unit discharge of bedload transport initiation; qs: unit bedload rate; S: thalweg slope): -qs-˜= 0,04ṡ(q- qc) S3/2 In particular, starting from the real rainfall data, the hydrograph and the sediment-graph have been reconstructed. Then, comparing the total volume calculated via the above cited equation to the real volume estimated using DoD techniques on post-event photogrammetric survey, a very satisfactory agreement has been obtained. The result further supports the thesis

  13. A review for identification of initiating events in event tree development process on nuclear power plants

    NASA Astrophysics Data System (ADS)

    Riyadi, Eko H.

    2014-09-01

    Initiating event is defined as any event either internal or external to the nuclear power plants (NPPs) that perturbs the steady state operation of the plant, if operating, thereby initiating an abnormal event such as transient or loss of coolant accident (LOCA) within the NPPs. These initiating events trigger sequences of events that challenge plant control and safety systems whose failure could potentially lead to core damage or large early release. Selection for initiating events consists of two steps i.e. first step, definition of possible events, such as by evaluating a comprehensive engineering, and by constructing a top level logic model. Then the second step, grouping of identified initiating event's by the safety function to be performed or combinations of systems responses. Therefore, the purpose of this paper is to discuss initiating events identification in event tree development process and to reviews other probabilistic safety assessments (PSA). The identification of initiating events also involves the past operating experience, review of other PSA, failure mode and effect analysis (FMEA), feedback from system modeling, and master logic diagram (special type of fault tree). By using the method of study for the condition of the traditional US PSA categorization in detail, could be obtained the important initiating events that are categorized into LOCA, transients and external events.

  14. Seismic event near Jarocin (Poland)

    NASA Astrophysics Data System (ADS)

    Lizurek, Grzegorz; Plesiewicz, Beata; Wiejacz, Paweł; Wiszniowski, Jan; Trojanowski, Jacek

    2013-02-01

    The earthquake of magnitude M L = 3:8 (EMSC) took place on Friday, 6 January 2012, north-east of the town of Jarocin in Wielkopolska Region, Poland. The only historical information about past earthquakes in the region was found in the diary from 1824; apart of it, there was a seismic event noticed in the vicinity of Wielkopolska in 1606 (Pagaczewski 1982). The scope of this paper is to describe the 6 January 2012 event in view of instrumental seismology, macroseismic data analysis and known tectonics of the region, which should be useful in future seismic hazard analysis of Poland.

  15. Rain-on-snow Events in Southwestern British Columbia: A Long-term Analysis of Meteorological Conditions and Snowpack Response

    NASA Astrophysics Data System (ADS)

    Trubilowicz, J. W.; Moore, D.

    2015-12-01

    Snowpack dynamics and runoff generation in coastal mountain regions are complicated by rain-on-snow (ROS) events. During major ROS events associated with warm, moist air and strong winds, turbulent heat fluxes can produce substantial melt to supplement rainfall, but previous studies suggest this may not be true for smaller, more frequent events. The internal temperature and water content of the snowpack are also expected to influence runoff generation during ROS events: a cold snowpack with no liquid water content will have the ability to store significant amounts of rainfall, whereas a 'ripe' snowpack may begin to melt and generate outflow with little rain input. However, it is not well understood how antecedent snowpack conditions and energy fluxes differ between ROS events that cause large runoff events and those that do not, in large part because major flood-producing ROS events occur infrequently, and thus are often not sampled during short-term research projects. To generate greater understanding of runoff generation over the spectrum of ROS magnitudes and frequencies, we analyzed data from Automated Snow Pillow (ASP) sites, which record hourly air temperature, precipitation and snowpack water equivalent and offer up to several decades of data at each site. We supplemented the ASP data with output from the North American Regional Reanalysis (NARR) product to support point scale snow modeling for 335 ROS event records from six ASP sites in southwestern BC from 2003 to 2013. Our analysis reconstructed the weather conditions, surface energy exchanges, internal mass and energy states of the snowpack, and generation of snow melt and water available for runoff (WAR) for each ROS event. Results indicate that WAR generation during large events is largely independent of the snowpack conditions, but for smaller events, the antecedent snow conditions play a significant role in either damping or enhancing WAR generation.

  16. Shift work and vascular events: systematic review and meta-analysis.

    PubMed

    Vyas, Manav V; Garg, Amit X; Iansavichus, Arthur V; Costella, John; Donner, Allan; Laugsand, Lars E; Janszky, Imre; Mrkobrada, Marko; Parraga, Grace; Hackam, Daniel G

    2012-07-26

    To synthesise the association of shift work with major vascular events as reported in the literature. Systematic searches of major bibliographic databases, contact with experts in the field, and review of reference lists of primary articles, review papers, and guidelines. Observational studies that reported risk ratios for vascular morbidity, vascular mortality, or all cause mortality in relation to shift work were included; control groups could be non-shift ("day") workers or the general population. Study quality was assessed with the Downs and Black scale for observational studies. The three primary outcomes were myocardial infarction, ischaemic stroke, and any coronary event. Heterogeneity was measured with the I(2) statistic and computed random effects models. 34 studies in 2,011,935 people were identified. Shift work was associated with myocardial infarction (risk ratio 1.23, 95% confidence interval 1.15 to 1.31; I(2)=0) and ischaemic stroke (1.05, 1.01 to 1.09; I(2)=0). Coronary events were also increased (risk ratio 1.24, 1.10 to 1.39), albeit with significant heterogeneity across studies (I(2)=85%). Pooled risk ratios were significant for both unadjusted analyses and analyses adjusted for risk factors. All shift work schedules with the exception of evening shifts were associated with a statistically higher risk of coronary events. Shift work was not associated with increased rates of mortality (whether vascular cause specific or overall). Presence or absence of adjustment for smoking and socioeconomic status was not a source of heterogeneity in the primary studies. 6598 myocardial infarctions, 17,359 coronary events, and 1854 ischaemic strokes occurred. On the basis of the Canadian prevalence of shift work of 32.8%, the population attributable risks related to shift work were 7.0% for myocardial infarction, 7.3% for all coronary events, and 1.6% for ischaemic stroke. Shift work is associated with vascular events, which may have implications for public

  17. Time-to-Event Analysis of Individual Variables Associated with Nursing Students' Academic Failure: A Longitudinal Study

    ERIC Educational Resources Information Center

    Dante, Angelo; Fabris, Stefano; Palese, Alvisa

    2013-01-01

    Empirical studies and conceptual frameworks presented in the extant literature offer a static imagining of academic failure. Time-to-event analysis, which captures the dynamism of individual factors, as when they determine the failure to properly tailor timely strategies, impose longitudinal studies which are still lacking within the field. The…

  18. Monitoring As A Helpful Means In Forensic Analysis Of Dams Static Instability Events

    NASA Astrophysics Data System (ADS)

    Solimene, Pellegrino

    2013-04-01

    Monitoring is a means of controlling the behavior of a structure, which during its operational life is subject to external actions as ordinary loading conditions and disturbing ones; these factors overlap with the random manner defined by the statistical parameter of the return period. The analysis of the monitoring data is crucial to gain a reasoned opinion on the reliability of the structure and its components, and also allows to identify, in the overall operational scenario, the time when preparing interventions aimed at maintaining the optimum levels of functionality and safety. The concept of monitoring in terms of prevention is coupled with the activity of Forensic Engineer who, by Judiciary appointment for the occurrence of an accident, turns its experience -the "Scientific knowledge"- in an "inverse analysis" in which he summed up the results of a survey, which also draws on data sets arising in the course of the constant control of the causes and effects, so to determine the correlations between these factors. His activity aims at giving a contribution to the identification of the typicality of an event, which represents, together with "causal link" between the conduct and events and contra-juridical, the factors judging if there an hypothesis of crime, and therefore liable according to law. In Italy there are about 10,000 dams of varying sizes, but only a small portion of them are considered "large dams" and subjected to a rigorous program of regular inspections and monitoring, in application of specific rules. The rest -"small" dams, conventionally defined as such by the standard, but not for the impact on the area- is affected by a heterogeneous response from the local authorities entrusted with this task: there is therefore a high potential risk scenario, as determined by the presence of not completely controlled structures that insist even on areas heavily populated. Risk can be traced back to acceptable levels if they were implemented with the

  19. Analysis of the variation of the 0°C isothermal altitude during rainfall events

    NASA Astrophysics Data System (ADS)

    Zeimetz, Fränz; Garcìa, Javier; Schaefli, Bettina; Schleiss, Anton J.

    2016-04-01

    In numerous countries of the world (USA, Canada, Sweden, Switzerland,…), the dam safety verifications for extreme floods are realized by referring to the so called Probable Maximum Flood (PMF). According to the World Meteorological Organization (WMO), this PMF is determined based on the PMP (Probable Maximum Precipitation). The PMF estimation is performed with a hydrological simulation model by routing the PMP. The PMP-PMF simulation is normally event based; therefore, if no further information is known, the simulation needs assumptions concerning the initial soil conditions such as saturation or snow cover. In addition, temperature series are also of interest for the PMP-PMF simulations. Temperature values can not only be deduced from temperature measurement but also using the temperature gradient method, the 0°C isothermal altitude can lead to temperature estimations on the ground. For practitioners, the usage of the isothermal altitude for referring to temperature is convenient and simpler because one value can give information over a large region under the assumption of a certain temperature gradient. The analysis of the evolution of the 0°C isothermal altitude during rainfall events is aimed here and based on meteorological soundings from the two sounding stations Payerne (CH) and Milan (I). Furthermore, hourly rainfall and temperature data are available from 110 pluviometers spread over the Swiss territory. The analysis of the evolution of the 0°C isothermal altitude is undertaken for different precipitation durations based on the meteorological measurements mentioned above. The results show that on average, the isothermal altitude tends to decrease during the rainfall events and that a correlation between the duration of the altitude loss and the duration of the rainfall exists. A significant difference in altitude loss is appearing when the soundings from Payerne and Milan are compared.

  20. Calibration, event reconstruction, data analysis, and limit calculation for the LUX dark matter experiment

    NASA Astrophysics Data System (ADS)

    Akerib, D. S.; Alsum, S.; Araújo, H. M.; Bai, X.; Bailey, A. J.; Balajthy, J.; Beltrame, P.; Bernard, E. P.; Bernstein, A.; Biesiadzinski, T. P.; Boulton, E. M.; Brás, P.; Byram, D.; Cahn, S. B.; Carmona-Benitez, M. C.; Chan, C.; Currie, A.; Cutter, J. E.; Davison, T. J. R.; Dobi, A.; Dobson, J. E. Y.; Druszkiewicz, E.; Edwards, B. N.; Faham, C. H.; Fallon, S. R.; Fan, A.; Fiorucci, S.; Gaitskell, R. J.; Gehman, V. M.; Genovesi, J.; Ghag, C.; Gilchriese, M. G. D.; Hall, C. R.; Hanhardt, M.; Haselschwardt, S. J.; Hertel, S. A.; Hogan, D. P.; Horn, M.; Huang, D. Q.; Ignarra, C. M.; Jacobsen, R. G.; Ji, W.; Kamdin, K.; Kazkaz, K.; Khaitan, D.; Knoche, R.; Larsen, N. A.; Lee, C.; Lenardo, B. G.; Lesko, K. T.; Lindote, A.; Lopes, M. I.; Manalaysay, A.; Mannino, R. L.; Marzioni, M. F.; McKinsey, D. N.; Mei, D.-M.; Mock, J.; Moongweluwan, M.; Morad, J. A.; Murphy, A. St. J.; Nehrkorn, C.; Nelson, H. N.; Neves, F.; O'Sullivan, K.; Oliver-Mallory, K. C.; Palladino, K. J.; Pease, E. K.; Reichhart, L.; Rhyne, C.; Shaw, S.; Shutt, T. A.; Silva, C.; Solmaz, M.; Solovov, V. N.; Sorensen, P.; Sumner, T. J.; Szydagis, M.; Taylor, D. J.; Taylor, W. C.; Tennyson, B. P.; Terman, P. A.; Tiedt, D. R.; To, W. H.; Tripathi, M.; Tvrznikova, L.; Uvarov, S.; Velan, V.; Verbus, J. R.; Webb, R. C.; White, J. T.; Whitis, T. J.; Witherell, M. S.; Wolfs, F. L. H.; Xu, J.; Yazdani, K.; Young, S. K.; Zhang, C.; LUX Collaboration

    2018-05-01

    The LUX experiment has performed searches for dark-matter particles scattering elastically on xenon nuclei, leading to stringent upper limits on the nuclear scattering cross sections for dark matter. Here, for results derived from 1.4 ×104 kg days of target exposure in 2013, details of the calibration, event-reconstruction, modeling, and statistical tests that underlie the results are presented. Detector performance is characterized, including measured efficiencies, stability of response, position resolution, and discrimination between electron- and nuclear-recoil populations. Models are developed for the drift field, optical properties, background populations, the electron- and nuclear-recoil responses, and the absolute rate of low-energy background events. Innovations in the analysis include in situ measurement of the photomultipliers' response to xenon scintillation photons, verification of fiducial mass with a low-energy internal calibration source, and new empirical models for low-energy signal yield based on large-sample, in situ calibrations.

  1. Hydroacoustic monitoring of a salt cavity: an analysis of precursory events of the collapse

    NASA Astrophysics Data System (ADS)

    Lebert, F.; Bernardie, S.; Mainsant, G.

    2011-09-01

    One of the main features of "post mining" research relates to available methods for monitoring mine-degradation processes that could directly threaten surface infrastructures. In this respect, GISOS, a French scientific interest group, is investigating techniques for monitoring the eventual collapse of underground cavities. One of the methods under investigation was monitoring the stability of a salt cavity through recording microseismic-precursor signals that may indicate the onset of rock failure. The data were recorded in a salt mine in Lorraine (France) when monitoring the controlled collapse of 2 000 000 m3 of rocks surrounding a cavity at 130 m depth. The monitoring in the 30 Hz to 3 kHz frequency range highlights the occurrence of events with high energy during periods of macroscopic movement, once the layers had ruptured; they appear to be the consequence of the post-rupture rock movements related to the intense deformation of the cavity roof. Moreover the analysis shows the presence of some interesting precursory signals before the cavity collapsed. They occurred a few hours before the failure phases, when the rocks were being weakened and damaged. They originated from the damaging and breaking process, when micro-cracks appear and then coalesce. From these results we expect that deeper signal analysis and statistical analysis on the complete event time distribution (several millions of files) will allow us to finalize a complete typology of each signal families and their relations with the evolution steps of the cavity over the five years monitoring.

  2. Leveraging Researcher Reflexivity to Consider a Classroom Event over Time: Reflexive Discourse Analysis of "What Counts"

    ERIC Educational Resources Information Center

    Anderson, Kate T.

    2017-01-01

    This article presents a reflexive and critical discourse analysis of classroom events that grew out of a cross-cultural partnership with a secondary school teacher in Singapore. I aim to illuminate how differences between researcher and teacher assumptions about what participation in classroom activities should look like came into high relief when…

  3. A framework for analysis of sentinel events in medical student education.

    PubMed

    Cohen, Daniel M; Clinchot, Daniel M; Werman, Howard A

    2013-11-01

    Although previous studies have addressed student factors contributing to dismissal or withdrawal from medical school for academic reasons, little information is available regarding institutional factors that may hinder student progress. The authors describe the development and application of a framework for sentinel event (SE) root cause analysis to evaluate cases in which students are dismissed or withdraw because of failure to progress in the medical school curriculum. The SE in medical student education (MSE) framework was piloted at the Ohio State University College of Medicine (OSUCOM) during 2010-2012. Faculty presented cases using the framework during academic oversight committee discussions. Nine SEs in MSE were presented using the framework. Major institution-level findings included the need for improved communication, documentation of cognitive and noncognitive (e.g., mental health) issues, clarification of requirements for remediation and fitness for duty, and additional psychological services. Challenges related to alternative and combined programs were identified as well. The OSUCOM undertook system changes based on the action plans developed through the discussions of these SEs. An SE analysis process appears to be a useful method for making system changes in response to institutional issues identified in evaluation of cases in which students fail to progress in the medical school curriculum. The authors plan to continue to refine the SE in MSE framework and analysis process. Next steps include assessing whether analysis using this framework yields improved student outcomes with universal applications for other institutions.

  4. Risk Analysis of Earth-Rock Dam Failures Based on Fuzzy Event Tree Method

    PubMed Central

    Fu, Xiao; Gu, Chong-Shi; Su, Huai-Zhi; Qin, Xiang-Nan

    2018-01-01

    Earth-rock dams make up a large proportion of the dams in China, and their failures can induce great risks. In this paper, the risks associated with earth-rock dam failure are analyzed from two aspects: the probability of a dam failure and the resulting life loss. An event tree analysis method based on fuzzy set theory is proposed to calculate the dam failure probability. The life loss associated with dam failure is summarized and refined to be suitable for Chinese dams from previous studies. The proposed method and model are applied to one reservoir dam in Jiangxi province. Both engineering and non-engineering measures are proposed to reduce the risk. The risk analysis of the dam failure has essential significance for reducing dam failure probability and improving dam risk management level. PMID:29710824

  5. Multipoint connectivity analysis of the May 2007 solar energetic particle events

    NASA Astrophysics Data System (ADS)

    Chollet, E. E.; Mewaldt, R. A.; Cummings, A. C.; Gosling, J. T.; Haggerty, D. K.; Hu, Q.; Larson, D.; Lavraud, B.; Leske, R. A.; Opitz, A.; Roelof, E. C.; Russell, C. T.; Sauvaud, J.-A.

    2010-12-01

    In May of 2007, the STEREO Ahead and Behind spacecraft, along with the ACE spacecraft situated between the two STEREO spacecraft, observed two small solar energetic particle (SEP) events. STEREO-A and -B observed nearly identical time profiles in the 19 May event, but in the 23 May event, the protons arrived significantly earlier at STEREO-A than at STEREO-B and the time-intensity profiles were markedly different. We present SEP anisotropy, suprathermal electron pitch angle and solar wind data to demonstrate distortion in the magnetic field topology produced by the passage of multiple interplanetary coronal mass ejections on 22 and 23 May, causing the two spacecraft to magnetically connect to different points back at the Sun. This pair of events illustrates the power of multipoint observations in detailed interpretation of complex events, since only a small shift in observer location results in different magnetic field line connections and different SEP time-intensity profiles.

  6. Paroxysmal non-epileptic events in infants and toddlers: A phenomenologic analysis.

    PubMed

    Chen, Li; Knight, Elia M Pestana; Tuxhorn, Ingrid; Shahid, Asim; Lüders, Hans O

    2015-06-01

    The aim of this study was to analyze in detail the clinical phenomenology of paroxysmal non-epileptic events (PNEE) in infants and toddlers. We studied all children aged ≤2 years who were diagnosed with PNEE based on video-electroencephalographic (VEEG) recordings. We analyzed the following four clinical domains of each clinical event: (i) motor manifestations (body/limb jerking, complex motor, and asymmetric limb posturing); (ii) oral/vocal (crying, vocalization, sighing); (iii) behavioral change (arrest of activity, staring); (iv) and autonomic (facial flushing, breath holding). Thirty-one of 81 (38.3%) infants and toddlers had 38 PNEE recorded during the study period (12 girls and 19 boys, mean age 10.5 months). The predominant clinical features were as follows: motor in 26/38 events, oral/verbal in 14/38 events, behavioral in 11/38 events, and autonomic in 8/38 events. Epileptic seizures and PNEE coexisted in four children (12.9%). Seventeen children (54.8%) had one or more risk factors suggestive of epilepsy. Twelve children (38.7%) had a normal neurologic examination, 10 (32.3%) had developmental delay, and eight (25.8%) had a family history of epilepsy or seizures. VEEG recorded PNEE in nearly 40% of 81 infants and toddlers referred for unclear paroxysmal events in our cohort. Non-epileptic staring spells and benign sleep myoclonus were the most common events recorded, followed by shuddering attacks and infantile masturbation. In addition, greater than one-half of the infants and toddlers had risk factors, raising a concern for epilepsy in the family and prompting the VEEG evaluation, suggesting that paroxysmal non-epileptic seizures may frequently coexist in young children with epilepsy. © 2014 The Authors. Psychiatry and Clinical Neurosciences © 2014 Japanese Society of Psychiatry and Neurology.

  7. Catchment process affecting drinking water quality, including the significance of rainfall events, using factor analysis and event mean concentrations.

    PubMed

    Cinque, Kathy; Jayasuriya, Niranjali

    2010-12-01

    To ensure the protection of drinking water an understanding of the catchment processes which can affect water quality is important as it enables targeted catchment management actions to be implemented. In this study factor analysis (FA) and comparing event mean concentrations (EMCs) with baseline values were techniques used to asses the relationships between water quality parameters and linking those parameters to processes within an agricultural drinking water catchment. FA found that 55% of the variance in the water quality data could be explained by the first factor, which was dominated by parameters usually associated with erosion. Inclusion of pathogenic indicators in an additional FA showed that Enterococcus and Clostridium perfringens (C. perfringens) were also related to the erosion factor. Analysis of the EMCs found that most parameters were significantly higher during periods of rainfall runoff. This study shows that the most dominant processes in an agricultural catchment are surface runoff and erosion. It also shows that it is these processes which mobilise pathogenic indicators and are therefore most likely to influence the transport of pathogens. Catchment management efforts need to focus on reducing the effect of these processes on water quality.

  8. Purpose in Life and Its Relationship to All-Cause Mortality and Cardiovascular Events: A Meta-Analysis.

    PubMed

    Cohen, Randy; Bavishi, Chirag; Rozanski, Alan

    2016-01-01

    To assess the net impact of purpose in life on all-cause mortality and cardiovascular events. The electronic databases PubMed, Embase, and PsycINFO were systematically searched through June 2015 to identify all studies investigating the relationship between purpose in life, mortality, and cardiovascular events. Articles were selected for inclusion if, a) they were prospective, b) evaluated the association between some measure of purpose in life and all-cause mortality and/or cardiovascular events, and c) unadjusted and/or adjusted risk estimates and confidence intervals (CIs) were reported. Ten prospective studies with a total of 136,265 participants were included in the analysis. A significant association was observed between having a higher purpose in life and reduced all-cause mortality (adjusted pooled relative risk = 0.83 [CI = 0.75-0.91], p < .001) and cardiovascular events (adjusted pooled relative risk = 0.83 [CI = 0.75-0.92], p = .001). Subgroup analyses by study country of origin, questionnaire used to measure purpose in life, age, and whether or not participants with baseline cardiovascular disease were included in the study all yielded similar results. Possessing a high sense of purpose in life is associated with a reduced risk for all-cause mortality and cardiovascular events. Future research should focus on mechanisms linking purpose in life to health outcomes, as well as interventions to assist individuals identified as having a low sense of purpose in life.

  9. A systematic review and meta-analysis on herpes zoster and the risk of cardiac and cerebrovascular events

    PubMed Central

    Erskine, Nathaniel; Tran, Hoang; Levin, Leonard; Ulbricht, Christine; Fingeroth, Joyce; Kiefe, Catarina; Singh, Sonal

    2017-01-01

    Background Patients who develop herpes zoster or herpes zoster ophthalmicus may be at risk for cerebrovascular and cardiac complications. We systematically reviewed the published literature to determine the association between herpes zoster and its subtypes with the occurrence of cerebrovascular and cardiac events. Methods/Results Systematic searches of PubMed (MEDLINE), SCOPUS (Embase) and Google Scholar were performed in December 2016. Eligible studies were cohort, case-control, and self-controlled case-series examining the association between herpes zoster or subtypes of herpes zoster with the occurrence of cerebrovascular and cardiac events including stroke, transient ischemic attack, coronary heart disease, and myocardial infarction. Data on the occurrence of the examined events were abstracted. Odds ratios and their accompanying confidence intervals were estimated using random and fixed effects models with statistical heterogeneity estimated with the I2 statistic. Twelve studies examining 7.9 million patients up to 28 years after the onset of herpes zoster met our pre-defined eligibility criteria. Random and fixed effects meta-analyses showed that herpes zoster, type unspecified, and herpes zoster ophthalmicus were associated with a significantly increased risk of cerebrovascular events, without any evidence of statistical heterogeneity. Our meta-analysis also found a significantly increased risk of cardiac events associated with herpes zoster, type unspecified. Conclusions Our results are consistent with the accumulating body of evidence that herpes zoster and herpes zoster ophthalmicus are significantly associated with cerebrovascular and cardiovascular events. PMID:28749981

  10. A review for identification of initiating events in event tree development process on nuclear power plants

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Riyadi, Eko H., E-mail: e.riyadi@bapeten.go.id

    2014-09-30

    Initiating event is defined as any event either internal or external to the nuclear power plants (NPPs) that perturbs the steady state operation of the plant, if operating, thereby initiating an abnormal event such as transient or loss of coolant accident (LOCA) within the NPPs. These initiating events trigger sequences of events that challenge plant control and safety systems whose failure could potentially lead to core damage or large early release. Selection for initiating events consists of two steps i.e. first step, definition of possible events, such as by evaluating a comprehensive engineering, and by constructing a top level logicmore » model. Then the second step, grouping of identified initiating event's by the safety function to be performed or combinations of systems responses. Therefore, the purpose of this paper is to discuss initiating events identification in event tree development process and to reviews other probabilistic safety assessments (PSA). The identification of initiating events also involves the past operating experience, review of other PSA, failure mode and effect analysis (FMEA), feedback from system modeling, and master logic diagram (special type of fault tree). By using the method of study for the condition of the traditional US PSA categorization in detail, could be obtained the important initiating events that are categorized into LOCA, transients and external events.« less

  11. The cost of nurse-sensitive adverse events.

    PubMed

    Pappas, Sharon Holcombe

    2008-05-01

    The aim of this study was to describe the methodology for nursing leaders to determine the cost of adverse events and effective levels of nurse staffing. The growing transparency of quality and cost outcomes motivates healthcare leaders to optimize the effectiveness of nurse staffing. Most hospitals have robust cost accounting systems that provide actual patient-level direct costs. These systems allow an analysis of the cost consumed by patients during a hospital stay. By knowing the cost of complications, leaders have the ability to justify the cost of improved staffing when quality evidence shows that higher nurse staffing improves quality. An analysis was performed on financial and clinical data from hospital databases of 3,200 inpatients. The purpose was to establish a methodology to determine actual cost per case. Three diagnosis-related groups were the focus of the analysis. Five adverse events were analyzed along with the costs. A regression analysis reported that the actual direct cost of an adverse event was dollars 1,029 per case in the congestive heart failure cases and dollars 903 in the surgical cases. There was a significant increase in the cost per case in medical patients with urinary tract infection and pressure ulcers and in surgical patients with urinary tract infection and pneumonia. The odds of pneumonia occurring in surgical patients decreased with additional registered nurse hours per patient day. Hospital cost accounting systems are useful in determining the cost of adverse events and can aid in decision making about nurse staffing. Adverse events add costs to patient care and should be measured at the unit level to adjust staffing to reduce adverse events and avoid costs.

  12. Proton pump inhibitor monotherapy and the risk of cardiovascular events in patients with gastro-esophageal reflux disease: a meta-analysis.

    PubMed

    Sun, S; Cui, Z; Zhou, M; Li, R; Li, H; Zhang, S; Ba, Y; Cheng, G

    2017-02-01

    Proton pump inhibitors (PPIs) are commonly used as potent gastric acid secretion antagonists for gastro-esophageal disorders and their overall safety in patients with gastro-esophageal reflux disease (GERD) is considered to be good and they are well-tolerated. However, recent studies have suggested that PPIs may be a potential independent risk factor for cardiovascular adverse events. The aim of our meta-analysis was to examine the association between PPI monotherapy and cardiovascular events in patients with GERD. A literature search involved examination of relevant databases up to July 2015 including PubMed, Cochrane Library, EMBASE, and ClinicalTrial.gov, as well as selected randomized controlled trials (RCTs) reporting cardiovascular events with PPI exposure in GERD patients. In addition, the pooled risk ratio (RR) and heterogeneity were assessed based on a fixed effects model of the meta-analysis and the I 2 statistic, respectively. Seventeen RCTs covering 7540 patients were selected. The pooled data suggested that the use of PPIs was associated with a 70% increased cardiovascular risk (RR=1.70, 95% CI: [1.13-2.56], P=.01, I 2 =0%). Furthermore, higher risks of adverse cardiovascular events in the omeprazole subgroup (RR=3.17, 95% CI: [1.43-7.03], P=.004, I 2 =25%) and long-term treatment subgroup (RR=2.33, 95% CI: [1.33-4.08], P=.003, I 2 =0%) were found. PPI monotherapy can be a risk factor for cardiovascular adverse events. Omeprazole could significantly increase the risk of cardiovascular events and, so, should be used carefully. © 2016 John Wiley & Sons Ltd.

  13. Increasing the Operational Value of Event Messages

    NASA Technical Reports Server (NTRS)

    Li, Zhenping; Savkli, Cetin; Smith, Dan

    2003-01-01

    Assessing the health of a space mission has traditionally been performed using telemetry analysis tools. Parameter values are compared to known operational limits and are plotted over various time periods. This presentation begins with the notion that there is an incredible amount of untapped information contained within the mission s event message logs. Through creative advancements in message handling tools, the event message logs can be used to better assess spacecraft and ground system status and to highlight and report on conditions not readily apparent when messages are evaluated one-at-a-time during a real-time pass. Work in this area is being funded as part of a larger NASA effort at the Goddard Space Flight Center to create component-based, middleware-based, standards-based general purpose ground system architecture referred to as GMSEC - the GSFC Mission Services Evolution Center. The new capabilities and operational concepts for event display, event data analyses and data mining are being developed by Lockheed Martin and the new subsystem has been named GREAT - the GMSEC Reusable Event Analysis Toolkit. Planned for use on existing and future missions, GREAT has the potential to increase operational efficiency in areas of problem detection and analysis, general status reporting, and real-time situational awareness.

  14. Heavy precipitation events in northern Switzerland

    NASA Astrophysics Data System (ADS)

    Giannakaki, Paraskevi; Martius, Olivia

    2013-04-01

    Heavy precipitation events in the Alpine region often cause floods, rock-falls and mud slides with severe consequences for population and economy. Breaking synoptic Rossby waves located over western Europe, play a central role in triggering such heavy rain events in southern Switzerland (e.g. Massacand et al. 1998). In contrast, synoptic scale structures triggering heavy precipitation on the north side of the Swiss Alps and orographic effects have so far not been studied comprehensively. An observation based high resolution precipitation data set for Switzerland and the Alps (MeteoSwiss) is used to identify heavy precipitation events affecting the north side of the Swiss Alps for the time period 1961-2010. For these events a detailed statistical and dynamical analysis of the upper level flow is conducted using ECMWFs ERA-40 and ERA-Interim reanalysis data sets. For the analysis north side of the Swiss Alps is divided in two investigation areas north-eastern and western Switzerland following the Swiss climate change scenarios (Bey et al. 2011). A subjective classification of upper level structures triggering heavy precipitation events in the areas of interest is presented. Four classes are defined based on the orientation and formation of the dynamical tropopause during extreme events in the northern part of Switzerland and its sub-regions. The analysis is extended by a climatology of breaking waves and cut-offs following the method of Wernli and Sprenger (2007) to examine their presence and location during extreme events. References Bey I., Croci-Maspoli M., Fuhrer J., Kull C, Appenzeller C., Knutti R. and Schär C. Swiss Climate Change Scenarios CH2011, C2SM, MeteoSwiss, ETH, NCCR Climate, OcCC (2011), http://dx.doi.org/10.3929/ethz-a-006720559 Massacand A., H. Wernli, and H.C. Davies, 1998. Heavy precipitation on the Alpine South side: An upper-level precursor. Geophys. Res. Lett., 25, 1435-1438. MeteoSwiss 2011. Documentation of Meteoswiss grid-data products

  15. Neural Events in the Reinforcement Contingency

    ERIC Educational Resources Information Center

    Silva, Maria Teresa Araujo; Goncalves, Fabio Leyser; Garcia-Mijares, Miriam

    2007-01-01

    When neural events are analyzed as stimuli and responses, functional relations among them and among overt stimuli and responses can be unveiled. The integration of neuroscience and the experimental analysis of behavior is beginning to provide empirical evidence of involvement of neural events in the three-term contingency relating discriminative…

  16. Single Event Upset Analysis: On-orbit performance of the Alpha Magnetic Spectrometer Digital Signal Processor Memory aboard the International Space Station

    NASA Astrophysics Data System (ADS)

    Li, Jiaqiang; Choutko, Vitaly; Xiao, Liyi

    2018-03-01

    Based on the collection of error data from the Alpha Magnetic Spectrometer (AMS) Digital Signal Processors (DSP), on-orbit Single Event Upsets (SEUs) of the DSP program memory are analyzed. The daily error distribution and time intervals between errors are calculated to evaluate the reliability of the system. The particle density distribution of International Space Station (ISS) orbit is presented and the effects from the South Atlantic Anomaly (SAA) and the geomagnetic poles are analyzed. The impact of solar events on the DSP program memory is carried out combining data analysis and Monte Carlo simulation (MC). From the analysis and simulation results, it is concluded that the area corresponding to the SAA is the main source of errors on the ISS orbit. Solar events can also cause errors on DSP program memory, but the effect depends on the on-orbit particle density.

  17. Calibration, event reconstruction, data analysis and limits calculation for the LUX dark matter experiment

    DOE PAGES

    Akerib, DS; Alsum, S; Araújo, HM; ...

    2018-01-05

    The LUX experiment has performed searches for dark matter particles scattering elastically on xenon nuclei, leading to stringent upper limits on the nuclear scattering cross sections for dark matter. Here, for results derived frommore » $${1.4}\\times 10^{4}\\;\\mathrm{kg\\,days}$$ of target exposure in 2013, details of the calibration, event-reconstruction, modeling, and statistical tests that underlie the results are presented. Detector performance is characterized, including measured efficiencies, stability of response, position resolution, and discrimination between electron- and nuclear-recoil populations. Models are developed for the drift field, optical properties, background populations, the electron- and nuclear-recoil responses, and the absolute rate of low-energy background events. Innovations in the analysis include in situ measurement of the photomultipliers' response to xenon scintillation photons, verification of fiducial mass with a low-energy internal calibration source, and new empirical models for low-energy signal yield based on large-sample, in situ calibrations.« less

  18. Calibration, event reconstruction, data analysis, and limit calculation for the LUX dark matter experiment

    DOE PAGES

    Akerib, D. S.; Alsum, S.; Araújo, H. M.; ...

    2018-05-31

    Here, the LUX experiment has performed searches for dark matter particles scattering elastically on xenon nuclei, leading to stringent upper limits on the nuclear scattering cross sections for dark matter. Here, for results derived frommore » $${1.4}\\times 10^{4}\\;\\mathrm{kg\\,days}$$ of target exposure in 2013, details of the calibration, event-reconstruction, modeling, and statistical tests that underlie the results are presented. Detector performance is characterized, including measured efficiencies, stability of response, position resolution, and discrimination between electron- and nuclear-recoil populations. Models are developed for the drift field, optical properties, background populations, the electron- and nuclear-recoil responses, and the absolute rate of low-energy background events. Innovations in the analysis include in situ measurement of the photomultipliers' response to xenon scintillation photons, verification of fiducial mass with a low-energy internal calibration source, and new empirical models for low-energy signal yield based on large-sample, in situ calibrations.« less

  19. Big Data Toolsets to Pharmacometrics: Application of Machine Learning for Time‐to‐Event Analysis

    PubMed Central

    Gong, Xiajing; Hu, Meng

    2018-01-01

    Abstract Additional value can be potentially created by applying big data tools to address pharmacometric problems. The performances of machine learning (ML) methods and the Cox regression model were evaluated based on simulated time‐to‐event data synthesized under various preset scenarios, i.e., with linear vs. nonlinear and dependent vs. independent predictors in the proportional hazard function, or with high‐dimensional data featured by a large number of predictor variables. Our results showed that ML‐based methods outperformed the Cox model in prediction performance as assessed by concordance index and in identifying the preset influential variables for high‐dimensional data. The prediction performances of ML‐based methods are also less sensitive to data size and censoring rates than the Cox regression model. In conclusion, ML‐based methods provide a powerful tool for time‐to‐event analysis, with a built‐in capacity for high‐dimensional data and better performance when the predictor variables assume nonlinear relationships in the hazard function. PMID:29536640

  20. Calibration, event reconstruction, data analysis, and limit calculation for the LUX dark matter experiment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Akerib, D. S.; Alsum, S.; Araújo, H. M.

    Here, the LUX experiment has performed searches for dark matter particles scattering elastically on xenon nuclei, leading to stringent upper limits on the nuclear scattering cross sections for dark matter. Here, for results derived frommore » $${1.4}\\times 10^{4}\\;\\mathrm{kg\\,days}$$ of target exposure in 2013, details of the calibration, event-reconstruction, modeling, and statistical tests that underlie the results are presented. Detector performance is characterized, including measured efficiencies, stability of response, position resolution, and discrimination between electron- and nuclear-recoil populations. Models are developed for the drift field, optical properties, background populations, the electron- and nuclear-recoil responses, and the absolute rate of low-energy background events. Innovations in the analysis include in situ measurement of the photomultipliers' response to xenon scintillation photons, verification of fiducial mass with a low-energy internal calibration source, and new empirical models for low-energy signal yield based on large-sample, in situ calibrations.« less

  1. Disruption Event Characterization and Forecasting in Tokamaks

    NASA Astrophysics Data System (ADS)

    Berkery, J. W.; Sabbagh, S. A.; Park, Y. S.; Ahn, J. H.; Jiang, Y.; Riquezes, J. D.; Gerhardt, S. P.; Myers, C. E.

    2017-10-01

    The Disruption Event Characterization and Forecasting (DECAF) code, being developed to meet the challenging goal of high reliability disruption prediction in tokamaks, automates data analysis to determine chains of events that lead to disruptions and to forecast their evolution. The relative timing of magnetohydrodynamic modes and other events including plasma vertical displacement, loss of boundary control, proximity to density limits, reduction of safety factor, and mismatch of the measured and desired plasma current are considered. NSTX/-U databases are examined with analysis expanding to DIII-D, KSTAR, and TCV. Characterization of tearing modes has determined mode bifurcation frequency and locking points. In an NSTX database exhibiting unstable resistive wall modes (RWM), the RWM event and loss of boundary control event were found in 100%, and the vertical displacement event in over 90% of cases. A reduced kinetic RWM stability physics model is evaluated to determine the proximity of discharges to marginal stability. The model shows high success as a disruption predictor (greater than 85%) with relatively low false positive rate. Supported by US DOE Contracts DE-FG02-99ER54524, DE-AC02-09CH11466, and DE-SC0016614.

  2. Sexual Partnership Types as Determinant of HIV Risk in South African MSM: An Event-Level Cluster Analysis

    PubMed Central

    Sandfort, Theo; Yi, Huso; Knox, Justin; Reddy, Vasu

    2012-01-01

    While individual determinants of HIV risk among MSM have been widely studied, there is limited understanding of how relational characteristics determine sexual risk. Based on data collected among 300 South African men who have sex with men (MSM) and using cluster analysis, this study developed a typology of four partnership types: the “Race-Economic Similar,” “Age-Race-Economic Discordant,” “Non-regular Neighbourhood,” and “Familiar” partnership types. Support for the meaningfulness of these types was found through associations of these partnership types with participant characteristics and characteristics of the last anal sex event. Furthermore, in a multivariate analysis, only partnership type independently predicted whether the last anal sex event was unprotected. Findings of the study illustrate the importance of taking into account the relational context in understanding unprotected sexual practices and present ways to target intervention efforts as well as identify relationship specific determinants of unprotected sex. PMID:22956229

  3. Climate-change driven increase in high intensity rainfall events: Analysis of development in the last decades and towards an extrapolation of future progression

    NASA Astrophysics Data System (ADS)

    Müller, Eva; Pfister, Angela; Gerd, Büger; Maik, Heistermann; Bronstert, Axel

    2015-04-01

    Hydrological extreme events can be triggered by rainfall on different spatiotemporal scales: river floods are typically caused by event durations of between hours and days, while urban flash floods as well as soil erosion or contaminant transport rather result from storms events of very short duration (minutes). Still, the analysis of climate change impacts on rainfall-induced extreme events is usually carried out using daily precipitation data at best. Trend analyses of extreme rainfall at sub-daily or even sub-hourly time scales are rare. In this contribution two lines of research are combined: first, we analyse sub-hourly rainfall data for several decades in three European regions.Second, we investigate the scaling behaviour of heavy short-term precipitation with temperature, i.e. the dependence of high intensity rainfall on the atmospheric temperature at that particular time and location. The trend analysis of high-resolution rainfall data shows for the first time that the frequency of short and intensive storm events in the temperate lowland regions in Germany has increased by up to 0.5 events per year over the last decades. I.e. this trend suggests that the occurrence of these types of storms have multiplied over only a few decades. Parallel to the changes in the rainfall regime, increases in the annual and seasonal average temperature and changes in the occurrence of circulation patterns responsible for the generation of high-intensity storms have been found. The analysis of temporally highly resolved rainfall records from three European regions further indicates that extreme precipitation events are more intense with warmer temperatures during the rainfall event. These observations follow partly the Clausius-Clapeyron relation. Based on this relation one may derive a general rule of maximum rainfall intensity associated to the event temperature, roughly following the Clausius-Clapeyron (CC) relation. This rule might be used for scenarios of future maximum

  4. Event visualization in ATLAS

    NASA Astrophysics Data System (ADS)

    Bianchi, R. M.; Boudreau, J.; Konstantinidis, N.; Martyniuk, A. C.; Moyse, E.; Thomas, J.; Waugh, B. M.; Yallup, D. P.; ATLAS Collaboration

    2017-10-01

    At the beginning, HEP experiments made use of photographical images both to record and store experimental data and to illustrate their findings. Then the experiments evolved and needed to find ways to visualize their data. With the availability of computer graphics, software packages to display event data and the detector geometry started to be developed. Here, an overview of the usage of event display tools in HEP is presented. Then the case of the ATLAS experiment is considered in more detail and two widely used event display packages are presented, Atlantis and VP1, focusing on the software technologies they employ, as well as their strengths, differences and their usage in the experiment: from physics analysis to detector development, and from online monitoring to outreach and communication. Towards the end, the other ATLAS visualization tools will be briefly presented as well. Future development plans and improvements in the ATLAS event display packages will also be discussed.

  5. ERPLAB: an open-source toolbox for the analysis of event-related potentials

    PubMed Central

    Lopez-Calderon, Javier; Luck, Steven J.

    2014-01-01

    ERPLAB toolbox is a freely available, open-source toolbox for processing and analyzing event-related potential (ERP) data in the MATLAB environment. ERPLAB is closely integrated with EEGLAB, a popular open-source toolbox that provides many EEG preprocessing steps and an excellent user interface design. ERPLAB adds to EEGLAB’s EEG processing functions, providing additional tools for filtering, artifact detection, re-referencing, and sorting of events, among others. ERPLAB also provides robust tools for averaging EEG segments together to create averaged ERPs, for creating difference waves and other recombinations of ERP waveforms through algebraic expressions, for filtering and re-referencing the averaged ERPs, for plotting ERP waveforms and scalp maps, and for quantifying several types of amplitudes and latencies. ERPLAB’s tools can be accessed either from an easy-to-learn graphical user interface or from MATLAB scripts, and a command history function makes it easy for users with no programming experience to write scripts. Consequently, ERPLAB provides both ease of use and virtually unlimited power and flexibility, making it appropriate for the analysis of both simple and complex ERP experiments. Several forms of documentation are available, including a detailed user’s guide, a step-by-step tutorial, a scripting guide, and a set of video-based demonstrations. PMID:24782741

  6. ERPLAB: an open-source toolbox for the analysis of event-related potentials.

    PubMed

    Lopez-Calderon, Javier; Luck, Steven J

    2014-01-01

    ERPLAB toolbox is a freely available, open-source toolbox for processing and analyzing event-related potential (ERP) data in the MATLAB environment. ERPLAB is closely integrated with EEGLAB, a popular open-source toolbox that provides many EEG preprocessing steps and an excellent user interface design. ERPLAB adds to EEGLAB's EEG processing functions, providing additional tools for filtering, artifact detection, re-referencing, and sorting of events, among others. ERPLAB also provides robust tools for averaging EEG segments together to create averaged ERPs, for creating difference waves and other recombinations of ERP waveforms through algebraic expressions, for filtering and re-referencing the averaged ERPs, for plotting ERP waveforms and scalp maps, and for quantifying several types of amplitudes and latencies. ERPLAB's tools can be accessed either from an easy-to-learn graphical user interface or from MATLAB scripts, and a command history function makes it easy for users with no programming experience to write scripts. Consequently, ERPLAB provides both ease of use and virtually unlimited power and flexibility, making it appropriate for the analysis of both simple and complex ERP experiments. Several forms of documentation are available, including a detailed user's guide, a step-by-step tutorial, a scripting guide, and a set of video-based demonstrations.

  7. Adverse event reporting in cancer clinical trial publications.

    PubMed

    Sivendran, Shanthi; Latif, Asma; McBride, Russell B; Stensland, Kristian D; Wisnivesky, Juan; Haines, Lindsay; Oh, William K; Galsky, Matthew D

    2014-01-10

    Reporting adverse events is a critical element of a clinical trial publication. In 2003, the Consolidated Standards of Reporting Trials (CONSORT) group generated recommendations regarding the appropriate reporting of adverse events. The degree to which these recommendations are followed in oncology publications has not been comprehensively evaluated. A review of citations from PubMed, Medline, and Embase published between Jan 1, 2009 and December 31, 2011, identified eligible randomized, controlled phase III trials in metastatic solid malignancies. Publications were assessed for 14 adverse event-reporting elements derived from the CONSORT harms extension statement; a completeness score (range, 0 to 14) was calculated by adding the number of elements reported. Linear regression analysis identified which publication characteristics associated with reporting completeness. A total of 175 publications, with data for 96,125 patients, were included in the analysis. The median completeness score was eight (range, three to 12). Most publications (96%) reported only adverse events occurring above a threshold rate or severity, 37% did not specify the criteria used to select which adverse events were reported, and 88% grouped together adverse events of varying severity. Regression analysis revealed that trials without a stated funding source and with an earlier year of publication had significantly lower completeness scores. Reporting of adverse events in oncology publications of randomized trials is suboptimal and characterized by substantial selectivity and heterogeneity. The development of oncology-specific standards for adverse event reporting should be established to ensure consistency and provide critical information required for medical decision-making.

  8. Single-Event Effects in High-Frequency Linear Amplifiers: Experiment and Analysis

    NASA Astrophysics Data System (ADS)

    Zeinolabedinzadeh, Saeed; Ying, Hanbin; Fleetwood, Zachary E.; Roche, Nicolas J.-H.; Khachatrian, Ani; McMorrow, Dale; Buchner, Stephen P.; Warner, Jeffrey H.; Paki-Amouzou, Pauline; Cressler, John D.

    2017-01-01

    The single-event transient (SET) response of two different silicon-germanium (SiGe) X-band (8-12 GHz) low noise amplifier (LNA) topologies is fully investigated in this paper. The two LNAs were designed and implemented in 130nm SiGe HBT BiCMOS process technology. Two-photon absorption (TPA) laser pulses were utilized to induce transients within various devices in these LNAs. Impulse response theory is identified as a useful tool for predicting the settling behavior of the LNAs subjected to heavy ion strikes. Comprehensive device and circuit level modeling and simulations were performed to accurately simulate the behavior of the circuits under ion strikes. The simulations agree well with TPA measurements. The simulation, modeling and analysis presented in this paper can be applied for any other circuit topologies for SET modeling and prediction.

  9. Life events, anxiety, social support, personality, and alexithymia in female patients with chronic pain: A path analysis.

    PubMed

    Zeng, Fanmin; Sun, Xueli; Yang, Bangxiang; Fu, Xiaoqian

    2016-03-01

    This study sought to identify a model that explains the relationship between psychosocial factors and chronic pain in female patients, and to explore all of these constructs in a single study and provide a more holistic examination of the overall psychosocial factors that female patients with chronic pain encounter. Female patients with chronic pain (n = 147), aged 20-65 (M = 34.9 years, SD = 11.25), from an outpatient pain clinic completed a cross-sectional self-report questionnaire on anxiety, life events, personality, social support, and alexithymia. Data were analyzed by means of path analysis. The direct effect of anxiety on female patients with chronic pain was greatest among all the paths. Personality and alexithymia led to chronic pain in female patients only indirectly, mediated by life events. The personality factors of neuroticism and extraversion were associated positively with social support, which had an indirect effect on the influence of life events on chronic pain. However, alexithymia was associated negatively with social support, which had an indirect effect on the influence of life events on chronic pain. Our findings provide evidence that life events are a mediator in the relationship between personality, social support, alexithymia, and chronic pain in female patients. © 2015 The Authors. Asia-Pacific Psychiatry published by Wiley Publishing Asia Pty Ltd.

  10. ASSET: Analysis of Sequences of Synchronous Events in Massively Parallel Spike Trains

    PubMed Central

    Canova, Carlos; Denker, Michael; Gerstein, George; Helias, Moritz

    2016-01-01

    With the ability to observe the activity from large numbers of neurons simultaneously using modern recording technologies, the chance to identify sub-networks involved in coordinated processing increases. Sequences of synchronous spike events (SSEs) constitute one type of such coordinated spiking that propagates activity in a temporally precise manner. The synfire chain was proposed as one potential model for such network processing. Previous work introduced a method for visualization of SSEs in massively parallel spike trains, based on an intersection matrix that contains in each entry the degree of overlap of active neurons in two corresponding time bins. Repeated SSEs are reflected in the matrix as diagonal structures of high overlap values. The method as such, however, leaves the task of identifying these diagonal structures to visual inspection rather than to a quantitative analysis. Here we present ASSET (Analysis of Sequences of Synchronous EvenTs), an improved, fully automated method which determines diagonal structures in the intersection matrix by a robust mathematical procedure. The method consists of a sequence of steps that i) assess which entries in the matrix potentially belong to a diagonal structure, ii) cluster these entries into individual diagonal structures and iii) determine the neurons composing the associated SSEs. We employ parallel point processes generated by stochastic simulations as test data to demonstrate the performance of the method under a wide range of realistic scenarios, including different types of non-stationarity of the spiking activity and different correlation structures. Finally, the ability of the method to discover SSEs is demonstrated on complex data from large network simulations with embedded synfire chains. Thus, ASSET represents an effective and efficient tool to analyze massively parallel spike data for temporal sequences of synchronous activity. PMID:27420734

  11. Event Rates, Hospital Utilization, and Costs Associated with Major Complications of Diabetes: A Multicountry Comparative Analysis

    PubMed Central

    Clarke, Philip M.; Glasziou, Paul; Patel, Anushka; Chalmers, John; Woodward, Mark; Harrap, Stephen B.; Salomon, Joshua A.

    2010-01-01

    Background Diabetes imposes a substantial burden globally in terms of premature mortality, morbidity, and health care costs. Estimates of economic outcomes associated with diabetes are essential inputs to policy analyses aimed at prevention and treatment of diabetes. Our objective was to estimate and compare event rates, hospital utilization, and costs associated with major diabetes-related complications in high-, middle-, and low-income countries. Methods and Findings Incidence and history of diabetes-related complications, hospital admissions, and length of stay were recorded in 11,140 patients with type 2 diabetes participating in the Action in Diabetes and Vascular Disease (ADVANCE) study (mean age at entry 66 y). The probability of hospital utilization and number of days in hospital for major events associated with coronary disease, cerebrovascular disease, congestive heart failure, peripheral vascular disease, and nephropathy were estimated for three regions (Asia, Eastern Europe, and Established Market Economies) using multiple regression analysis. The resulting estimates of days spent in hospital were multiplied by regional estimates of the costs per hospital bed-day from the World Health Organization to compute annual acute and long-term costs associated with the different types of complications. To assist, comparability, costs are reported in international dollars (Int$), which represent a hypothetical currency that allows for the same quantities of goods or services to be purchased regardless of country, standardized on purchasing power in the United States. A cost calculator accompanying this paper enables the estimation of costs for individual countries and translation of these costs into local currency units. The probability of attending a hospital following an event was highest for heart failure (93%–96% across regions) and lowest for nephropathy (15%–26%). The average numbers of days in hospital given at least one admission were greatest for

  12. Towards Real-Time Detection of Gait Events on Different Terrains Using Time-Frequency Analysis and Peak Heuristics Algorithm.

    PubMed

    Zhou, Hui; Ji, Ning; Samuel, Oluwarotimi Williams; Cao, Yafei; Zhao, Zheyi; Chen, Shixiong; Li, Guanglin

    2016-10-01

    Real-time detection of gait events can be applied as a reliable input to control drop foot correction devices and lower-limb prostheses. Among the different sensors used to acquire the signals associated with walking for gait event detection, the accelerometer is considered as a preferable sensor due to its convenience of use, small size, low cost, reliability, and low power consumption. Based on the acceleration signals, different algorithms have been proposed to detect toe off (TO) and heel strike (HS) gait events in previous studies. While these algorithms could achieve a relatively reasonable performance in gait event detection, they suffer from limitations such as poor real-time performance and are less reliable in the cases of up stair and down stair terrains. In this study, a new algorithm is proposed to detect the gait events on three walking terrains in real-time based on the analysis of acceleration jerk signals with a time-frequency method to obtain gait parameters, and then the determination of the peaks of jerk signals using peak heuristics. The performance of the newly proposed algorithm was evaluated with eight healthy subjects when they were walking on level ground, up stairs, and down stairs. Our experimental results showed that the mean F1 scores of the proposed algorithm were above 0.98 for HS event detection and 0.95 for TO event detection on the three terrains. This indicates that the current algorithm would be robust and accurate for gait event detection on different terrains. Findings from the current study suggest that the proposed method may be a preferable option in some applications such as drop foot correction devices and leg prostheses.

  13. Towards Real-Time Detection of Gait Events on Different Terrains Using Time-Frequency Analysis and Peak Heuristics Algorithm

    PubMed Central

    Zhou, Hui; Ji, Ning; Samuel, Oluwarotimi Williams; Cao, Yafei; Zhao, Zheyi; Chen, Shixiong; Li, Guanglin

    2016-01-01

    Real-time detection of gait events can be applied as a reliable input to control drop foot correction devices and lower-limb prostheses. Among the different sensors used to acquire the signals associated with walking for gait event detection, the accelerometer is considered as a preferable sensor due to its convenience of use, small size, low cost, reliability, and low power consumption. Based on the acceleration signals, different algorithms have been proposed to detect toe off (TO) and heel strike (HS) gait events in previous studies. While these algorithms could achieve a relatively reasonable performance in gait event detection, they suffer from limitations such as poor real-time performance and are less reliable in the cases of up stair and down stair terrains. In this study, a new algorithm is proposed to detect the gait events on three walking terrains in real-time based on the analysis of acceleration jerk signals with a time-frequency method to obtain gait parameters, and then the determination of the peaks of jerk signals using peak heuristics. The performance of the newly proposed algorithm was evaluated with eight healthy subjects when they were walking on level ground, up stairs, and down stairs. Our experimental results showed that the mean F1 scores of the proposed algorithm were above 0.98 for HS event detection and 0.95 for TO event detection on the three terrains. This indicates that the current algorithm would be robust and accurate for gait event detection on different terrains. Findings from the current study suggest that the proposed method may be a preferable option in some applications such as drop foot correction devices and leg prostheses. PMID:27706086

  14. Development and application of a multi-targeting reference plasmid as calibrator for analysis of five genetically modified soybean events.

    PubMed

    Pi, Liqun; Li, Xiang; Cao, Yiwei; Wang, Canhua; Pan, Liangwen; Yang, Litao

    2015-04-01

    Reference materials are important in accurate analysis of genetically modified organism (GMO) contents in food/feeds, and development of novel reference plasmid is a new trend in the research of GMO reference materials. Herein, we constructed a novel multi-targeting plasmid, pSOY, which contained seven event-specific sequences of five GM soybeans (MON89788-5', A2704-12-3', A5547-127-3', DP356043-5', DP305423-3', A2704-12-5', and A5547-127-5') and sequence of soybean endogenous reference gene Lectin. We evaluated the specificity, limit of detection and quantification, and applicability of pSOY in both qualitative and quantitative PCR analyses. The limit of detection (LOD) was as low as 20 copies in qualitative PCR, and the limit of quantification (LOQ) in quantitative PCR was 10 copies. In quantitative real-time PCR analysis, the PCR efficiencies of all event-specific and Lectin assays were higher than 90%, and the squared regression coefficients (R(2)) were more than 0.999. The quantification bias varied from 0.21% to 19.29%, and the relative standard deviations were from 1.08% to 9.84% in simulated samples analysis. All the results demonstrated that the developed multi-targeting plasmid, pSOY, was a credible substitute of matrix reference materials, and could be used as a reliable reference calibrator in the identification and quantification of multiple GM soybean events.

  15. Social network changes and life events across the life span: a meta-analysis.

    PubMed

    Wrzus, Cornelia; Hänel, Martha; Wagner, Jenny; Neyer, Franz J

    2013-01-01

    For researchers and practitioners interested in social relationships, the question remains as to how large social networks typically are, and how their size and composition change across adulthood. On the basis of predictions of socioemotional selectivity theory and social convoy theory, we conducted a meta-analysis on age-related social network changes and the effects of life events on social networks using 277 studies with 177,635 participants from adolescence to old age. Cross-sectional as well as longitudinal studies consistently showed that (a) the global social network increased up until young adulthood and then decreased steadily, (b) both the personal network and the friendship network decreased throughout adulthood, (c) the family network was stable in size from adolescence to old age, and (d) other networks with coworkers or neighbors were important only in specific age ranges. Studies focusing on life events that occur at specific ages, such as transition to parenthood, job entry, or widowhood, demonstrated network changes similar to such age-related network changes. Moderator analyses detected that the type of network assessment affected the reported size of global, personal, and family networks. Period effects on network sizes occurred for personal and friendship networks, which have decreased in size over the last 35 years. Together the findings are consistent with the view that a portion of normative, age-related social network changes are due to normative, age-related life events. We discuss how these patterns of normative social network development inform research in social, evolutionary, cultural, and personality psychology. (PsycINFO Database Record (c) 2013 APA, all rights reserved).

  16. Radionuclide data analysis in connection of DPRK event in May 2009

    NASA Astrophysics Data System (ADS)

    Nikkinen, Mika; Becker, Andreas; Zähringer, Matthias; Polphong, Pornsri; Pires, Carla; Assef, Thierry; Han, Dongmei

    2010-05-01

    The seismic event detected in DPRK on 25.5.2009 was triggering a series of actions within CTBTO/PTS to ensure its preparedness to detect any radionuclide emissions possibly linked with the event. Despite meticulous work to detect and verify, traces linked to the DPRK event were not found. After three weeks of high alert the PTS resumed back to normal operational routine. This case illuminates the importance of objectivity and procedural approach in the data evaluation. All the data coming from particulate and noble gas stations were evaluated daily, some of the samples even outside of office hours and during the weekends. Standard procedures were used to determine the network detection thresholds of the key (CTBT relevant) radionuclides achieved across the DPRK event area and for the assessment of radionuclides typically occurring at IMS stations (background history). Noble gas system has sometimes detections that are typical for the sites due to legitimate non-nuclear test related activities. Therefore, set of hypothesis were used to see if the detection is consistent with event time and location through atmospheric transport modelling. Also the consistency of event timing and isotopic ratios was used in the evaluation work. As a result it was concluded that if even 1/1000 of noble gasses from a nuclear detonation would had leaked, the IMS system would not had problems to detect it. This case also showed the importance of on-site inspections to verify the nuclear traces of possible tests.

  17. The four faces of rumination to stressful events: A psychometric analysis.

    PubMed

    García, Felipe E; Duque, Almudena; Cova, Félix

    2017-11-01

    To increase the knowledge of rumination and its associations with stressful events, we explored the relationships between 4 types of rumination (brooding, reflection, intrusive, and deliberate rumination) in a sample of 750 adult participants who experienced a highly stressful event. We also explored the predictive value of the different types of rumination on posttraumatic stress symptoms and posttraumatic growth 6 months after the highly stressful event occurred. Participants completed the Ruminative Response Scale and the Event-Related Rumination Inventory. Brooding and reflection rumination were obtained from the Ruminative Response Scale, whereas deliberate and intrusive rumination were obtained from the Event-Related Rumination Inventory. Confirmatory factorial analyses were conducted using the 4 types of rumination to test 3 different models: (a) 4-factor model (brooding, reflection, intrusive, and deliberate rumination), (b) 2-factor model: adaptive rumination (reflection and deliberate) and maladaptive rumination (brooding and intrusive), and (c) 2-factor model: depressive rumination (brooding and reflection) and posttraumatic rumination (intrusive and deliberate). It was observed that the 4-factor model showed the best fit to the data. Moreover, 6 months later it was observed that the most significant predictor of posttraumatic symptoms was intrusive rumination, whereas deliberate rumination was the most significant predictor of posttraumatic growth. Results indicate that the 4 types of rumination are differentiated constructs. Ruminative thoughts experienced after a stressful event predicted posttraumatic consequences 6 months later. Implications of these findings are discussed. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  18. Analysis of a potential meteorite-dropping event over the south of Spain in 2007

    NASA Astrophysics Data System (ADS)

    Madiedo, J. M.; Trigo-Rodríguez, J. M.

    2008-09-01

    the case of Puerto Lápice, there are no pictures or videos of the June 29, 2007 bolide and just some images of the distorted train taken several minutes later are available. A forth potential meteoritedropping bolide could be directly recorded by SPMN video cameras on March 25, 2007. We were lucky enough of having this event near to the zenith of two SPMN stations, exhibiting all its magnificence (Fig. 2). We focus here on the preliminary analysis of this event, which was observed over an

  19. Identification and analysis of long duration low frequency events from microseismic data

    NASA Astrophysics Data System (ADS)

    Hu, H.; Li, A.

    2016-12-01

    Long duration low frequency (LDLF) earthquakes, which are commonly present in volcanic fields and subduction zones, have been observed from microseismic data. In this research, we have identified and located several LDLF events from a microseismic dataset acquired by surface receivers in the Eagle Ford Shale. The LDLF events are clearly identified on frequency-time plots with the central frequencies at 5-25 Hz and the duration time from tens of seconds up to 100 seconds. We pick the arrival times of the events using the envelops of the filtered data and apply a grid search method to find the source locations. These events are located at the depth around 1500 m, close to the horizontal treatment well for hydraulic fracturing. The associated phase arrivals show typical P-wave moveout trends. In addition, these events tend to migrate away from the horizontal well with time. Furthermore, these events are recorded only during the time when the rock is breaking according to the treating pressure records. Considering all these observations, we conclude that the observed LDLF events are caused by the pressure change related to fluid flow in fractures. The time-dependence source locations could have an important application to characterize the fluid path inside fractures.

  20. Robust non-parametric one-sample tests for the analysis of recurrent events.

    PubMed

    Rebora, Paola; Galimberti, Stefania; Valsecchi, Maria Grazia

    2010-12-30

    One-sample non-parametric tests are proposed here for inference on recurring events. The focus is on the marginal mean function of events and the basis for inference is the standardized distance between the observed and the expected number of events under a specified reference rate. Different weights are considered in order to account for various types of alternative hypotheses on the mean function of the recurrent events process. A robust version and a stratified version of the test are also proposed. The performance of these tests was investigated through simulation studies under various underlying event generation processes, such as homogeneous and nonhomogeneous Poisson processes, autoregressive and renewal processes, with and without frailty effects. The robust versions of the test have been shown to be suitable in a wide variety of event generating processes. The motivating context is a study on gene therapy in a very rare immunodeficiency in children, where a major end-point is the recurrence of severe infections. Robust non-parametric one-sample tests for recurrent events can be useful to assess efficacy and especially safety in non-randomized studies or in epidemiological studies for comparison with a standard population. Copyright © 2010 John Wiley & Sons, Ltd.

  1. Dementia Medications and Risk of Falls, Syncope, and Related Adverse Events Meta-Analysis of Randomized Controlled Trials

    PubMed Central

    Kim, Dae Hyun; Brown, Rebecca T.; Ding, Eric L.; Kiel, Douglas P.; Berry, Sarah D.

    2012-01-01

    Background Conflicting evidence exists on whether cholinesterase inhibitors and memantine increase the risk of falls, syncope, and related events, defined as fracture and accidental injury. Objectives To evaluate the effect of cholinesterase inhibitors and memantine on the risk of falls, syncope, and related events Design, Setting, Participants, and Intervention Meta-analysis of 54 placebo-controlled randomized trials and extension studies of cholinesterase inhibitors and memantine that reported falls, syncope, and related events in cognitively impaired older adults. Trials were identified from MEDLINE, EMBASE, Cochrane Central Register of Controlled Trials (no language restriction, through July 2009), and manual search. Measurements Falls, syncope, fracture, and accidental injury Results Compared to placebo, cholinesterase inhibitor use was associated with an increased risk of syncope (odds ratio [95% confidence interval]: 1.53 [1.02-2.30]), but not with other events (falls: 0.88 [0.74-1.04]; fracture: 1.39 [0.75-2.56]; accidental injury: 1.13 [0.87-1.45]). Memantine use was associated with fewer fractures (0.21 [0.05-0.85]), but not with other events (fall: 0.92 [0.72-1.18]; syncope: 1.04 [0.35-3.04]; accidental injury: 0.80 [0.56-1.12]). There was no differential effect by type and severity of cognitive impairment, residential status, nor length of follow-up. However, due to underreporting and small number of events, a potential benefit or risk cannot be excluded. Conclusion Cholinesterase inhibitors may increase the risk of syncope, with no effects on falls, fracture, and accidental injury in cognitively impaired older adults. Memantine may have a favorable effect on fracture, with no effects on other events. More research is needed to confirm the reduction in fractures observed for memantine. PMID:21649634

  2. Event Segmentation Ability Uniquely Predicts Event Memory

    PubMed Central

    Sargent, Jesse Q.; Zacks, Jeffrey M.; Hambrick, David Z.; Zacks, Rose T.; Kurby, Christopher A.; Bailey, Heather R.; Eisenberg, Michelle L.; Beck, Taylor M.

    2013-01-01

    Memory for everyday events plays a central role in tasks of daily living, autobiographical memory, and planning. Event memory depends in part on segmenting ongoing activity into meaningful units. This study examined the relationship between event segmentation and memory in a lifespan sample to answer the following question: Is the ability to segment activity into meaningful events a unique predictor of subsequent memory, or is the relationship between event perception and memory accounted for by general cognitive abilities? Two hundred and eight adults ranging from 20 to 79 years old segmented movies of everyday events and attempted to remember the events afterwards. They also completed psychometric ability tests and tests measuring script knowledge for everyday events. Event segmentation and script knowledge both explained unique variance in event memory above and beyond the psychometric measures, and did so as strongly in older as in younger adults. These results suggest that event segmentation is a basic cognitive mechanism, important for memory across the lifespan. PMID:23942350

  3. Considerations for analysis of time-to-event outcomes measured with error: Bias and correction with SIMEX.

    PubMed

    Oh, Eric J; Shepherd, Bryan E; Lumley, Thomas; Shaw, Pamela A

    2018-04-15

    For time-to-event outcomes, a rich literature exists on the bias introduced by covariate measurement error in regression models, such as the Cox model, and methods of analysis to address this bias. By comparison, less attention has been given to understanding the impact or addressing errors in the failure time outcome. For many diseases, the timing of an event of interest (such as progression-free survival or time to AIDS progression) can be difficult to assess or reliant on self-report and therefore prone to measurement error. For linear models, it is well known that random errors in the outcome variable do not bias regression estimates. With nonlinear models, however, even random error or misclassification can introduce bias into estimated parameters. We compare the performance of 2 common regression models, the Cox and Weibull models, in the setting of measurement error in the failure time outcome. We introduce an extension of the SIMEX method to correct for bias in hazard ratio estimates from the Cox model and discuss other analysis options to address measurement error in the response. A formula to estimate the bias induced into the hazard ratio by classical measurement error in the event time for a log-linear survival model is presented. Detailed numerical studies are presented to examine the performance of the proposed SIMEX method under varying levels and parametric forms of the error in the outcome. We further illustrate the method with observational data on HIV outcomes from the Vanderbilt Comprehensive Care Clinic. Copyright © 2017 John Wiley & Sons, Ltd.

  4. gPhoton: Time-tagged GALEX photon events analysis tools

    NASA Astrophysics Data System (ADS)

    Million, Chase C.; Fleming, S. W.; Shiao, B.; Loyd, P.; Seibert, M.; Smith, M.

    2016-03-01

    Written in Python, gPhoton calibrates and sky-projects the ~1.1 trillion ultraviolet photon events detected by the microchannel plates on the Galaxy Evolution Explorer Spacecraft (GALEX), archives these events in a publicly accessible database at the Mikulski Archive for Space Telescopes (MAST), and provides tools for working with the database to extract scientific results, particularly over short time domains. The software includes a re-implementation of core functionality of the GALEX mission calibration pipeline to produce photon list files from raw spacecraft data as well as a suite of command line tools to generate calibrated light curves, images, and movies from the MAST database.

  5. Analysis and verification of a prediction model of solar energetic proton events

    NASA Astrophysics Data System (ADS)

    Wang, J.; Zhong, Q.

    2017-12-01

    The solar energetic particle event can cause severe radiation damages near Earth. The alerts and summary products of the solar energetic proton events were provided by the Space Environment Prediction Center (SEPC) according to the flux of the greater than 10 MeV protons taken by GOES satellite in geosynchronous orbit. The start of a solar energetic proton event is defined as the time when the flux of the greater than 10 MeV protons equals or exceeds 10 proton flux units (pfu). In this study, a model was developed to predict the solar energetic proton events, provide the warning for the solar energetic proton events at least minutes in advance, based on both the soft X-ray flux and integral proton flux taken by GOES. The quality of the forecast model was measured against verifications of accuracy, reliability, discrimination capability, and forecast skills. The peak flux and rise time of the solar energetic proton events in the six channels, >1MeV, >5 MeV, >10 MeV, >30 MeV, >50 MeV, >100 MeV, were also simulated and analyzed.

  6. Event-Based Analysis of Rainfall-Runoff Response to Assess Wetland-Stream Interaction in the Prairie Pothole Region

    NASA Astrophysics Data System (ADS)

    Haque, M. A.; Ross, C.; Schmall, A.; Bansah, S.; Ali, G.

    2016-12-01

    Process-based understanding of wetland response to precipitation is needed to quantify the extent to which non-floodplain wetlands - such as Prairie potholes - generate flow and transmit that flow to nearby streams. While measuring wetland-stream (W-S) interaction is difficult, it is possible to infer it by examining hysteresis characteristics between wetland and stream stage during individual precipitation events. Hence, to evaluate W-S interaction, 10 intact and 10 altered/lost potholes were selected for study; they are located in Broughton's Creek Watershed (Manitoba, Canada) on both sides of a 5 km creek reach. Stilling wells (i.e., above ground wells) were deployed in the intact and altered wetlands to monitor surface water level fluctuations while water table wells were drilled below drainage ditches to a depth of 1 m to monitor shallow groundwater fluctuations. All stilling wells and water table wells were equipped with capacitance water level loggers to monitor fluctuations in surface water and shallow groundwater every 15 minutes. In 2013 (normal year) and 2014 (wet year), 15+ precipitation events were identified and scatter plots of wetland (x-axis) versus stream (y-axis) stage were built to identify W-S hysteretic dynamics. Initial data analysis reveals that in dry antecedent conditions, intact and altered wetlands show clockwise W-S relations, while drained wetlands show anticlockwise W-S hysteresis. However, in wetter antecedent conditions, all wetland types show anticlockwise hysteresis. Future analysis will target the identification of thresholds in antecedent moisture conditions that determine significant changes in event wetland response characteristics (e.g., the delay between the start of rainfall and stream stage, the maximum water level rise in each wetland during each event, the delay between the start of rainfall and peak wetland stage) as well as hysteresis properties (e.g., gradient and area of the hysteresis loop).

  7. Work stress and the risk of recurrent coronary heart disease events: A systematic review and meta-analysis.

    PubMed

    Li, Jian; Zhang, Min; Loerbroks, Adrian; Angerer, Peter; Siegrist, Johannes

    2015-01-01

    Though much evidence indicates that work stress increases the risk of incident of coronary heart disease (CHD), little is known about the role of work stress in the development of recurrent CHD events. The objective of this study was to review and synthesize the existing epidemiological evidence on whether work stress increases the risk of recurrent CHD events in patients with the first CHD. A systematic literature search in the PubMed database (January 1990 - December 2013) for prospective studies was performed. Inclusion criteria included: peer-reviewed English papers with original data, studies with substantial follow-up (> 3 years), end points defined as cardiac death or nonfatal myocardial infarction, as well as work stress assessed with reliable and valid instruments. Meta-analysis using random-effects modeling was conducted in order to synthesize the observed effects across the studies. Five papers derived from 4 prospective studies conducted in Sweden and Canada were included in this systematic review. The measurement of work stress was based on the Demand- Control model (4 papers) or the Effort-Reward Imbalance model (1 paper). According to the estimation by meta-analysis based on 4 papers, a significant effect of work stress on the risk of recurrent CHD events (hazard ratio: 1.65, 95% confidence interval: 1.23-2.22) was observed. Our findings suggest that, in patients with the first CHD, work stress is associated with an increased relative risk of recurrent CHD events by 65%. Due to the limited literature, more well-designed prospective research is needed to examine this association, in particular, from other than western regions of the world. This work is available in Open Access model and licensed under a CC BY-NC 3.0 PL license.

  8. Wood anatomical analysis of Alnus incana and Betula pendula injured by a debris-flow event.

    PubMed

    Arbellay, Estelle; Stoffel, Markus; Bollschweiler, Michelle

    2010-10-01

    Vessel chronologies in ring-porous species have been successfully employed in the past to extract the climate signal from tree rings. Environmental signals recorded in vessels of ring-porous species have also been used in previous studies to reconstruct discrete events of drought, flooding and insect defoliation. However, very little is known about the ability of diffuse-porous species to record environmental signals in their xylem cells. Moreover, time series of wood anatomical features have only rarely been used to reconstruct former geomorphic events. This study was therefore undertaken to characterize the wood anatomical response of diffuse-porous Alnus incana (L.) Moench and Betula pendula Roth to debris-flow-induced wounding. Tree microscopic response to wounding was assessed through the analysis of wood anatomical differences between injured rings formed in the debris-flow event year and uninjured rings formed in the previous year. The two ring types were examined close and opposite to the injury in order to determine whether wound effects on xylem cells decrease with increasing tangential distance from the injury. Image analysis was used to measure vessel parameters as well as fiber and parenchyma cell (FPC) parameters. The results of this study indicate that injured rings are characterized by smaller vessels as compared with uninjured rings. By contrast, FPC parameters were not found to significantly differ between injured and uninjured rings. Vessel and FPC parameters mainly remained constant with increasing tangential distance from the injury, except for a higher proportion of vessel lumen area opposite to the injury within A. incana. This study highlights the existence of anatomical tree-ring signatures-in the form of smaller vessels-related to past debris-flow activity and addresses a new methodological approach to date injuries inflicted on trees by geomorphic processes.

  9. Systematic inference of functional phosphorylation events in yeast metabolism.

    PubMed

    Chen, Yu; Wang, Yonghong; Nielsen, Jens

    2017-07-01

    Protein phosphorylation is a post-translational modification that affects proteins by changing their structure and conformation in a rapid and reversible way, and it is an important mechanism for metabolic regulation in cells. Phosphoproteomics enables high-throughput identification of phosphorylation events on metabolic enzymes, but identifying functional phosphorylation events still requires more detailed biochemical characterization. Therefore, development of computational methods for investigating unknown functions of a large number of phosphorylation events identified by phosphoproteomics has received increased attention. We developed a mathematical framework that describes the relationship between phosphorylation level of a metabolic enzyme and the corresponding flux through the enzyme. Using this framework, it is possible to quantitatively estimate contribution of phosphorylation events to flux changes. We showed that phosphorylation regulation analysis, combined with a systematic workflow and correlation analysis, can be used for inference of functional phosphorylation events in steady and dynamic conditions, respectively. Using this analysis, we assigned functionality to phosphorylation events of 17 metabolic enzymes in the yeast Saccharomyces cerevisiae , among which 10 are novel. Phosphorylation regulation analysis cannot only be extended for inference of other functional post-translational modifications but also be a promising scaffold for multi-omics data integration in systems biology. Matlab codes for flux balance analysis in this study are available in Supplementary material. yhwang@ecust.edu.cn or nielsenj@chalmers.se. Supplementary data are available at Bioinformatics online. © The Author 2017. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com

  10. The analysis of a complex fire event using multispaceborne observations

    NASA Astrophysics Data System (ADS)

    Andrei, Simona; Carstea, Emil; Marmureanu, Luminita; Ene, Dragos; Binietoglou, Ioannis; Nicolae, Doina; Konsta, Dimitra; Amiridis, Vassilis; Proestakis, Emmanouil

    2018-04-01

    This study documents a complex fire event that occurred on October 2016, in Middle East belligerent area. Two fire outbreaks were detected by different spacecraft monitoring instruments on board of TERRA, CALIPSO and AURA Earth Observation missions. Link with local weather conditions was examined using ERA Interim Reanalysis and CAMS datasets. The detection of the event by multiple sensors enabled a detailed characterization of fires and the comparison with different observational data.

  11. Cardiovascular safety of linagliptin in type 2 diabetes: a comprehensive patient-level pooled analysis of prospectively adjudicated cardiovascular events.

    PubMed

    Rosenstock, Julio; Marx, Nikolaus; Neubacher, Dietmar; Seck, Thomas; Patel, Sanjay; Woerle, Hans-Juergen; Johansen, Odd Erik

    2015-05-21

    The cardiovascular (CV) safety of linagliptin was evaluated in subjects with type 2 diabetes (T2DM). Pre-specified patient-level pooled analysis of all available double-blind, randomized, controlled trials, ≥ 12 weeks' duration (19 trials, 9459 subjects) of linagliptin versus placebo/active treatment. Primary end point: composite of prospectively adjudicated CV death, non-fatal myocardial infarction, non-fatal stroke, and hospitalization for unstable angina (4P-MACE). Hospitalization for congestive heart failure (CHF) was also evaluated; adjudication of CHF was introduced during the phase 3 program (8 trials; 3314 subjects). 4P-MACE was assessed in placebo-controlled trials (subgroup of 18 trials; 7746 subjects). Investigator-reported events suggestive of CHF from 24 placebo-controlled trials (including trials <12 weeks' duration, 8778 subjects) were also analyzed. 5847 patients received linagliptin (5 mg: 5687, 10 mg: 160) and 3612 comparator (glimepiride: 775, voglibose: 162, placebo: 2675); cumulative exposure, 4421.3 and 3254.7 patient-years, respectively. 4P-MACE incidence rates: 13.4 per 1000 patient-years, linagliptin (60 events), 18.9, total comparators (62 events); overall hazard ratio (HR), 0.78 (95% confidence interval [CI], 0.55-1.12). HR for adjudicated hospitalization for CHF (n = 21): 1.04 (0.43-2.47). For placebo-controlled trials, 4P-MACE incidence rates: 14.9 per 1000 patient-years, linagliptin (43 events), 16.4, total comparators (29 events); overall HR, 1.09 (95% CI, 0.68-1.75). Occurrence of investigator-reported events suggestive of CHF was low for linagliptin- (26 events, 0.5%; serious: 16 events, 0.3%) and placebo-treated (8 events, 0.2%; serious: 6 events, 0.2%) patients. Linagliptin is not associated with increased CV risk versus pooled active comparators or placebo in patients with T2DM.

  12. Frequency Domain Analysis of Sensor Data for Event Classification in Real-Time Robot Assisted Deburring

    PubMed Central

    Pappachan, Bobby K; Caesarendra, Wahyu; Tjahjowidodo, Tegoeh; Wijaya, Tomi

    2017-01-01

    Process monitoring using indirect methods relies on the usage of sensors. Using sensors to acquire vital process related information also presents itself with the problem of big data management and analysis. Due to uncertainty in the frequency of events occurring, a higher sampling rate is often used in real-time monitoring applications to increase the chances of capturing and understanding all possible events related to the process. Advanced signal processing methods are used to further decipher meaningful information from the acquired data. In this research work, power spectrum density (PSD) of sensor data acquired at sampling rates between 40–51.2 kHz was calculated and the corelation between PSD and completed number of cycles/passes is presented. Here, the progress in number of cycles/passes is the event this research work intends to classify and the algorithm used to compute PSD is Welch’s estimate method. A comparison between Welch’s estimate method and statistical methods is also discussed. A clear co-relation was observed using Welch’s estimate to classify the number of cycles/passes. The paper also succeeds in classifying vibration signal generated by the spindle from the vibration signal acquired during finishing process. PMID:28556809

  13. Event segmentation ability uniquely predicts event memory.

    PubMed

    Sargent, Jesse Q; Zacks, Jeffrey M; Hambrick, David Z; Zacks, Rose T; Kurby, Christopher A; Bailey, Heather R; Eisenberg, Michelle L; Beck, Taylor M

    2013-11-01

    Memory for everyday events plays a central role in tasks of daily living, autobiographical memory, and planning. Event memory depends in part on segmenting ongoing activity into meaningful units. This study examined the relationship between event segmentation and memory in a lifespan sample to answer the following question: Is the ability to segment activity into meaningful events a unique predictor of subsequent memory, or is the relationship between event perception and memory accounted for by general cognitive abilities? Two hundred and eight adults ranging from 20 to 79years old segmented movies of everyday events and attempted to remember the events afterwards. They also completed psychometric ability tests and tests measuring script knowledge for everyday events. Event segmentation and script knowledge both explained unique variance in event memory above and beyond the psychometric measures, and did so as strongly in older as in younger adults. These results suggest that event segmentation is a basic cognitive mechanism, important for memory across the lifespan. Copyright © 2013 Elsevier B.V. All rights reserved.

  14. Immune-related adverse events for anti-PD-1 and anti-PD-L1 drugs: systematic review and meta-analysis

    PubMed Central

    Baxi, Shrujal; Yang, Annie; Gennarelli, Renee L; Khan, Niloufer; Wang, Ziwei; Boyce, Lindsay

    2018-01-01

    Abstract Objective To evaluate rates of serious organ specific immune-related adverse events, general adverse events related to immune activation, and adverse events consistent with musculoskeletal problems for anti-programmed cell death 1 (PD-1) drugs overall and compared with control treatments. Design Systematic review and meta-analysis. Data sources Medline, Embase, Cochrane Library, Web of Science, and Scopus searched to 16 March 2017 and combined with data from ClinicalTrials.gov. Study selection Eligible studies included primary clinical trial data on patients with cancer with recurrent or metastatic disease. Data extraction Three independent investigators extracted data on adverse events from ClinicalTrials.gov and the published studies. Risk of bias was assessed using the Cochrane tool by three independent investigators. Results 13 relevant studies were included; adverse event data were available on ClinicalTrials.gov for eight. Studies compared nivolumab (n=6), pembrolizumab (5), or atezolizumab (2) with chemotherapy (11), targeted drugs (1), or both (1). Serious organ specific immune-related adverse events were rare, but compared with standard treatment, rates of hypothyroidism (odds ratio 7.56, 95% confidence interval 4.53 to 12.61), pneumonitis (5.37, 2.73 to 10.56), colitis (2.88, 1.30 to 6.37), and hypophysitis (3.38, 1.02 to 11.08) were increased with anti-PD-1 drugs. Of the general adverse events related to immune activation, only the rate of rash (2.34, 2.73 to 10.56) increased. Incidence of fatigue (32%) and diarrhea (19%) were high but similar to control. Reporting of adverse events consistent with musculoskeletal problems was inconsistent; rates varied but were over 20% in some studies for arthraligia and back pain. Conclusions Organ specific immune-related adverse events are uncommon with anti-PD-1 drugs but the risk is increased compared with control treatments. General adverse events related to immune activation are largely similar. Adverse

  15. Immune-related adverse events for anti-PD-1 and anti-PD-L1 drugs: systematic review and meta-analysis.

    PubMed

    Baxi, Shrujal; Yang, Annie; Gennarelli, Renee L; Khan, Niloufer; Wang, Ziwei; Boyce, Lindsay; Korenstein, Deborah

    2018-03-14

    To evaluate rates of serious organ specific immune-related adverse events, general adverse events related to immune activation, and adverse events consistent with musculoskeletal problems for anti-programmed cell death 1 (PD-1) drugs overall and compared with control treatments. Systematic review and meta-analysis. Medline, Embase, Cochrane Library, Web of Science, and Scopus searched to 16 March 2017 and combined with data from ClinicalTrials.gov. Eligible studies included primary clinical trial data on patients with cancer with recurrent or metastatic disease. Three independent investigators extracted data on adverse events from ClinicalTrials.gov and the published studies. Risk of bias was assessed using the Cochrane tool by three independent investigators. 13 relevant studies were included; adverse event data were available on ClinicalTrials.gov for eight. Studies compared nivolumab (n=6), pembrolizumab (5), or atezolizumab (2) with chemotherapy (11), targeted drugs (1), or both (1). Serious organ specific immune-related adverse events were rare, but compared with standard treatment, rates of hypothyroidism (odds ratio 7.56, 95% confidence interval 4.53 to 12.61), pneumonitis (5.37, 2.73 to 10.56), colitis (2.88, 1.30 to 6.37), and hypophysitis (3.38, 1.02 to 11.08) were increased with anti-PD-1 drugs. Of the general adverse events related to immune activation, only the rate of rash (2.34, 2.73 to 10.56) increased. Incidence of fatigue (32%) and diarrhea (19%) were high but similar to control. Reporting of adverse events consistent with musculoskeletal problems was inconsistent; rates varied but were over 20% in some studies for arthraligia and back pain. Organ specific immune-related adverse events are uncommon with anti-PD-1 drugs but the risk is increased compared with control treatments. General adverse events related to immune activation are largely similar. Adverse events consistent with musculoskeletal problems are inconsistently reported but adverse

  16. Corrective interpersonal experience in psychodrama group therapy: a comprehensive process analysis of significant therapeutic events.

    PubMed

    McVea, Charmaine S; Gow, Kathryn; Lowe, Roger

    2011-07-01

    This study investigated the process of resolving painful emotional experience during psychodrama group therapy, by examining significant therapeutic events within seven psychodrama enactments. A comprehensive process analysis of four resolved and three not-resolved cases identified five meta-processes which were linked to in-session resolution. One was a readiness to engage in the therapeutic process, which was influenced by client characteristics and the client's experience of the group; and four were therapeutic events: (1) re-experiencing with insight; (2) activating resourcefulness; (3) social atom repair with emotional release; and (4) integration. A corrective interpersonal experience (social atom repair) healed the sense of fragmentation and interpersonal disconnection associated with unresolved emotional pain, and emotional release was therapeutically helpful when located within the enactment of this new role relationship. Protagonists who experienced resolution reported important improvements in interpersonal functioning and sense of self which they attributed to this experience.

  17. Mitigation of Manhole Events Caused by Secondary Cable Failure

    NASA Astrophysics Data System (ADS)

    Zhang, Lili

    "Manhole event" refers to a range of phenomena, such as smokers, fires and explosions which occur on underground electrical infrastructure, primarily in major cities. The most common cause of manhole events is decomposition of secondary cable initiated by an electric fault. The work presented in this thesis addresses various aspects related to the evolution and mitigation of the manhole events caused by secondary cable insulation failure. Manhole events develop as a result of thermal decomposition of organic materials present in the cable duct and manholes. Polymer characterization techniques are applied to intensively study the materials properties as related to manhole events, mainly the thermal decomposition behaviors of the polymers present in the cable duct. Though evolved gas analysis, the combustible gases have been quantitatively identified. Based on analysis and knowledge of field conditions, manhole events is divided into at least two classes, those in which exothermic chemical reactions dominate and those in which electrical energy dominates. The more common form of manhole event is driven by air flow down the duct. Numerical modeling of smolder propagation in the cable duct demonstrated that limiting air flow is effective in reducing the generation rate of combustible gas, in other words, limiting manhole events to relatively minor "smokers". Besides manhole events, another by-product of secondary cable insulation breakdown is stray voltage. The danger to personnel due to stray voltage is mostly caused by the 'step potential'. The amplitude of step potential as a result of various types of insulation defects is calculated using Finite Element Analysis (FEA) program.

  18. Adverse events after anthrax vaccination reported to the Vaccine Adverse Event Reporting System (VAERS), 1990-2007.

    PubMed

    Niu, Manette T; Ball, Robert; Woo, Emily Jane; Burwen, Dale R; Knippen, Maureen; Braun, M Miles

    2009-01-07

    During the period March 1, 1998 to January 14, 2007, approximately 6 million doses of Anthrax vaccine adsorbed (AVA) vaccine were administered. As of January 16, 2007, 4753 reports of adverse events following receipt of AVA vaccination had been submitted to the Vaccine Adverse Event Reporting System (VAERS). Taken together, reports to VAERS did not definitively link any serious unexpected risk to this vaccine, and review of death and serious reports did not show a distinctive pattern indicative of a causal relationship to AVA vaccination. Continued monitoring of VAERS and analysis of potential associations between AVA vaccination and rare, serious events is warranted.

  19. The pros and cons of researching events ethnographically

    PubMed Central

    2017-01-01

    Events (remarkable, disruptive happenings) are important subjects of study for understanding processes of change. In this essay, I reflect upon the issue of what the ethnographic method has to offer for the analysis of this social phenomenon. To do so, I review three recently published ethnographic studies of events. My conclusion is that it is indeed a very useful method for understanding the feelings and ideas of people who are experiencing eventful situations, for instance around protests or natural disasters. However, using this method also brings about practical difficulties, such as the ‘luck’ that an event occurs at the ethnographic fieldwork site. Next, as transformative responses to events are not bound by the place or time of the happening, other methods (interviews, discourse analysis, surveys) that make it easier to follow them in varying locations and periods might be more suitable for getting a comprehensive picture of their meaning-making dynamics. PMID:29081715

  20. What drives high flow events in the Swiss Alps? Recent developments in wavelet spectral analysis and their application to hydrology

    NASA Astrophysics Data System (ADS)

    Schaefli, B.; Maraun, D.; Holschneider, M.

    2007-12-01

    Extreme hydrological events are often triggered by exceptional co-variations of the relevant hydrometeorological processes and in particular by exceptional co-oscillations at various temporal scales. Wavelet and cross wavelet spectral analysis offers promising time-scale resolved analysis methods to detect and analyze such exceptional co-oscillations. This paper presents the state-of-the-art methods of wavelet spectral analysis, discusses related subtleties, potential pitfalls and recently developed solutions to overcome them and shows how wavelet spectral analysis, if combined to a rigorous significance test, can lead to reliable new insights into hydrometeorological processes for real-world applications. The presented methods are applied to detect potentially flood triggering situations in a high Alpine catchment for which a recent re-estimation of design floods encountered significant problems simulating the observed high flows. For this case study, wavelet spectral analysis of precipitation, temperature and discharge offers a powerful tool to help detecting potentially flood producing meteorological situations and to distinguish between different types of floods with respect to the prevailing critical hydrometeorological conditions. This opens very new perspectives for the analysis of model performances focusing on the occurrence and non-occurrence of different types of high flow events. Based on the obtained results, the paper summarizes important recommendations for future applications of wavelet spectral analysis in hydrology.

  1. Modeling propensity to move after job change using event history analysis and temporal GIS

    NASA Astrophysics Data System (ADS)

    Vandersmissen, Marie-Hélène; Séguin, Anne-Marie; Thériault, Marius; Claramunt, Christophe

    2009-03-01

    The research presented in this paper analyzes the emergent residential behaviors of individual actors in a context of profound social changes in the work sphere. It incorporates a long-term view in the analysis of the relationships between social changes in the work sphere and these behaviors. The general hypothesis is that social changes produce complex changes in the long-term dynamics of residential location behavior. More precisely, the objective of this paper is to estimate the propensity for professional workers to move house after a change of workplace. Our analysis draws on data from a biographical survey using a retrospective questionnaire that enables a posteriori reconstitution of the familial, professional and residential lifelines of professional workers since their departure from their parents’ home. The survey was conducted in 1996 in the Quebec City Metropolitan Area, which, much like other Canadian cities, has experienced a substantial increase in “unstable” work, even for professionals. The approach is based on event history analysis, a Temporal Geographic Information System and exploratory spatial analysis of model’s residuals. Results indicate that 48.9% of respondents moved after a job change and that the most important factors influencing the propensity to move house after a job change are home tenure (for lone adults as for couple) and number of children (for couples only). We also found that moving is associated with changing neighborhood for owners while tenants or co-tenants tend to stay in the same neighborhood. The probability of moving 1 year after a job change is 0.10 for lone adults and couples while after 2 years, the household structure seems to have an impact: the probability increased to 0.23 for lone adults and to 0.21 for couples. The outcome of this research contributes to furthering our understanding of a familial decision (to move) following a professional event (change of job), controlling for household structure

  2. Structural monitoring for rare events in remote locations

    NASA Astrophysics Data System (ADS)

    Hale, J. M.

    2005-01-01

    A structural monitoring system has been developed for use on high value engineering structures, which is particularly suitable for use in remote locations where rare events such as accidental impacts, seismic activity or terrorist attack might otherwise go undetected. The system comprises a low power intelligent on-site data logger and a remote analysis computer that communicate with one another using the internet and mobile telephone technology. The analysis computer also generates e-mail alarms and maintains a web page that displays detected events in near real-time to authorised users. The application of the prototype system to pipeline monitoring is described in which the analysis of detected events is used to differentiate between impacts and pressure surges. The system has been demonstrated successfully and is ready for deployment.

  3. The effect of folic acid based homocysteine lowering on cardiovascular events in people with kidney disease: systematic review and meta-analysis.

    PubMed

    Jardine, Meg J; Kang, Amy; Zoungas, Sophia; Navaneethan, Sankar D; Ninomiya, Toshiharu; Nigwekar, Sagar U; Gallagher, Martin P; Cass, Alan; Strippoli, Giovanni; Perkovic, Vlado

    2012-06-13

    To systematically review the effect of folic acid based homocysteine lowering on cardiovascular outcomes in people with kidney disease. Systematic review and meta-analysis. Medline, Embase, the Cochrane Library, and ClinicalTrials.gov to June 2011. Randomised trials in people with non-dialysis dependent chronic kidney disease or end stage kidney disease or with a functioning kidney transplant reporting at least 100 patient years of follow-up and assessing the effect of folic acid based homocysteine lowering therapy. No language restrictions were applied. Two reviewers independently extracted data on study setting, design, and outcomes using a standardised form. The primary endpoint was cardiovascular events (myocardial infarction, stroke, and cardiovascular mortality, or as defined by study author). Secondary endpoints included the individual composite components, all cause mortality, access thrombosis, requirement for renal replacement therapy, and reported adverse events, including haematological and neurological events. The effect of folic acid based homocysteine lowering on outcomes was assessed with meta-analysis using random effects models. 11 trials were identified that reported on 4389 people with chronic kidney disease, 2452 with end stage kidney disease, and 4110 with functioning kidney transplants (10,951 participants in total). Folic acid based homocysteine therapy did not prevent cardiovascular events (relative risk 0.97, 95% confidence interval 0.92 to 1.03, P = 0.326) or any of the secondary outcomes. There was no evidence of heterogeneity in subgroup analyses, including those of kidney disease category, background fortification, rates of pre-existing disease, or baseline homocysteine level. The definitions of chronic kidney disease varied widely between the studies. Non-cardiovascular events could not be analysed as few studies reported these outcomes. Folic acid based homocysteine lowering does not reduce cardiovascular events in people with kidney

  4. Clinical usefulness and feasibility of time-frequency analysis of chemosensory event-related potentials.

    PubMed

    Huart, C; Rombaux, Ph; Hummel, T; Mouraux, A

    2013-09-01

    The clinical usefulness of olfactory event-related brain potentials (OERPs) to assess olfactory function is limited by the relatively low signal-to-noise ratio of the responses identified using conventional time-domain averaging. Recently, it was shown that time-frequency analysis of the obtained EEG signals can markedly improve the signal-to-noise ratio of OERPs in healthy controls, because it enhances both phase-locked and non phase-locked EEG responses. The aim of the present study was to investigate the clinical usefulness of this approach and evaluate its feasibility in a clinical setting. We retrospectively analysed EEG recordings obtained from 45 patients (15 anosmic, 15 hyposmic and 15 normos- mic). The responses to olfactory stimulation were analysed using conventional time-domain analysis and joint time-frequency analysis. The ability of the two methods to discriminate between anosmic, hyposmic and normosmic patients was assessed using a Receiver Operating Characteristic analysis. The discrimination performance of OERPs identified using conventional time-domain averaging was poor. In contrast, the discrimination performance of the EEG response identified in the time-frequency domain was relatively high. Furthermore, we found a significant correlation between the magnitude of this response and the psychophysical olfactory score. Time-frequency analysis of the EEG responses to olfactory stimulation could be used as an effective and reliable diagnostic tool for the objective clinical evaluation of olfactory function in patients.

  5. Bayesian Analysis for Risk Assessment of Selected Medical Events in Support of the Integrated Medical Model Effort

    NASA Technical Reports Server (NTRS)

    Gilkey, Kelly M.; Myers, Jerry G.; McRae, Michael P.; Griffin, Elise A.; Kallrui, Aditya S.

    2012-01-01

    The Exploration Medical Capability project is creating a catalog of risk assessments using the Integrated Medical Model (IMM). The IMM is a software-based system intended to assist mission planners in preparing for spaceflight missions by helping them to make informed decisions about medical preparations and supplies needed for combating and treating various medical events using Probabilistic Risk Assessment. The objective is to use statistical analyses to inform the IMM decision tool with estimated probabilities of medical events occurring during an exploration mission. Because data regarding astronaut health are limited, Bayesian statistical analysis is used. Bayesian inference combines prior knowledge, such as data from the general U.S. population, the U.S. Submarine Force, or the analog astronaut population located at the NASA Johnson Space Center, with observed data for the medical condition of interest. The posterior results reflect the best evidence for specific medical events occurring in flight. Bayes theorem provides a formal mechanism for combining available observed data with data from similar studies to support the quantification process. The IMM team performed Bayesian updates on the following medical events: angina, appendicitis, atrial fibrillation, atrial flutter, dental abscess, dental caries, dental periodontal disease, gallstone disease, herpes zoster, renal stones, seizure, and stroke.

  6. 'It was a freak accident': an analysis of the labelling of injury events in the US press.

    PubMed

    Smith, Katherine C; Girasek, Deborah C; Baker, Susan P; Manganello, Jennifer A; Bowman, Stephen M; Samuels, Alicia; Gielen, Andrea C

    2012-02-01

    Given that the news media shape our understanding of health issues, a study was undertaken to examine the use by the US media of the expression 'freak accident' in relation to injury events. This analysis is intended to contribute to the ongoing consideration of lay conceptualisation of injuries as 'accidents'. LexisNexis Academic was used to search three purposively selected US news sources (Associated Press, New York Times and Philadelphia Inquirer) for the expression 'freak accident' over 5 years (2005-9). Textual analysis included both structured and open coding. Coding included measures for who used the expression within the story, the nature of the injury event and the injured person(s) being reported upon, incorporation of prevention information within the story and finally a phenomenological consideration of the uses and meanings of the expression within the story context. Results The search yielded a dataset of 250 human injury stories incorporating the term 'freak accident'. Injuries sustained by professional athletes dominated coverage (61%). Fewer than 10% of stories provided a clear and explicit injury prevention message. Stories in which journalists employed the expression 'freak accident' were less likely to include prevention information than stories in which the expression was used by people quoted in the story. Journalists who frame injury events as freak accidents may be an appropriate focus for advocacy efforts. Effective prevention messages should be developed and disseminated to accompany injury reporting in order to educate and protect the public.

  7. Budget impact analysis of thrombolysis for stroke in Spain: a discrete event simulation model.

    PubMed

    Mar, Javier; Arrospide, Arantzazu; Comas, Mercè

    2010-01-01

    Thrombolysis within the first 3 hours after the onset of symptoms of a stroke has been shown to be a cost-effective treatment because treated patients are 30% more likely than nontreated patients to have no residual disability. The objective of this study was to calculate by means of a discrete event simulation model the budget impact of thrombolysis in Spain. The budget impact analysis was based on stroke incidence rates and the estimation of the prevalence of stroke-related disability in Spain and its translation to hospital and social costs. A discrete event simulation model was constructed to represent the flow of patients with stroke in Spain. If 10% of patients with stroke from 2000 to 2015 would receive thrombolytic treatment, the prevalence of dependent patients in 2015 would decrease from 149,953 to 145,922. For the first 6 years, the cost of intervention would surpass the savings. Nevertheless, the number of cases in which patient dependency was avoided would steadily increase, and after 2006 the cost savings would be greater, with a widening difference between the cost of intervention and the cost of nonintervention, until 2015. The impact of thrombolysis on society's health and social budget indicates a net benefit after 6 years, and the improvement in health grows continuously. The validation of the model demonstrates the adequacy of the discrete event simulation approach in representing the epidemiology of stroke to calculate the budget impact.

  8. Big Data Toolsets to Pharmacometrics: Application of Machine Learning for Time-to-Event Analysis.

    PubMed

    Gong, Xiajing; Hu, Meng; Zhao, Liang

    2018-05-01

    Additional value can be potentially created by applying big data tools to address pharmacometric problems. The performances of machine learning (ML) methods and the Cox regression model were evaluated based on simulated time-to-event data synthesized under various preset scenarios, i.e., with linear vs. nonlinear and dependent vs. independent predictors in the proportional hazard function, or with high-dimensional data featured by a large number of predictor variables. Our results showed that ML-based methods outperformed the Cox model in prediction performance as assessed by concordance index and in identifying the preset influential variables for high-dimensional data. The prediction performances of ML-based methods are also less sensitive to data size and censoring rates than the Cox regression model. In conclusion, ML-based methods provide a powerful tool for time-to-event analysis, with a built-in capacity for high-dimensional data and better performance when the predictor variables assume nonlinear relationships in the hazard function. © 2018 The Authors. Clinical and Translational Science published by Wiley Periodicals, Inc. on behalf of American Society for Clinical Pharmacology and Therapeutics.

  9. Passive (Micro-) Seismic Event Detection by Identifying Embedded "Event" Anomalies Within Statistically Describable Background Noise

    NASA Astrophysics Data System (ADS)

    Baziw, Erick; Verbeek, Gerald

    2012-12-01

    Among engineers there is considerable interest in the real-time identification of "events" within time series data with a low signal to noise ratio. This is especially true for acoustic emission analysis, which is utilized to assess the integrity and safety of many structures and is also applied in the field of passive seismic monitoring (PSM). Here an array of seismic receivers are used to acquire acoustic signals to monitor locations where seismic activity is expected: underground excavations, deep open pits and quarries, reservoirs into which fluids are injected or from which fluids are produced, permeable subsurface formations, or sites of large underground explosions. The most important element of PSM is event detection: the monitoring of seismic acoustic emissions is a continuous, real-time process which typically runs 24 h a day, 7 days a week, and therefore a PSM system with poor event detection can easily acquire terabytes of useless data as it does not identify crucial acoustic events. This paper outlines a new algorithm developed for this application, the so-called SEED™ (Signal Enhancement and Event Detection) algorithm. The SEED™ algorithm uses real-time Bayesian recursive estimation digital filtering techniques for PSM signal enhancement and event detection.

  10. Aspirin and the risk of cardiovascular events in atherosclerosis patients with and without prior ischemic events.

    PubMed

    Bavry, Anthony A; Elgendy, Islam Y; Elbez, Yedid; Mahmoud, Ahmed N; Sorbets, Emmanuel; Steg, Philippe Gabriel; Bhatt, Deepak L

    2017-09-01

    The benefit of aspirin among patients with stable atherosclerosis without a prior ischemic event is not well defined. Aspirin would be of benefit in outpatients with atherosclerosis with prior ischemic events, but not in those without ischemic events. Subjects from the Reduction of Atherothrombosis for Continued Health registry were divided according to prior ischemic event (n =21 724) vs stable atherosclerosis, but no prior ischemic event (n = 11 872). Analyses were propensity score matched. Aspirin use was updated at each clinic visit and considered as a time-varying covariate. The primary outcome was the first occurrence of cardiovascular death, myocardial infarction, or stroke. In the group with a prior ischemic event, aspirin use was associated with a marginally lower risk of the primary outcome at a median of 41 months (hazard ratio [HR]: 0.81, 95% confidence interval [CI]: 0.65-1.01, P = 0.06). In the group without a prior ischemic event, aspirin use was not associated with a lower risk of the primary outcome at a median of 36 months (HR: 1.03, 95% CI: 0.73-1.45, P = 0.86). In this observational analysis of outpatients with stable atherosclerosis, aspirin was marginally beneficial among patients with a prior ischemic event; however, there was no apparent benefit among those with no prior ischemic event. © 2017 Wiley Periodicals, Inc.

  11. Systematic review and meta-analysis: Multi-strain probiotics as adjunct therapy for Helicobacter pylori eradication and prevention of adverse events

    PubMed Central

    Huang, Ying; Wang, Lin; Malfertheiner, Peter

    2015-01-01

    Background Eradication rates with triple therapy for Helicobacter pylori infections have currently declined to unacceptable levels worldwide. Newer quadruple therapies are burdened with a high rate of adverse events. Whether multi-strain probiotics can improve eradication rates or diminish adverse events remains uncertain. Methods Relevant publications in which patients with H. pylori infections were randomized to a multi-strain probiotic or control were identified in PubMed, Cochrane Databases, and other sources from 1 January 1960–3 June 2015. Primary outcomes included eradication rates, incidence of any adverse event and the incidence of antibiotic-associated diarrhea. As probiotic efficacy is strain-specific, pooled relative risks and 95% confidence intervals were calculated using meta-analysis stratified by similar multi-strain probiotic mixtures. Results A total of 19 randomized controlled trials (20 treatment arms, n = 2730) assessing one of six mixtures of strains of probiotics were included. Four multi-strain probiotics significantly improved H. pylori eradication rates, five significantly prevented any adverse reactions and three significantly reduced antibiotic-associated diarrhea. Only two probiotic mixtures (Lactobacillus acidophilus/Bifidobacterium animalis and an eight-strain mixture) had significant efficacy for all three outcomes. Conclusions Our meta-analysis found adjunctive use of some multi-strain probiotics may improve H. pylori eradication rates and prevent the development of adverse events and antibiotic-associated diarrhea, but not all mixtures were effective. PMID:27536365

  12. Event-by-event fluctuations and inclusive distributions

    NASA Astrophysics Data System (ADS)

    Bialas, A.; Koch, V.

    1999-06-01

    Event-by-event observables are compared with conventional inclusive measurements. We find that moments of event-by-event fluctuations are closely related to inclusive correlation functions. Implications for upcoming heavy ion experiments are discussed.

  13. Bayesian regression model for recurrent event data with event-varying covariate effects and event effect.

    PubMed

    Lin, Li-An; Luo, Sheng; Davis, Barry R

    2018-01-01

    In the course of hypertension, cardiovascular disease events (e.g., stroke, heart failure) occur frequently and recurrently. The scientific interest in such study may lie in the estimation of treatment effect while accounting for the correlation among event times. The correlation among recurrent event times come from two sources: subject-specific heterogeneity (e.g., varied lifestyles, genetic variations, and other unmeasurable effects) and event dependence (i.e., event incidences may change the risk of future recurrent events). Moreover, event incidences may change the disease progression so that there may exist event-varying covariate effects (the covariate effects may change after each event) and event effect (the effect of prior events on the future events). In this article, we propose a Bayesian regression model that not only accommodates correlation among recurrent events from both sources, but also explicitly characterizes the event-varying covariate effects and event effect. This model is especially useful in quantifying how the incidences of events change the effects of covariates and risk of future events. We compare the proposed model with several commonly used recurrent event models and apply our model to the motivating lipid-lowering trial (LLT) component of the Antihypertensive and Lipid-Lowering Treatment to Prevent Heart Attack Trial (ALLHAT) (ALLHAT-LLT).

  14. Bayesian regression model for recurrent event data with event-varying covariate effects and event effect

    PubMed Central

    Lin, Li-An; Luo, Sheng; Davis, Barry R.

    2017-01-01

    In the course of hypertension, cardiovascular disease events (e.g., stroke, heart failure) occur frequently and recurrently. The scientific interest in such study may lie in the estimation of treatment effect while accounting for the correlation among event times. The correlation among recurrent event times come from two sources: subject-specific heterogeneity (e.g., varied lifestyles, genetic variations, and other unmeasurable effects) and event dependence (i.e., event incidences may change the risk of future recurrent events). Moreover, event incidences may change the disease progression so that there may exist event-varying covariate effects (the covariate effects may change after each event) and event effect (the effect of prior events on the future events). In this article, we propose a Bayesian regression model that not only accommodates correlation among recurrent events from both sources, but also explicitly characterizes the event-varying covariate effects and event effect. This model is especially useful in quantifying how the incidences of events change the effects of covariates and risk of future events. We compare the proposed model with several commonly used recurrent event models and apply our model to the motivating lipid-lowering trial (LLT) component of the Antihypertensive and Lipid-Lowering Treatment to Prevent Heart Attack Trial (ALLHAT) (ALLHAT-LLT). PMID:29755162

  15. Arrests, Recent Life Circumstances, and Recurrent Job Loss for At-Risk Young Men: An Event-History Analysis

    ERIC Educational Resources Information Center

    Wiesner, Margit; Capaldi, Deborah M.; Kim, Hyoun K.

    2010-01-01

    This study used longitudinal data from 202 at-risk young men to examine effects of arrests, prior risk factors, and recent life circumstances on job loss across a 7-year period in early adulthood. Repeated failure-time continuous event-history analysis indicated that occurrence of job loss was primarily related to prior mental health problems,…

  16. Sensitivity analysis of the FEMA HAZUS-MH MR4 Earthquake Model using seismic events affecting King County Washington

    NASA Astrophysics Data System (ADS)

    Neighbors, C.; Noriega, G. R.; Caras, Y.; Cochran, E. S.

    2010-12-01

    HAZUS-MH MR4 (HAZards U. S. Multi-Hazard Maintenance Release 4) is a risk-estimation software developed by FEMA to calculate potential losses due to natural disasters. Federal, state, regional, and local government use the HAZUS-MH Earthquake Model for earthquake risk mitigation, preparedness, response, and recovery planning (FEMA, 2003). In this study, we examine several parameters used by the HAZUS-MH Earthquake Model methodology to understand how modifying the user-defined settings affect ground motion analysis, seismic risk assessment and earthquake loss estimates. This analysis focuses on both shallow crustal and deep intraslab events in the American Pacific Northwest. Specifically, the historic 1949 Mw 6.8 Olympia, 1965 Mw 6.6 Seattle-Tacoma and 2001 Mw 6.8 Nisqually normal fault intraslab events and scenario large-magnitude Seattle reverse fault crustal events are modeled. Inputs analyzed include variations of deterministic event scenarios combined with hazard maps and USGS ShakeMaps. This approach utilizes the capacity of the HAZUS-MH Earthquake Model to define landslide- and liquefaction- susceptibility hazards with local groundwater level and slope stability information. Where Shakemap inputs are not used, events are run in combination with NEHRP soil classifications to determine site amplification effects. The earthquake component of HAZUS-MH applies a series of empirical ground motion attenuation relationships developed from source parameters of both regional and global historical earthquakes to estimate strong ground motion. Ground motion and resulting ground failure due to earthquakes are then used to calculate, direct physical damage for general building stock, essential facilities, and lifelines, including transportation systems and utility systems. Earthquake losses are expressed in structural, economic and social terms. Where available, comparisons between recorded earthquake losses and HAZUS-MH earthquake losses are used to determine how region

  17. Analyzing System on A Chip Single Event Upset Responses using Single Event Upset Data, Classical Reliability Models, and Space Environment Data

    NASA Technical Reports Server (NTRS)

    Berg, Melanie; LaBel, Kenneth; Campola, Michael; Xapsos, Michael

    2017-01-01

    We are investigating the application of classical reliability performance metrics combined with standard single event upset (SEU) analysis data. We expect to relate SEU behavior to system performance requirements. Our proposed methodology will provide better prediction of SEU responses in harsh radiation environments with confidence metrics. single event upset (SEU), single event effect (SEE), field programmable gate array devises (FPGAs)

  18. Real-time detection of organic contamination events in water distribution systems by principal components analysis of ultraviolet spectral data.

    PubMed

    Zhang, Jian; Hou, Dibo; Wang, Ke; Huang, Pingjie; Zhang, Guangxin; Loáiciga, Hugo

    2017-05-01

    The detection of organic contaminants in water distribution systems is essential to protect public health from potential harmful compounds resulting from accidental spills or intentional releases. Existing methods for detecting organic contaminants are based on quantitative analyses such as chemical testing and gas/liquid chromatography, which are time- and reagent-consuming and involve costly maintenance. This study proposes a novel procedure based on discrete wavelet transform and principal component analysis for detecting organic contamination events from ultraviolet spectral data. Firstly, the spectrum of each observation is transformed using discrete wavelet with a coiflet mother wavelet to capture the abrupt change along the wavelength. Principal component analysis is then employed to approximate the spectra based on capture and fusion features. The significant value of Hotelling's T 2 statistics is calculated and used to detect outliers. An alarm of contamination event is triggered by sequential Bayesian analysis when the outliers appear continuously in several observations. The effectiveness of the proposed procedure is tested on-line using a pilot-scale setup and experimental data.

  19. Methods for external event screening quantification: Risk Methods Integration and Evaluation Program (RMIEP) methods development

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ravindra, M.K.; Banon, H.

    1992-07-01

    In this report, the scoping quantification procedures for external events in probabilistic risk assessments of nuclear power plants are described. External event analysis in a PRA has three important goals; (1) the analysis should be complete in that all events are considered; (2) by following some selected screening criteria, the more significant events are identified for detailed analysis; (3) the selected events are analyzed in depth by taking into account the unique features of the events: hazard, fragility of structures and equipment, external-event initiated accident sequences, etc. Based on the above goals, external event analysis may be considered as amore » three-stage process: Stage I: Identification and Initial Screening of External Events; Stage II: Bounding Analysis; Stage III: Detailed Risk Analysis. In the present report, first, a review of published PRAs is given to focus on the significance and treatment of external events in full-scope PRAs. Except for seismic, flooding, fire, and extreme wind events, the contributions of other external events to plant risk have been found to be negligible. Second, scoping methods for external events not covered in detail in the NRC's PRA Procedures Guide are provided. For this purpose, bounding analyses for transportation accidents, extreme winds and tornadoes, aircraft impacts, turbine missiles, and chemical release are described.« less

  20. Methods for external event screening quantification: Risk Methods Integration and Evaluation Program (RMIEP) methods development

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ravindra, M.K.; Banon, H.

    1992-07-01

    In this report, the scoping quantification procedures for external events in probabilistic risk assessments of nuclear power plants are described. External event analysis in a PRA has three important goals; (1) the analysis should be complete in that all events are considered; (2) by following some selected screening criteria, the more significant events are identified for detailed analysis; (3) the selected events are analyzed in depth by taking into account the unique features of the events: hazard, fragility of structures and equipment, external-event initiated accident sequences, etc. Based on the above goals, external event analysis may be considered as amore » three-stage process: Stage I: Identification and Initial Screening of External Events; Stage II: Bounding Analysis; Stage III: Detailed Risk Analysis. In the present report, first, a review of published PRAs is given to focus on the significance and treatment of external events in full-scope PRAs. Except for seismic, flooding, fire, and extreme wind events, the contributions of other external events to plant risk have been found to be negligible. Second, scoping methods for external events not covered in detail in the NRC`s PRA Procedures Guide are provided. For this purpose, bounding analyses for transportation accidents, extreme winds and tornadoes, aircraft impacts, turbine missiles, and chemical release are described.« less

  1. The DPRK event of May 25, 2009: an analysis carried out at INGV from a multidisciplinary perspective

    NASA Astrophysics Data System (ADS)

    Console, R.; Carluccio, R.; Chiappini, M.; Chiappini, S.; D'Ajello Caracciolo, F.; Damiani, K.; Giuntini, A.; Materni, V.; Pignatelli, A.

    2009-12-01

    On the early morning of May 25, 2009 INGV detected an event located in a predefined area of interest, including the DPRK. Such detection triggered the request of raw data from the seismological International Agencies that operate global seismic networks. Around 6:00 UTC of the same morning, the INGV staff started the standard procedures of seismological analysis on the data collected from such Agencies, in order to locate, identify and characterize the event from a National perspective. At the same time, the DPRK Government announced the conduction of an underground nuclear test in their territory, confirming the suspected explosive nature of the seismic event. The seismological analysis carried out at the INGV included hypocentral location, mb and Ms computation, application of identification criteria developed at the INGV, and estimation of a possible range for the yield. Here the basic parameters for the event, as obtained at the INGV are reported: Origin time: 2009/05/25 00:54:43.039 Latitude: 41.286 deg. N +/- 12.214 km. Longitude: 129.174 deg. E +/- 14.767 km. Depth: 0 (fixed by the analyst) mb: 4.5 +/- 0.1; Ms: 3.2 +/- 0.2 The criteria adopted at the INGV for event screening led us to classify this event as an explosion with high probability. To reach this conclusion, a rigorous statistical method known as “Discriminant Analysis” has been applied. Particular care has been devoted to the comparison with the nuclear test announced by the DPRK on 9 October 2006. The two locations appear very close to each other (within a distance of the order of 10 km), with the respective error ellipses nearly totally overlapping (the error ellipse area of the recent event is smaller due to the better quality and more numerous recordings). A more accurate relative location has been carried out by the application of the algorithm of DDJHD specifically developed at the INGV for these purposes. In this case the epicentral distance drops to less than 3 km, with an error of 3

  2. Multivariate Regression Analysis of Winter Ozone Events in the Uinta Basin of Eastern Utah, USA

    NASA Astrophysics Data System (ADS)

    Mansfield, M. L.

    2012-12-01

    I report on a regression analysis of a number of variables that are involved in the formation of winter ozone in the Uinta Basin of Eastern Utah. One goal of the analysis is to develop a mathematical model capable of predicting the daily maximum ozone concentration from values of a number of independent variables. The dependent variable is the daily maximum ozone concentration at a particular site in the basin. Independent variables are (1) daily lapse rate, (2) daily "basin temperature" (defined below), (3) snow cover, (4) midday solar zenith angle, (5) monthly oil production, (6) monthly gas production, and (7) the number of days since the beginning of a multi-day inversion event. Daily maximum temperature and daily snow cover data are available at ten or fifteen different sites throughout the basin. The daily lapse rate is defined operationally as the slope of the linear least-squares fit to the temperature-altitude plot, and the "basin temperature" is defined as the value assumed by the same least-squares line at an altitude of 1400 m. A multi-day inversion event is defined as a set of consecutive days for which the lapse rate remains positive. The standard deviation in the accuracy of the model is about 10 ppb. The model has been combined with historical climate and oil & gas production data to estimate historical ozone levels.

  3. Detection of goal events in soccer videos

    NASA Astrophysics Data System (ADS)

    Kim, Hyoung-Gook; Roeber, Steffen; Samour, Amjad; Sikora, Thomas

    2005-01-01

    In this paper, we present an automatic extraction of goal events in soccer videos by using audio track features alone without relying on expensive-to-compute video track features. The extracted goal events can be used for high-level indexing and selective browsing of soccer videos. The detection of soccer video highlights using audio contents comprises three steps: 1) extraction of audio features from a video sequence, 2) event candidate detection of highlight events based on the information provided by the feature extraction Methods and the Hidden Markov Model (HMM), 3) goal event selection to finally determine the video intervals to be included in the summary. For this purpose we compared the performance of the well known Mel-scale Frequency Cepstral Coefficients (MFCC) feature extraction method vs. MPEG-7 Audio Spectrum Projection feature (ASP) extraction method based on three different decomposition methods namely Principal Component Analysis( PCA), Independent Component Analysis (ICA) and Non-Negative Matrix Factorization (NMF). To evaluate our system we collected five soccer game videos from various sources. In total we have seven hours of soccer games consisting of eight gigabytes of data. One of five soccer games is used as the training data (e.g., announcers' excited speech, audience ambient speech noise, audience clapping, environmental sounds). Our goal event detection results are encouraging.

  4. Evaluation of thromboembolic events in cancer patients receiving bevacizumab according to the Japanese Adverse Drug Event Report database.

    PubMed

    Matsumura, Chikako; Chisaki, Yugo; Sakimoto, Satoko; Sakae, Honoka; Yano, Yoshitaka

    2018-01-01

    Purpose We aimed to examine the risk factors, time of onset, incidence rates, and outcomes of thromboembolic events induced by bevacizumab in patients with cancer using the Japanese Adverse Drug Event Report (JADER) database of the Pharmaceuticals and Medical Devices Agency. Methods Adverse event data recorded in the JADER database between January 2004 and January 2015 were used. After screening the data using the generic drug name bevacizumab, patient data were classified into two groups by age and five groups by cancer type. The histories of disorders were also categorized. Arterial thromboembolic event and venous thromboembolic event were classified as "favorable" or "unfavorable" outcomes. Results In total, 6076 patients were reported to have developed adverse events during the sample period, of which 233 and 453 developed arterial thromboembolic event and venous thromboembolic event, respectively. Logistic analysis suggested that the presence of cancer was a significant risk factor for both arterial thromboembolic event and venous thromboembolic event. Age (≥70 years), histories of either hypertension or diabetes mellitus were also risk factors for arterial thromboembolic event. Median cumulative times of onset for arterial thromboembolic event and venous thromboembolic event were 60 and 80 days, respectively, and were not significantly different by the log-rank test. By the chi-square test, the rate of unfavorable outcomes was found to be higher after developing arterial thromboembolic event than after venous thromboembolic event. Conclusion Thromboembolism is a leading cause of mortality in patients with cancer. Patients should be monitored for the symptoms of thromboembolic events right from the initial stages of bevacizumab treatment.

  5. Top-down proteomics for the analysis of proteolytic events - Methods, applications and perspectives.

    PubMed

    Tholey, Andreas; Becker, Alexander

    2017-11-01

    Mass spectrometry based proteomics is an indispensable tool for almost all research areas relevant for the understanding of proteolytic processing, ranging from the identification of substrates, products and cleavage sites up to the analysis of structural features influencing protease activity. The majority of methods for these studies are based on bottom-up proteomics performing analysis at peptide level. As this approach is characterized by a number of pitfalls, e.g. loss of molecular information, there is an ongoing effort to establish top-down proteomics, performing separation and MS analysis both at intact protein level. We briefly introduce major approaches of bottom-up proteomics used in the field of protease research and highlight the shortcomings of these methods. We then discuss the present state-of-the-art of top-down proteomics. Together with the discussion of known challenges we show the potential of this approach and present a number of successful applications of top-down proteomics in protease research. This article is part of a Special Issue entitled: Proteolysis as a Regulatory Event in Pathophysiology edited by Stefan Rose-John. Copyright © 2017 Elsevier B.V. All rights reserved.

  6. Hypersensitivity Events, Including Potentially Hypersensitivity-Related Skin Events, with Dapagliflozin in Patients with Type 2 Diabetes Mellitus: A Pooled Analysis.

    PubMed

    Mellander, Annika; Billger, Martin; Johnsson, Eva; Träff, Anna Karin; Yoshida, Shigeru; Johnsson, Kristina

    2016-11-01

    In patients with type 2 diabetes mellitus (T2DM), dapagliflozin improves glycemic control and has a safety profile typically related to its mechanism of action. Hypersensitivity adverse events (AEs) have been reported in some patients with sodium-glucose cotransporter 2 (SGLT2) inhibitors, including a recent report of dermatological AEs in Japan. We investigated the frequency and characteristics of hypersensitivity AEs, including potentially hypersensitivity-related skin AEs, across 21 phase IIb/III trials of dapagliflozin (N = 5936) versus active or placebo comparators (N = 3403), including the subpopulation of Asian patients (N = 1563). Overall, AEs and serious AEs (SAEs) of hypersensitivity were infrequent and were reported in a similar proportion of patients with dapagliflozin versus active or placebo comparators (AEs: 4.5 vs. 4.3 %; SAEs: 0.2 vs. 0.1 %, respectively). The most common events affected the skin or subcutaneous tissue: rash (dapagliflozin: 1.1 %, comparator: 1.1 %), eczema (0.6, 0.8 %), dermatitis (0.5, 0.4 %), and urticaria (0.5, 0.2 %). Few patients discontinued as a result of hypersensitivity AEs (≤0.2 %). In patients of Asian descent, a lower frequency of hypersensitivity AEs was observed with dapagliflozin versus comparators (2.0 vs. 4.5 %). In the subset of placebo-controlled trials, hypersensitivity AEs were slightly more frequent with dapagliflozin than with placebo across the overall population (4.7 vs. 3.8 %), and less frequent with dapagliflozin in Asian patients (1.5 vs. 5.0 %). The findings of this post hoc analysis indicate that dapagliflozin does not lead to an increased risk of serious hypersensitivity reactions or potentially hypersensitivity-related skin events among patients with T2DM, including Asian patients. Long-term outcome studies and postmarketing surveillance will provide further information on hypersensitivity reactions with SGLT2 inhibitors. CLINICALTRIALS. NCT01042977, NCT01031680, NCT00855166, NCT

  7. SCADA data and the quantification of hazardous events for QMRA.

    PubMed

    Nilsson, P; Roser, D; Thorwaldsdotter, R; Petterson, S; Davies, C; Signor, R; Bergstedt, O; Ashbolt, N

    2007-01-01

    The objective of this study was to assess the use of on-line monitoring to support the QMRA at water treatment plants studied in the EU MicroRisk project. SCADA data were obtained from three Catchment-to-Tap Systems (CTS) along with system descriptions, diary records, grab sample data and deviation reports. Particular attention was paid to estimating hazardous event frequency, duration and magnitude. Using Shewart and CUSUM we identified 'change-points' corresponding to events of between 10 min and >1 month duration in timeseries data. Our analysis confirmed it is possible to quantify hazardous event durations from turbidity, chlorine residual and pH records and distinguish them from non-hazardous variability in the timeseries dataset. The durations of most 'events' were short-term (0.5-2.3 h). These data were combined with QMRA to estimate pathogen infection risk arising from such events as chlorination failure. While analysis of SCADA data alone could identify events provisionally, its interpretation was severely constrained in the absence of diary records and other system information. SCADA data analysis should only complement traditional water sampling, rather than replace it. More work on on-line data management, quality control and interpretation is needed before it can be used routinely for event characterization.

  8. Global patterns and impacts of El Niño events on coral reefs: A meta-analysis.

    PubMed

    Claar, Danielle C; Szostek, Lisa; McDevitt-Irwin, Jamie M; Schanze, Julian J; Baum, Julia K

    2018-01-01

    Impacts of global climate change on coral reefs are being amplified by pulse heat stress events, including El Niño, the warm phase of the El Niño Southern Oscillation (ENSO). Despite reports of extensive coral bleaching and up to 97% coral mortality induced by El Niño events, a quantitative synthesis of the nature, intensity, and drivers of El Niño and La Niña impacts on corals is lacking. Herein, we first present a global meta-analysis of studies quantifying the effects of El Niño/La Niña-warming on corals, surveying studies from both the primary literature and International Coral Reef Symposium (ICRS) Proceedings. Overall, the strongest signal for El Niño/La Niña-associated coral bleaching was long-term mean temperature; bleaching decreased with decreasing long-term mean temperature (n = 20 studies). Additionally, coral cover losses during El Niño/La Niña were shaped by localized maximum heat stress and long-term mean temperature (n = 28 studies). Second, we present a method for quantifying coral heat stress which, for any coral reef location in the world, allows extraction of remotely-sensed degree heating weeks (DHW) for any date (since 1982), quantification of the maximum DHW, and the time lag since the maximum DHW. Using this method, we show that the 2015/16 El Niño event instigated unprecedented global coral heat stress across the world's oceans. With El Niño events expected to increase in frequency and severity this century, it is imperative that we gain a clear understanding of how these thermal stress anomalies impact different coral species and coral reef regions. We therefore finish with recommendations for future coral bleaching studies that will foster improved syntheses, as well as predictive and adaptive capacity to extreme warming events.

  9. Global patterns and impacts of El Niño events on coral reefs: A meta-analysis

    PubMed Central

    Szostek, Lisa; McDevitt-Irwin, Jamie M.; Schanze, Julian J.; Baum, Julia K.

    2018-01-01

    Impacts of global climate change on coral reefs are being amplified by pulse heat stress events, including El Niño, the warm phase of the El Niño Southern Oscillation (ENSO). Despite reports of extensive coral bleaching and up to 97% coral mortality induced by El Niño events, a quantitative synthesis of the nature, intensity, and drivers of El Niño and La Niña impacts on corals is lacking. Herein, we first present a global meta-analysis of studies quantifying the effects of El Niño/La Niña-warming on corals, surveying studies from both the primary literature and International Coral Reef Symposium (ICRS) Proceedings. Overall, the strongest signal for El Niño/La Niña-associated coral bleaching was long-term mean temperature; bleaching decreased with decreasing long-term mean temperature (n = 20 studies). Additionally, coral cover losses during El Niño/La Niña were shaped by localized maximum heat stress and long-term mean temperature (n = 28 studies). Second, we present a method for quantifying coral heat stress which, for any coral reef location in the world, allows extraction of remotely-sensed degree heating weeks (DHW) for any date (since 1982), quantification of the maximum DHW, and the time lag since the maximum DHW. Using this method, we show that the 2015/16 El Niño event instigated unprecedented global coral heat stress across the world's oceans. With El Niño events expected to increase in frequency and severity this century, it is imperative that we gain a clear understanding of how these thermal stress anomalies impact different coral species and coral reef regions. We therefore finish with recommendations for future coral bleaching studies that will foster improved syntheses, as well as predictive and adaptive capacity to extreme warming events. PMID:29401493

  10. Risk Factors for Vascular Occlusive Events and Death Due to Bleeding in Trauma Patients; an Analysis of the CRASH-2 Cohort

    PubMed Central

    Pealing, Louise; Perel, Pablo; Prieto-Merino, David; Roberts, Ian

    2012-01-01

    Background Vascular occlusive events can complicate recovery following trauma. We examined risk factors for venous and arterial vascular occlusive events in trauma patients and the extent to which the risk of vascular occlusive events varies with the severity of bleeding. Methods and Findings We conducted a cohort analysis using data from a large international, double-blind, randomised, placebo-controlled trial (The CRASH-2 trial) [1]. We studied the association between patient demographic and physiological parameters at hospital admission and the risk of vascular occlusive events. To assess the extent to which risk of vascular occlusive events varies with severity of bleeding, we constructed a prognostic model for the risk of death due to bleeding and assessed the relationship between risk of death due to bleeding and risk of vascular occlusive events. There were 20,127 trauma patients with outcome data including 204 (1.01%) patients with a venous event (pulmonary embolism or deep vein thrombosis) and 200 (0.99%) with an arterial event (myocardial infarction or stroke). There were 81 deaths due to vascular occlusive events. Increasing age, decreasing systolic blood pressure, increased respiratory rates, longer central capillary refill times, higher heart rates and lower Glasgow Coma Scores (all p<0.02) were strong risk factors for venous and arterial vascular occlusive events. Patients with more severe bleeding as assessed by predicted risk of haemorrhage death had a greatly increased risk for all types of vascular occlusive event (all p<0.001). Conclusions Patients with severe traumatic bleeding are at greatly increased risk of venous and arterial vascular occlusive events. Older age and blunt trauma are also risk factors for vascular occlusive events. Effective treatment of bleeding may reduce venous and arterial vascular occlusive complications in trauma patients. PMID:23251374

  11. Interval estimation of the overall treatment effect in a meta-analysis of a few small studies with zero events.

    PubMed

    Pateras, Konstantinos; Nikolakopoulos, Stavros; Mavridis, Dimitris; Roes, Kit C B

    2018-03-01

    When a meta-analysis consists of a few small trials that report zero events, accounting for heterogeneity in the (interval) estimation of the overall effect is challenging. Typically, we predefine meta-analytical methods to be employed. In practice, data poses restrictions that lead to deviations from the pre-planned analysis, such as the presence of zero events in at least one study arm. We aim to explore heterogeneity estimators behaviour in estimating the overall effect across different levels of sparsity of events. We performed a simulation study that consists of two evaluations. We considered an overall comparison of estimators unconditional on the number of observed zero cells and an additional one by conditioning on the number of observed zero cells. Estimators that performed modestly robust when (interval) estimating the overall treatment effect across a range of heterogeneity assumptions were the Sidik-Jonkman, Hartung-Makambi and improved Paul-Mandel. The relative performance of estimators did not materially differ between making a predefined or data-driven choice. Our investigations confirmed that heterogeneity in such settings cannot be estimated reliably. Estimators whose performance depends strongly on the presence of heterogeneity should be avoided. The choice of estimator does not need to depend on whether or not zero cells are observed.

  12. Flares, ejections, proton events

    NASA Astrophysics Data System (ADS)

    Belov, A. V.

    2017-11-01

    Statistical analysis is performed for the relationship of coronal mass ejections (CMEs) and X-ray flares with the fluxes of solar protons with energies >10 and >100 MeV observed near the Earth. The basis for this analysis was the events that took place in 1976-2015, for which there are reliable observations of X-ray flares on GOES satellites and CME observations with SOHO/LASCO coronagraphs. A fairly good correlation has been revealed between the magnitude of proton enhancements and the power and duration of flares, as well as the initial CME speed. The statistics do not give a clear advantage either to CMEs or the flares concerning their relation with proton events, but the characteristics of the flares and ejections complement each other well and are reasonable to use together in the forecast models. Numerical dependences are obtained that allow estimation of the proton fluxes to the Earth expected from solar observations; possibilities for improving the model are discussed.

  13. Analysis and reconstructed modelling of the debris flow event of the 21st of July 2012 of St. Lorenzen (Styria, Austira)

    NASA Astrophysics Data System (ADS)

    Janu, Stefan; Mehlhorn, Susanne; Moser, Markus

    2013-04-01

    Analysis and reconstructed modelling of the debris flow event of the 21st of July 2012 of St. Lorenzen (Styria, Austria) Authors: Stefan Janu, Susanne Mehlhorn, Markus Moser The village of St. Lorenzen, in the Styrian Palten valley is situated on the banks of the Lorenz torrent, in which a debris flow event occurred in the early morning hours of the 21st of July 2012, causing catastrophic damage to residential buildings and other infrastructural facilities. In the ministry-approved hazard zone map of 2009, the flood water discharge and bedload volume associated with a 150-year event was estimated at 34 m³/s and 25,000 m³ respectively for the 5.84 km² catchment area. The bedload transport capacity of the torrent was classified as ranging from 'heavy' to 'capable of producing debris flows'. The dominant process type of the mass movement event may be described as a fine-grained debris flow. The damage in the residential area of St.Lorenzen was caused by a debris flow pulse in the lower reach of the Lorenz torrent. This debris flow pulse was in turn caused by numerous landslides along the middle reaches of the torrent, some of which caused blockages, ultimately leading to an outburst event in the main torrent. Discharge cross-sections ranging from 65 - 90 m², and over 100 m² in a few instances, were measured upstream of the St. Lorenzen residential area. Back-calculations of velocities yielded an average debris flow velocity along the middle reaches of the torrent between 11 and 16 m/s. An average velocity of 9 m/s was calculated for the debris flow at the neck of the alluvial fan directly behind the center of the village. Due to both the high discharge values as well as to the height of the mass movement deposits, the natural hazard event of 21 July 2012 in St. Lorenzen is clearly to be described as having had an extreme intensity. A total of 67 buildings were damaged along the Lorenz torrent, 7 of were completely destroyed. According to the Austrian Service for

  14. Tracking the evolution of stream DOM source during storm events using end member mixing analysis based on DOM quality

    NASA Astrophysics Data System (ADS)

    Yang, Liyang; Chang, Soon-Woong; Shin, Hyun-Sang; Hur, Jin

    2015-04-01

    The source of river dissolved organic matter (DOM) during storm events has not been well constrained, which is critical in determining the quality and reactivity of DOM. This study assessed temporal changes in the contributions of four end members (weeds, leaf litter, soil, and groundwater), which exist in a small forested watershed (the Ehwa Brook, South Korea), to the stream DOM during two storm events, using end member mixing analysis (EMMA) based on spectroscopic properties of DOM. The instantaneous export fluxes of dissolved organic carbon (DOC), chromophoric DOM (CDOM), and fluorescent components were all enhanced during peak flows. The DOC concentration increased with the flow rate, while CDOM and humic-like fluorescent components were diluted around the peak flows. Leaf litter was dominant for the DOM source in event 2 with a higher rainfall, although there were temporal variations in the contributions of the four end members to the stream DOM for both events. The contribution of leaf litter peaked while that of deeper soils decreased to minima at peak flows. Our results demonstrated that EMMA based on DOM properties could be used to trace the DOM source, which is of fundamental importance for understanding the factors responsible for river DOM dynamics during storm events.

  15. Cardiovascular Events in Systemic Lupus Erythematosus

    PubMed Central

    Fernández-Nebro, Antonio; Rúa-Figueroa, Íñigo; López-Longo, Francisco J.; Galindo-Izquierdo, María; Calvo-Alén, Jaime; Olivé-Marqués, Alejandro; Ordóñez-Cañizares, Carmen; Martín-Martínez, María A.; Blanco, Ricardo; Melero-González, Rafael; Ibáñez-Rúan, Jesús; Bernal-Vidal, José Antonio; Tomero-Muriel, Eva; Uriarte-Isacelaya, Esther; Horcada-Rubio, Loreto; Freire-González, Mercedes; Narváez, Javier; Boteanu, Alina L.; Santos-Soler, Gregorio; Andreu, José L.; Pego-Reigosa, José M.

    2015-01-01

    Abstract This article estimates the frequency of cardiovascular (CV) events that occurred after diagnosis in a large Spanish cohort of patients with systemic lupus erythematosus (SLE) and investigates the main risk factors for atherosclerosis. RELESSER is a nationwide multicenter, hospital-based registry of SLE patients. This is a cross-sectional study. Demographic and clinical variables, the presence of traditional risk factors, and CV events were collected. A CV event was defined as a myocardial infarction, angina, stroke, and/or peripheral artery disease. Multiple logistic regression analysis was performed to investigate the possible risk factors for atherosclerosis. From 2011 to 2012, 3658 SLE patients were enrolled. Of these, 374 (10.9%) patients suffered at least a CV event. In 269 (7.4%) patients, the CV events occurred after SLE diagnosis (86.2% women, median [interquartile range] age 54.9 years [43.2–66.1], and SLE duration of 212.0 months [120.8–289.0]). Strokes (5.7%) were the most frequent CV event, followed by ischemic heart disease (3.8%) and peripheral artery disease (2.2%). Multivariate analysis identified age (odds ratio [95% confidence interval], 1.03 [1.02–1.04]), hypertension (1.71 [1.20–2.44]), smoking (1.48 [1.06–2.07]), diabetes (2.2 [1.32–3.74]), dyslipidemia (2.18 [1.54–3.09]), neurolupus (2.42 [1.56–3.75]), valvulopathy (2.44 [1.34–4.26]), serositis (1.54 [1.09–2.18]), antiphospholipid antibodies (1.57 [1.13–2.17]), low complement (1.81 [1.12–2.93]), and azathioprine (1.47 [1.04–2.07]) as risk factors for CV events. We have confirmed that SLE patients suffer a high prevalence of premature CV disease. Both traditional and nontraditional risk factors contribute to this higher prevalence. Although it needs to be verified with future studies, our study also shows—for the first time—an association between diabetes and CV events in SLE patients. PMID:26200625

  16. What can we learn from the deadly flash floods? Post Event Review Capability (PERC) analysis of the Bavaria and Baden-Wurttemberg flood events in Summer 2016

    NASA Astrophysics Data System (ADS)

    Szoenyi, Michael

    2017-04-01

    In May/June 2016, stationary low pressure systems brought intense rainfall with record-braking intensities of well above 100 mm rain in few hours locally in the southern states of Baden-Wurttemberg and Bavaria, Germany. In steep terrains, small channels and creeks became devastating torrents impacting, among others, the villages of Simbach/Inn, Schwäbisch-Gmünd and Braunsbach. Just few days prior, France had also seen devastating rainfall and flooding. Damage in Germany alone is estimated at 2.8 M USD, of which less than 50% are insured. The loss of life was significant, with 18 fatalities reported across the events. This new forensic event analysis as part of Zurich's Post Event Review Capability (PERC) investigates the flash flood events following these record rainfalls in Southern Germany and tries to answer the following questions holistically, across the five capitals (5C) and the full disaster risk management (DRM) cycle, which are key to understanding how to become more resilient to such flood events: - Why have these intense rainfall events led to such devastating consequences? The EU Floods directive and its implementation in the various member states, as well as the 2002 and 2013 Germany floods, have focused on larger rivers and the main asset concentration. The pathway and mechanism of the 2016 floods are very different and need to be better understood. Flash floods and surface flooding may need to become the new focus and be much better communicated to people at risk, as the awareness for such perils has been identified as low. - How can the prevalence for such flash floods be better identified and mapped? Research indicated that affected people and decision makers alike attribute the occurrence of such flash floods as arbitrary, but we argue that hotspots can and must be identified based on an overlay of rainfall intensity maps, topography leading to flash flood processes, and vulnerable assets. In Germany, there are currently no comprehensive hazard

  17. Analysis of geohazards events along Swiss roads from autumn 2011 to present

    NASA Astrophysics Data System (ADS)

    Voumard, Jérémie; Jaboyedoff, Michel; Derron, Marc-Henri

    2014-05-01

    In Switzerland, roads and railways are threatened throughout the year by several natural hazards. Some of these events reach transport infrastructure many time per year leading to the closing of transportation corridors, loss of access, deviation travels and sometimes infrastructures damages and loss of human lives (3 fatalities during the period considered). The aim of this inventory of events is to investigate the number of natural events affecting roads and railways in Switzerland since autumn 2011 until now. Natural hazards affecting roads and railway can be classified in five categories: rockfalls, landslides, debris flows, snow avalanches and floods. They potentially cause several important direct damages on transportation infrastructure (roads, railway), vehicles (slightly or very damaged) or human life (slightly or seriously injured person, death). These direct damages can be easily evaluated from press articles or from Swiss police press releases. Indirect damages such as deviation cost are not taken into account in this work. During the two a half last years, about 50 events affecting the Swiss roads and Swiss railways infrastructures were inventoried. The proportion of events due to rockfalls is 45%, to landslides 25%, to debris flows 15%, to snow avalanches 10% and to floods 5%. During this period, three fatalities and two persons were injured while 23 vehicles (car, trains and coach) and 24 roads and railways were damaged. We can see that floods occur mainly on the Swiss Plateau whereas rockfalls, debris flow, snow avalanches and landslides are mostly located in the Alpine area. Most of events occur on secondary mountain roads and railways. The events are well distributed on the whole Alpine area except for the Gotthard hotspot, where an important European North-South motorway (hit in 2003 with two fatalities) and railway (hit three times in 2012 with one fatalities) are more frequently affected. According to the observed events in border regions of

  18. Predicting ground contact events for a continuum of gait types: An application of targeted machine learning using principal component analysis.

    PubMed

    Osis, Sean T; Hettinga, Blayne A; Ferber, Reed

    2016-05-01

    An ongoing challenge in the application of gait analysis to clinical settings is the standardized detection of temporal events, with unobtrusive and cost-effective equipment, for a wide range of gait types. The purpose of the current study was to investigate a targeted machine learning approach for the prediction of timing for foot strike (or initial contact) and toe-off, using only kinematics for walking, forefoot running, and heel-toe running. Data were categorized by gait type and split into a training set (∼30%) and a validation set (∼70%). A principal component analysis was performed, and separate linear models were trained and validated for foot strike and toe-off, using ground reaction force data as a gold-standard for event timing. Results indicate the model predicted both foot strike and toe-off timing to within 20ms of the gold-standard for more than 95% of cases in walking and running gaits. The machine learning approach continues to provide robust timing predictions for clinical use, and may offer a flexible methodology to handle new events and gait types. Copyright © 2016 Elsevier B.V. All rights reserved.

  19. Predicting the occurrence of embolic events: an analysis of 1456 episodes of infective endocarditis from the Italian Study on Endocarditis (SEI).

    PubMed

    Rizzi, Marco; Ravasio, Veronica; Carobbio, Alessandra; Mattucci, Irene; Crapis, Massimo; Stellini, Roberto; Pasticci, Maria Bruna; Chinello, Pierangelo; Falcone, Marco; Grossi, Paolo; Barbaro, Francesco; Pan, Angelo; Viale, Pierluigi; Durante-Mangoni, Emanuele

    2014-04-29

    Embolic events are a major cause of morbidity and mortality in patients with infective endocarditis. We analyzed the database of the prospective cohort study SEI in order to identify factors associated with the occurrence of embolic events and to develop a scoring system for the assessment of the risk of embolism. We retrospectively analyzed 1456 episodes of infective endocarditis from the multicenter study SEI. Predictors of embolism were identified. Risk factors identified at multivariate analysis as predictive of embolism in left-sided endocarditis, were used for the development of a risk score: 1 point was assigned to each risk factor (total risk score range: minimum 0 points; maximum 2 points). Three categories were defined by the score: low (0 points), intermediate (1 point), or high risk (2 points); the probability of embolic events per risk category was calculated for each day on treatment (day 0 through day 30). There were 499 episodes of infective endocarditis (34%) that were complicated by ≥ 1 embolic event. Most embolic events occurred early in the clinical course (first week of therapy: 15.5 episodes per 1000 patient days; second week: 3.7 episodes per 1000 patient days). In the total cohort, the factors associated with the occurrence of embolism at multivariate analysis were prosthetic valve localization (odds ratio, 1.84), right-sided endocarditis (odds ratio, 3.93), Staphylococcus aureus etiology (odds ratio, 2.23) and vegetation size ≥ 13 mm (odds ratio, 1.86). In left-sided endocarditis, Staphylococcus aureus etiology (odds ratio, 2.1) and vegetation size ≥ 13 mm (odds ratio, 2.1) were independently associated with embolic events; the 30-day cumulative incidence of embolism varied with risk score category (low risk, 12%; intermediate risk, 25%; high risk, 38%; p < 0.001). Staphylococcus aureus etiology and vegetation size are associated with an increased risk of embolism. In left-sided endocarditis, a simple scoring system

  20. Predicting the occurrence of embolic events: an analysis of 1456 episodes of infective endocarditis from the Italian Study on Endocarditis (SEI)

    PubMed Central

    2014-01-01

    Background Embolic events are a major cause of morbidity and mortality in patients with infective endocarditis. We analyzed the database of the prospective cohort study SEI in order to identify factors associated with the occurrence of embolic events and to develop a scoring system for the assessment of the risk of embolism. Methods We retrospectively analyzed 1456 episodes of infective endocarditis from the multicenter study SEI. Predictors of embolism were identified. Risk factors identified at multivariate analysis as predictive of embolism in left-sided endocarditis, were used for the development of a risk score: 1 point was assigned to each risk factor (total risk score range: minimum 0 points; maximum 2 points). Three categories were defined by the score: low (0 points), intermediate (1 point), or high risk (2 points); the probability of embolic events per risk category was calculated for each day on treatment (day 0 through day 30). Results There were 499 episodes of infective endocarditis (34%) that were complicated by ≥ 1 embolic event. Most embolic events occurred early in the clinical course (first week of therapy: 15.5 episodes per 1000 patient days; second week: 3.7 episodes per 1000 patient days). In the total cohort, the factors associated with the occurrence of embolism at multivariate analysis were prosthetic valve localization (odds ratio, 1.84), right-sided endocarditis (odds ratio, 3.93), Staphylococcus aureus etiology (odds ratio, 2.23) and vegetation size ≥ 13 mm (odds ratio, 1.86). In left-sided endocarditis, Staphylococcus aureus etiology (odds ratio, 2.1) and vegetation size ≥ 13 mm (odds ratio, 2.1) were independently associated with embolic events; the 30-day cumulative incidence of embolism varied with risk score category (low risk, 12%; intermediate risk, 25%; high risk, 38%; p < 0.001). Conclusions Staphylococcus aureus etiology and vegetation size are associated with an increased risk of embolism. In left

  1. Potential of Breastmilk Analysis to Inform Early Events in Breast Carcinogenesis: Rationale and Considerations

    PubMed Central

    Murphy, Jeanne; Sherman, Mark E.; Browne, Eva P.; Caballero, Ana I.; Punska, Elizabeth C.; Pfeiffer, Ruth M.; Yang, Hannah P.; Lee, Maxwell; Yang, Howard; Gierach, Gretchen L.; Arcaro, Kathleen F.

    2016-01-01

    This review summarizes methods related to the study of human breastmilk in etiologic and biomarkers research. Despite the importance of reproductive factors in breast carcinogenesis, factors that act early in life are difficult to study because young women rarely require breast imaging or biopsy, and analysis of critical circulating factors (e.g. hormones) is often complicated by the requirement to accurately account for menstrual cycle date. Accordingly, novel approaches are needed to understand how events such as pregnancy, breastfeeding, weaning, and post-weaning breast remodeling influence breast cancer risk. Analysis of breastmilk offers opportunities to understand mechanisms related to carcinogenesis in the breast, and to identify risk markers that may inform efforts to identify high-risk women early in the carcinogenic process. In addition, analysis of breastmilk could have value in early detection or diagnosis of breast cancer. In this article we describe the potential for using breastmilk to characterize the microenvironment of the lactating breast with the goal of advancing research on risk assessment, prevention, and detection of breast cancer. PMID:27107568

  2. Chronodes: Interactive Multifocus Exploration of Event Sequences

    PubMed Central

    POLACK, PETER J.; CHEN, SHANG-TSE; KAHNG, MINSUK; DE BARBARO, KAYA; BASOLE, RAHUL; SHARMIN, MOUSHUMI; CHAU, DUEN HORNG

    2018-01-01

    The advent of mobile health (mHealth) technologies challenges the capabilities of current visualizations, interactive tools, and algorithms. We present Chronodes, an interactive system that unifies data mining and human-centric visualization techniques to support explorative analysis of longitudinal mHealth data. Chronodes extracts and visualizes frequent event sequences that reveal chronological patterns across multiple participant timelines of mHealth data. It then combines novel interaction and visualization techniques to enable multifocus event sequence analysis, which allows health researchers to interactively define, explore, and compare groups of participant behaviors using event sequence combinations. Through summarizing insights gained from a pilot study with 20 behavioral and biomedical health experts, we discuss Chronodes’s efficacy and potential impact in the mHealth domain. Ultimately, we outline important open challenges in mHealth, and offer recommendations and design guidelines for future research. PMID:29515937

  3. Meta-analysis of transcatheter closure versus medical therapy for patent foramen ovale in prevention of recurrent neurological events after presumed paradoxical embolism.

    PubMed

    Agarwal, Shikhar; Bajaj, Navkaranbir Singh; Kumbhani, Dharam J; Tuzcu, E Murat; Kapadia, Samir R

    2012-07-01

    In this study, a meta-analysis of observational studies was performed to compare the rate of recurrent neurological events (RNE) between transcatheter closure and medical management of patients with cryptogenic stroke/transient ischemic attack (TIA) and concomitant patent foramen ovale (PFO). A significant controversy surrounds the optimal strategy for treatment of cryptogenic stroke/TIA and coexistent PFO. We conducted a MEDLINE search with standard search terms to determine eligible studies. Adjusted incidence rates of RNE were 0.8 (95% confidence interval [CI]: 0.5 to 1.1) events and 5.0 (95% CI: 3.6 to 6.9) events/100 person-years (PY) in the transcatheter closure and medical management arms, respectively. Meta-analysis of the limited number of comparative studies and meta-regression analysis suggested that the transcatheter closure might be superior to the medical therapy in prevention of RNE after cryptogenic stroke. Comparison of the anticoagulation and antiplatelet therapy subgroups of the medical arm yielded a significantly lower risk of RNE within patients treated with anticoagulants. Device-related complications were encountered at the rate of 4.1 (95% CI: 3.2 to 5.0) events/100 PY, with atrial arrhythmias being the most frequent complication. After transcatheter closure, RNE did not seem to be related to the pre-treatment shunt size or the presence of residual shunting in the follow-up period. Significant benefit of transcatheter PFO closure was apparent in elderly patients, patients with concomitant atrial septal aneurysm, and patients with thrombophilia. Rates of RNE with transcatheter closure and medical therapy in patients presenting with cryptogenic stroke or TIA were estimated at 0.8 and 5.0 events/100 PY. Further randomized controlled trials are needed to conclusively compare these 2 management strategies. Copyright © 2012 American College of Cardiology Foundation. Published by Elsevier Inc. All rights reserved.

  4. Conversion events in gene clusters

    PubMed Central

    2011-01-01

    Background Gene clusters containing multiple similar genomic regions in close proximity are of great interest for biomedical studies because of their associations with inherited diseases. However, such regions are difficult to analyze due to their structural complexity and their complicated evolutionary histories, reflecting a variety of large-scale mutational events. In particular, conversion events can mislead inferences about the relationships among these regions, as traced by traditional methods such as construction of phylogenetic trees or multi-species alignments. Results To correct the distorted information generated by such methods, we have developed an automated pipeline called CHAP (Cluster History Analysis Package) for detecting conversion events. We used this pipeline to analyze the conversion events that affected two well-studied gene clusters (α-globin and β-globin) and three gene clusters for which comparative sequence data were generated from seven primate species: CCL (chemokine ligand), IFN (interferon), and CYP2abf (part of cytochrome P450 family 2). CHAP is freely available at http://www.bx.psu.edu/miller_lab. Conclusions These studies reveal the value of characterizing conversion events in the context of studying gene clusters in complex genomes. PMID:21798034

  5. Towards cross-lingual alerting for bursty epidemic events.

    PubMed

    Collier, Nigel

    2011-10-06

    Online news reports are increasingly becoming a source for event-based early warning systems that detect natural disasters. Harnessing the massive volume of information available from multilingual newswire presents as many challanges as opportunities due to the patterns of reporting complex spatio-temporal events. In this article we study the problem of utilising correlated event reports across languages. We track the evolution of 16 disease outbreaks using 5 temporal aberration detection algorithms on text-mined events classified according to disease and outbreak country. Using ProMED reports as a silver standard, comparative analysis of news data for 13 languages over a 129 day trial period showed improved sensitivity, F1 and timeliness across most models using cross-lingual events. We report a detailed case study analysis for Cholera in Angola 2010 which highlights the challenges faced in correlating news events with the silver standard. The results show that automated health surveillance using multilingual text mining has the potential to turn low value news into high value alerts if informed choices are used to govern the selection of models and data sources. An implementation of the C2 alerting algorithm using multilingual news is available at the BioCaster portal http://born.nii.ac.jp/?page=globalroundup.

  6. Metamizole-Associated Adverse Events: A Systematic Review and Meta-Analysis

    PubMed Central

    Fässler, Margrit; Blozik, Eva; Linde, Klaus; Jüni, Peter; Reichenbach, Stephan; Scherer, Martin

    2015-01-01

    Background Metamizole is used to treat pain in many parts of the world. Information on the safety profile of metamizole is scarce; no conclusive summary of the literature exists. Objective To determine whether metamizole is clinically safe compared to placebo and other analgesics. Methods We searched CENTRAL, MEDLINE, EMBASE, CINAHL, and several clinical trial registries. We screened the reference lists of included trials and previous systematic reviews. We included randomized controlled trials that compared the effects of metamizole, administered to adults in any form and for any indication, to other analgesics or to placebo. Two authors extracted data regarding trial design and size, indications for pain medication, patient characteristics, treatment regimens, and methodological characteristics. Adverse events (AEs), serious adverse events (SAEs), and dropouts were assessed. We conducted separate meta-analyses for each metamizole comparator, using standard inverse-variance random effects meta-analysis to pool the estimates across trials, reported as risk ratios (RRs). We calculated the DerSimonian and Laird variance estimate T2 to measure heterogeneity between trials. The pre-specified primary end point was any AE during the trial period. Results Of the 696 potentially eligible trials, 79 trials including almost 4000 patients with short-term metamizole use of less than two weeks met our inclusion criteria. Fewer AEs were reported for metamizole compared to opioids, RR = 0.79 (confidence interval 0.79 to 0.96). We found no differences between metamizole and placebo, paracetamol and NSAIDs. Only a few SAEs were reported, with no difference between metamizole and other analgesics. No agranulocytosis or deaths were reported. Our results were limited by the mediocre overall quality of the reports. Conclusion For short-term use in the hospital setting, metamizole seems to be a safe choice when compared to other widely used analgesics. High-quality, adequately sized

  7. Analysis of respiratory events in obstructive sleep apnea syndrome: Inter-relations and association to simple nocturnal features.

    PubMed

    Ghandeharioun, H; Rezaeitalab, F; Lotfi, R

    2016-01-01

    This study carefully evaluates the association of different respiration-related events to each other and to simple nocturnal features in obstructive sleep apnea-hypopnea syndrome (OSAS). The events include apneas, hypopneas, respiratory event-related arousals and snores. We conducted a statistical study on 158 adults who underwent polysomnography between July 2012 and May 2014. To monitor relevance, along with linear statistical strategies like analysis of variance and bootstrapping a correlation coefficient standard error, the non-linear method of mutual information is also applied to illuminate vague results of linear techniques. Based on normalized mutual information weights (NMIW), indices of apnea are 1.3 times more relevant to AHI values than those of hypopnea. NMIW for the number of blood oxygen desaturation below 95% is considerable (0.531). The next relevant feature is "respiratory arousals index" with NMIW of 0.501. Snore indices (0.314), and BMI (0.203) take the next place. Based on NMIW values, snoring events are nearly one-third (29.9%) more dependent to hypopneas than RERAs. 1. The more sever the OSAS is, the more frequently the apneic events happen. 2. The association of snore with hypopnea/RERA revealed which is routinely ignored in regression-based OSAS modeling. 3. The statistical dependencies of oximetry features potentially can lead to home-based screening of OSAS. 4. Poor ESS-AHI relevance in the database under study indicates its disability for the OSA diagnosis compared to oximetry. 5. Based on poor RERA-snore/ESS relevance, detailed history of the symptoms plus polysomnography is suggested for accurate diagnosis of RERAs. Copyright © 2015 Sociedade Portuguesa de Pneumologia. Published by Elsevier España, S.L.U. All rights reserved.

  8. Perioperative outcomes and adverse events of minimally invasive versus open posterior lumbar fusion: meta-analysis and systematic review.

    PubMed

    Goldstein, Christina L; Macwan, Kevin; Sundararajan, Kala; Rampersaud, Y Raja

    2016-03-01

    The objective of this study was to determine the clinical comparative effectiveness and adverse event rates of posterior minimally invasive surgery (MIS) compared with open transforaminal or posterior lumbar interbody fusion (TLIF/PLIF). A systematic review of the Medline, EMBASE, PubMed, Web of Science, and Cochrane databases was performed. A hand search of reference lists was conducted. Studies were reviewed by 2 independent assessors to identify randomized controlled trials (RCTs) or comparative cohort studies including at least 10 patients undergoing MIS or open TLIF/PLIF for degenerative lumbar spinal disorders and reporting at least 1 of the following: clinical outcome measure, perioperative clinical or process measure, radiographic outcome, or adverse events. Study quality was assessed using the Grades of Recommendation, Assessment, Development, and Evaluation (GRADE) protocol. When appropriate, a meta-analysis of outcomes data was conducted. The systematic review and reference list search identified 3301 articles, with 26 meeting study inclusion criteria. All studies, including 1 RCT, were of low or very low quality. No significant difference regarding age, sex, surgical levels, or diagnosis was identified between the 2 cohorts (856 patients in the MIS cohort, 806 patients in the open cohort). The meta-analysis revealed changes in the perioperative outcomes of mean estimated blood loss, time to ambulation, and length of stay favoring an MIS approach by 260 ml (p < 0.00001), 3.5 days (p = 0.0006), and 2.9 days (p < 0.00001), respectively. Operative time was not significantly different between the surgical techniques (p = 0.78). There was no significant difference in surgical adverse events (p = 0.97), but MIS cases were significantly less likely to experience medical adverse events (risk ratio [MIS vs open] = 0.39, 95% confidence interval 0.23-0.69, p = 0.001). No difference in nonunion (p = 0.97) or reoperation rates (p = 0.97) was observed. Mean Oswestry

  9. Modification of the SAS4A Safety Analysis Code for Integration with the ADAPT Discrete Dynamic Event Tree Framework.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jankovsky, Zachary Kyle; Denman, Matthew R.

    It is difficult to assess the consequences of a transient in a sodium-cooled fast reactor (SFR) using traditional probabilistic risk assessment (PRA) methods, as numerous safety-related sys- tems have passive characteristics. Often there is significant dependence on the value of con- tinuous stochastic parameters rather than binary success/failure determinations. One form of dynamic PRA uses a system simulator to represent the progression of a transient, tracking events through time in a discrete dynamic event tree (DDET). In order to function in a DDET environment, a simulator must have characteristics that make it amenable to changing physical parameters midway through themore » analysis. The SAS4A SFR system analysis code did not have these characteristics as received. This report describes the code modifications made to allow dynamic operation as well as the linking to a Sandia DDET driver code. A test case is briefly described to demonstrate the utility of the changes.« less

  10. Single event time series analysis in a binary karst catchment evaluated using a groundwater model (Lurbach system, Austria).

    PubMed

    Mayaud, C; Wagner, T; Benischke, R; Birk, S

    2014-04-16

    The Lurbach karst system (Styria, Austria) is drained by two major springs and replenished by both autogenic recharge from the karst massif itself and a sinking stream that originates in low permeable schists (allogenic recharge). Detailed data from two events recorded during a tracer experiment in 2008 demonstrate that an overflow from one of the sub-catchments to the other is activated if the discharge of the main spring exceeds a certain threshold. Time series analysis (autocorrelation and cross-correlation) was applied to examine to what extent the various available methods support the identification of the transient inter-catchment flow observed in this binary karst system. As inter-catchment flow is found to be intermittent, the evaluation was focused on single events. In order to support the interpretation of the results from the time series analysis a simplified groundwater flow model was built using MODFLOW. The groundwater model is based on the current conceptual understanding of the karst system and represents a synthetic karst aquifer for which the same methods were applied. Using the wetting capability package of MODFLOW, the model simulated an overflow similar to what has been observed during the tracer experiment. Various intensities of allogenic recharge were employed to generate synthetic discharge data for the time series analysis. In addition, geometric and hydraulic properties of the karst system were varied in several model scenarios. This approach helps to identify effects of allogenic recharge and aquifer properties in the results from the time series analysis. Comparing the results from the time series analysis of the observed data with those of the synthetic data a good agreement was found. For instance, the cross-correlograms show similar patterns with respect to time lags and maximum cross-correlation coefficients if appropriate hydraulic parameters are assigned to the groundwater model. The comparable behaviors of the real and the

  11. Single event time series analysis in a binary karst catchment evaluated using a groundwater model (Lurbach system, Austria)

    PubMed Central

    Mayaud, C.; Wagner, T.; Benischke, R.; Birk, S.

    2014-01-01

    Summary The Lurbach karst system (Styria, Austria) is drained by two major springs and replenished by both autogenic recharge from the karst massif itself and a sinking stream that originates in low permeable schists (allogenic recharge). Detailed data from two events recorded during a tracer experiment in 2008 demonstrate that an overflow from one of the sub-catchments to the other is activated if the discharge of the main spring exceeds a certain threshold. Time series analysis (autocorrelation and cross-correlation) was applied to examine to what extent the various available methods support the identification of the transient inter-catchment flow observed in this binary karst system. As inter-catchment flow is found to be intermittent, the evaluation was focused on single events. In order to support the interpretation of the results from the time series analysis a simplified groundwater flow model was built using MODFLOW. The groundwater model is based on the current conceptual understanding of the karst system and represents a synthetic karst aquifer for which the same methods were applied. Using the wetting capability package of MODFLOW, the model simulated an overflow similar to what has been observed during the tracer experiment. Various intensities of allogenic recharge were employed to generate synthetic discharge data for the time series analysis. In addition, geometric and hydraulic properties of the karst system were varied in several model scenarios. This approach helps to identify effects of allogenic recharge and aquifer properties in the results from the time series analysis. Comparing the results from the time series analysis of the observed data with those of the synthetic data a good agreement was found. For instance, the cross-correlograms show similar patterns with respect to time lags and maximum cross-correlation coefficients if appropriate hydraulic parameters are assigned to the groundwater model. The comparable behaviors of the real and

  12. Assessment of Adverse Events in Protocols, Clinical Study Reports, and Published Papers of Trials of Orlistat: A Document Analysis.

    PubMed

    Schroll, Jeppe Bennekou; Penninga, Elisabeth I; Gøtzsche, Peter C

    2016-08-01

    stated that "all adverse events were recorded." For one trial, we identified an additional 1,318 adverse events that were not listed or mentioned in the CSR itself but could be identified through manually counting individual adverse events reported in an appendix. We discovered that the majority of patients had multiple episodes of the same adverse event that were only counted once, though this was not described in the CSRs. We also discovered that participants treated with orlistat experienced twice as many days with adverse events as participants treated with placebo (22.7 d versus 14.9 d, p-value < 0.0001, Student's t test). Furthermore, compared with the placebo group, adverse events in the orlistat group were more severe. None of this was stated in the CSR or in the published paper. Our analysis was restricted to one drug tested in the mid-1990s; our results might therefore not be applicable for newer drugs. In the orlistat trials, we identified important disparities in the reporting of adverse events between protocols, clinical study reports, and published papers. Reports of these trials seemed to have systematically understated adverse events. Based on these findings, systematic reviews of drugs might be improved by including protocols and CSRs in addition to published articles.

  13. Life events and Tourette syndrome.

    PubMed

    Steinberg, Tamar; Shmuel-Baruch, Sharona; Horesh, Netta; Apter, Alan

    2013-07-01

    Tourette syndrome (TS) is a neuropsychiatric developmental disorder characterized by the presence of multiple motor tics and one or more vocal tics. Although TS is primarily biological in origin, stress-diatheses interactions most probably play a role in the course of the illness. The precise influence of the environment on this basically biological disorder is difficult to ascertain, particularly when TS is complicated by comorbidities. Among the many questions that remain unresolved are the differential impact of positive and negative events and specific subtypes of events, and the importance of major crucial events relative to minor daily ones to tic severity. To examine the relationships between life events, tic severity and comorbid disorders in Tourette Syndrome (TS), including OCD, ADHD, anxiety, depression and rage attacks. Life events were classified by quantity, quality (positive or negative) and classification types of events (family, friends etc.). Sixty patients aged 7-17 years with Tourette syndrome or a chronic tic disorder were recruited from Psychological Medicine Clinic in Schneider Children's Medical Center of Israel. Yale Global Tic Severity Scale; Children's Yale Brown Obsessive Compulsive Scale; Life Experiences Survey; Brief Adolescent Life Events Scale; Screen for Child Anxiety Related Emotional Disorders; Child Depression Inventory/Beck Depression Inventory; ADHD Rating Scale IV; Overt Aggression Scale. Regarding tics and minor life events, there was a weak but significant correlation between severity of motor tics and the quantity of negative events. No significant correlation was found between tic severity and quantity of positive events. Analysis of the BALES categories yielded a significant direct correlation between severity of vocal tics and quantity of negative events involving friends. Regarding comorbidities and minor life events, highly significant correlations were found with depression and anxiety. Regarding tics and major life

  14. Paraesthesia after local anaesthetics: an analysis of reports to the FDA Adverse Event Reporting System.

    PubMed

    Piccinni, Carlo; Gissi, Davide B; Gabusi, Andrea; Montebugnoli, Lucio; Poluzzi, Elisabetta

    2015-07-01

    This study was aimed to evaluate the possible alert signals of paraesthesia by local anaesthetics, focusing on those used in dentistry. A case/non-case study of spontaneous adverse events recorded in FAERS (FDA Adverse Event Reporting System) between 2004 and 2011 was performed. Cases were represented by the reports of reactions grouped under the term 'Paraesthesias and dysaesthesias' involving local anaesthetics (ATC: N01B*); non-cases were all other reports of the same drugs. Reporting odds ratios (ROR) with the relevant 95% confidence intervals (95CI) were calculated. Alert signal was considered when number of cases >3 and lower limit of ROR 95CI > 1. To estimate the specificity of signals for dentistry, the analysis was restricted to the specific term "Oral Paraesthesia" and to reports concerning dental practice. Overall, 528 reports of 'Paraesthesias and dysaesthesias' were retrieved, corresponding to 573 drug-reaction pairs (247 lidocaine, 99 bupivacaine, 85 articaine, 30 prilocaine, 112 others). The signal was significant only for articaine (ROR=18.38; 95CI = 13.95-24.21) and prilocaine (2.66; 1.82-3.90). The analysis of the specific term "Oral Paraesthesia" retrieved 82 reports corresponding to 90 drug-reaction pairs (37 articaine, 19 lidocaine, 14 prilocaine, 7 bupivacaine, 13 others) and confirmed the signal for articaine (58.77; 37.82-91.31) and prilocaine (8.73; 4.89-15.57). The analysis of reports concerning dental procedures retrieved a signal for articaine, both for any procedures (8.84; 2.79-27.97) and for non-surgical ones (15.79; 1.87-133.46). In conclusion, among local anaesthetics, only articaine and prilocaine generated a signal of paraesthesia, especially when used in dentistry. © 2015 Nordic Association for the Publication of BCPT (former Nordic Pharmacological Society).

  15. The impact of economic austerity and prosperity events on suicide in Greece: a 30-year interrupted time-series analysis

    PubMed Central

    Branas, Charles C; Kastanaki, Anastasia E; Michalodimitrakis, Manolis; Tzougas, John; Kranioti, Elena F; Theodorakis, Pavlos N; Carr, Brendan G; Wiebe, Douglas J

    2015-01-01

    Objectives To complete a 30-year interrupted time-series analysis of the impact of austerity-related and prosperity-related events on the occurrence of suicide across Greece. Setting Greece from 1 January 1983 to 31 December 2012. Participants A total of 11 505 suicides, 9079 by men and 2426 by women, occurring in Greece over the study period. Primary and secondary outcomes National data from the Hellenic Statistical Authority assembled as 360 monthly counts of: all suicides, male suicides, female suicides and all suicides plus potentially misclassified suicides. Results In 30 years, the highest months of suicide in Greece occurred in 2012. The passage of new austerity measures in June 2011 marked the beginning of significant, abrupt and sustained increases in total suicides (+35.7%, p<0.001) and male suicides (+18.5%, p<0.01). Sensitivity analyses that figured in undercounting of suicides also found a significant, abrupt and sustained increase in June 2011 (+20.5%, p<0.001). Suicides by men in Greece also underwent a significant, abrupt and sustained increase in October 2008 when the Greek recession began (+13.1%, p<0.01), and an abrupt but temporary increase in April 2012 following a public suicide committed in response to austerity conditions (+29.7%, p<0.05). Suicides by women in Greece also underwent an abrupt and sustained increase in May 2011 following austerity-related events (+35.8%, p<0.05). One prosperity-related event, the January 2002 launch of the Euro in Greece, marked an abrupt but temporary decrease in male suicides (−27.1%, p<0.05). Conclusions This is the first multidecade, national analysis of suicide in Greece using monthly data. Select austerity-related events in Greece corresponded to statistically significant increases for suicides overall, as well as for suicides among men and women. The consideration of future austerity measures should give greater weight to the unintended mental health consequences that may follow and the public

  16. Influence of risk factors and past events on flood resilience in coastal megacities: Comparative analysis of NYC and Shanghai.

    PubMed

    Xian, Siyuan; Yin, Jie; Lin, Ning; Oppenheimer, Michael

    2018-01-01

    Coastal flood protection measures have been widely implemented to improve flood resilience. However, protection levels vary among coastal megacities globally. This study compares the distinct flood protection standards for two coastal megacities, New York City and Shanghai, and investigates potential influences such as risk factors and past flood events. Extreme value analysis reveals that, compared to NYC, Shanghai faces a significantly higher flood hazard. Flood inundation analysis indicates that Shanghai has a higher exposure to extreme flooding. Meanwhile, Shanghai's urban development, population, and economy have increased much faster than NYC's over the last three decades. These risk factors provide part of the explanation for the implementation of a relatively high level of protection (e.g. reinforced concrete sea-wall designed for a 200-year flood return level) in Shanghai and low protection (e.g. vertical brick and stone walls and sand dunes) in NYC. However, individual extreme flood events (typhoons in 1962, 1974, and 1981) seem to have had a greater impact on flood protection decision-making in Shanghai, while NYC responded significantly less to past events (with the exception of Hurricane Sandy). Climate change, sea level rise, and ongoing coastal development are rapidly changing the hazard and risk calculus for both cities and both would benefit from a more systematic and dynamic approach to coastal protection. Copyright © 2017 Elsevier B.V. All rights reserved.

  17. Transition-Region Ultraviolet Explosive Events in IRIS Si IV: A Statistical Analysis

    NASA Astrophysics Data System (ADS)

    Bartz, Allison

    2018-01-01

    Explosive events (EEs) in the solar transition region are characterized by broad, non-Gaussian line profiles with wings at Doppler velocities exceeding the speed of sound. We present a statistical analysis of 23 IRIS (Interface Region Imaging Spectrograph) sit-and-stare observations, observed between April 2014 and March 2017. Using the IRIS Si IV 1394 Å and 1403 Å spectral windows and the 1400Å Slit Jaw images we have identified 581 EEs. We found that most EEs last less than 20 min. and have a spatial scale on the slit less than 10”, agreeing with measurements in previous work. We observed most EEs in active regions, regardless of date of observation, but selection bias of IRIS observations cannot be ruled out. We also present preliminary findings of optical depth effects from our statistical study.

  18. Antigen-antibody biorecognition events as discriminated by noise analysis of force spectroscopy curves.

    PubMed

    Bizzarri, Anna Rita; Cannistraro, Salvatore

    2014-08-22

    Atomic force spectroscopy is able to extract kinetic and thermodynamic parameters of biomolecular complexes provided that the registered unbinding force curves could be reliably attributed to the rupture of the specific complex interactions. To this aim, a commonly used strategy is based on the analysis of the stretching features of polymeric linkers which are suitably introduced in the biomolecule-substrate immobilization procedure. Alternatively, we present a method to select force curves corresponding to specific biorecognition events, which relies on a careful analysis of the force fluctuations of the biomolecule-functionalized cantilever tip during its approach to the partner molecules immobilized on a substrate. In the low frequency region, a characteristic 1/f (α) noise with α equal to one (flickering noise) is found to replace white noise in the cantilever fluctuation power spectrum when, and only when, a specific biorecognition process between the partners occurs. The method, which has been validated on a well-characterized antigen-antibody complex, represents a fast, yet reliable alternative to the use of linkers which may involve additional surface chemistry and reproducibility concerns.

  19. Event identification by acoustic signature recognition

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dress, W.B.; Kercel, S.W.

    1995-07-01

    Many events of interest to the security commnnity produce acoustic emissions that are, in principle, identifiable as to cause. Some obvious examples are gunshots, breaking glass, takeoffs and landings of small aircraft, vehicular engine noises, footsteps (high frequencies when on gravel, very low frequencies. when on soil), and voices (whispers to shouts). We are investigating wavelet-based methods to extract unique features of such events for classification and identification. We also discuss methods of classification and pattern recognition specifically tailored for acoustic signatures obtained by wavelet analysis. The paper is divided into three parts: completed work, work in progress, and futuremore » applications. The completed phase has led to the successful recognition of aircraft types on landing and takeoff. Both small aircraft (twin-engine turboprop) and large (commercial airliners) were included in the study. The project considered the design of a small, field-deployable, inexpensive device. The techniques developed during the aircraft identification phase were then adapted to a multispectral electromagnetic interference monitoring device now deployed in a nuclear power plant. This is a general-purpose wavelet analysis engine, spanning 14 octaves, and can be adapted for other specific tasks. Work in progress is focused on applying the methods previously developed to speaker identification. Some of the problems to be overcome include recognition of sounds as voice patterns and as distinct from possible background noises (e.g., music), as well as identification of the speaker from a short-duration voice sample. A generalization of the completed work and the work in progress is a device capable of classifying any number of acoustic events-particularly quasi-stationary events such as engine noises and voices and singular events such as gunshots and breaking glass. We will show examples of both kinds of events and discuss their recognition likelihood.« less

  20. A general framework for time series data mining based on event analysis: application to the medical domains of electroencephalography and stabilometry.

    PubMed

    Lara, Juan A; Lizcano, David; Pérez, Aurora; Valente, Juan P

    2014-10-01

    There are now domains where information is recorded over a period of time, leading to sequences of data known as time series. In many domains, like medicine, time series analysis requires to focus on certain regions of interest, known as events, rather than analyzing the whole time series. In this paper, we propose a framework for knowledge discovery in both one-dimensional and multidimensional time series containing events. We show how our approach can be used to classify medical time series by means of a process that identifies events in time series, generates time series reference models of representative events and compares two time series by analyzing the events they have in common. We have applied our framework on time series generated in the areas of electroencephalography (EEG) and stabilometry. Framework performance was evaluated in terms of classification accuracy, and the results confirmed that the proposed schema has potential for classifying EEG and stabilometric signals. The proposed framework is useful for discovering knowledge from medical time series containing events, such as stabilometric and electroencephalographic time series. These results would be equally applicable to other medical domains generating iconographic time series, such as, for example, electrocardiography (ECG). Copyright © 2014 Elsevier Inc. All rights reserved.

  1. Interferometric observation of microlensing events

    NASA Astrophysics Data System (ADS)

    Cassan, Arnaud; Ranc, Clément

    2016-05-01

    Interferometric observations of microlensing events have the potential to provide unique constraints on the physical properties of the lensing systems. In this work, we first present a formalism that closely combines interferometric and microlensing observable quantities, which lead us to define an original microlensing (u, v) plane. We run simulations of long-baseline interferometric observations and photometric light curves to decide which observational strategy is required to obtain a precise measurement on vector Einstein radius. We finally perform a detailed analysis of the expected number of targets in the light of new microlensing surveys (2011+) which currently deliver 2000 alerts per year. We find that a few events are already at reach of long-baseline interferometers (CHARA, VLTI), and a rate of about six events per year is expected with a limiting magnitude of K ≃ 10. This number would increase by an order of magnitude by raising it to K ≃ 11. We thus expect that a new route for characterizing microlensing events will be opened by the upcoming generations of interferometers.

  2. High fold computer disk storage DATABASE for fast extended analysis of γ-rays events

    NASA Astrophysics Data System (ADS)

    Stézowski, O.; Finck, Ch.; Prévost, D.

    1999-03-01

    Recently spectacular technical developments have been achieved to increase the resolving power of large γ-ray spectrometers. With these new eyes, physicists are able to study the intricate nature of atomic nuclei. Concurrently more and more complex multidimensional analyses are needed to investigate very weak phenomena. In this article, we first present a software (DATABASE) allowing high fold coincidences γ-rays events to be stored on hard disk. Then, a non-conventional method of analysis, anti-gating procedure, is described. Two physical examples are given to explain how it can be used and Monte Carlo simulations have been performed to test the validity of this method.

  3. Getting the right blood to the right patient: the contribution of near-miss event reporting and barrier analysis.

    PubMed

    Kaplan, H S

    2005-11-01

    Safety and reliability in blood transfusion are not static, but are dynamic non-events. Since performance deviations continually occur in complex systems, their detection and correction must be accomplished over and over again. Non-conformance must be detected early enough to allow for recovery or mitigation. Near-miss events afford early detection of possible system weaknesses and provide an early chance at correction. National event reporting systems, both voluntary and involuntary, have begun to include near-miss reporting in their classification schemes, raising awareness for their detection. MERS-TM is a voluntary safety reporting initiative in transfusion. Currently 22 hospitals submit reports anonymously to a central database which supports analysis of a hospital's own data and that of an aggregate database. The system encourages reporting of near-miss events, where the patient is protected from receiving an unsuitable or incorrect blood component due to a planned or unplanned recovery step. MERS-TM data suggest approximately 90% of events are near-misses, with 10% caught after issue but before transfusion. Near-miss reporting may increase total reports ten-fold. The ratio of near-misses to events with harm is 339:1, consistent with other industries' ratio of 300:1, which has been proposed as a measure of reporting in event reporting systems. Use of a risk matrix and an event's relation to protective barriers allow prioritization of these events. Near-misses recovered by planned barriers occur ten times more frequently then unplanned recoveries. A bedside check of the patient's identity with that on the blood component is an essential, final barrier. How the typical two person check is performed, is critical. Even properly done, this check is ineffective against sampling and testing errors. Blood testing at bedside just prior to transfusion minimizes the risk of such upstream events. However, even with simple and well designed devices, training may be a

  4. Defining Extreme Events: A Cross-Disciplinary Review

    NASA Astrophysics Data System (ADS)

    McPhillips, Lauren E.; Chang, Heejun; Chester, Mikhail V.; Depietri, Yaella; Friedman, Erin; Grimm, Nancy B.; Kominoski, John S.; McPhearson, Timon; Méndez-Lázaro, Pablo; Rosi, Emma J.; Shafiei Shiva, Javad

    2018-03-01

    Extreme events are of interest worldwide given their potential for substantial impacts on social, ecological, and technical systems. Many climate-related extreme events are increasing in frequency and/or magnitude due to anthropogenic climate change, and there is increased potential for impacts due to the location of urbanization and the expansion of urban centers and infrastructures. Many disciplines are engaged in research and management of these events. However, a lack of coherence exists in what constitutes and defines an extreme event across these fields, which impedes our ability to holistically understand and manage these events. Here, we review 10 years of academic literature and use text analysis to elucidate how six major disciplines—climatology, earth sciences, ecology, engineering, hydrology, and social sciences—define and communicate extreme events. Our results highlight critical disciplinary differences in the language used to communicate extreme events. Additionally, we found a wide range in definitions and thresholds, with more than half of examined papers not providing an explicit definition, and disagreement over whether impacts are included in the definition. We urge distinction between extreme events and their impacts, so that we can better assess when responses to extreme events have actually enhanced resilience. Additionally, we suggest that all researchers and managers of extreme events be more explicit in their definition of such events as well as be more cognizant of how they are communicating extreme events. We believe clearer and more consistent definitions and communication can support transdisciplinary understanding and management of extreme events.

  5. Adverse Events following 12 and 18 Month Vaccinations: a Population-Based, Self-Controlled Case Series Analysis

    PubMed Central

    Wilson, Kumanan; Hawken, Steven; Kwong, Jeffrey C.; Deeks, Shelley; Crowcroft, Natasha S.; Van Walraven, Carl; Potter, Beth K.; Chakraborty, Pranesh; Keelan, Jennifer; Pluscauskas, Michael; Manuel, Doug

    2011-01-01

    Background Live vaccines have distinct safety profiles, potentially causing systemic reactions one to 2 weeks after administration. In the province of Ontario, Canada, live MMR vaccine is currently recommended at age 12 months and 18 months. Methods Using the self-controlled case series design we examined 271,495 12 month vaccinations and 184,312 18 month vaccinations to examine the relative incidence of the composite endpoint of emergency room visits or hospital admissions in consecutive one day intervals following vaccination. These were compared to a control period 20 to 28 days later. In a post-hoc analysis we examined the reasons for emergency room visits and the average acuity score at presentation for children during the at-risk period following the 12 month vaccine. Results Four to 12 days post 12 month vaccination, children had a 1.33 (1.29–1.38) increased relative incidence of the combined endpoint compared to the control period, or at least one event during the risk interval for every 168 children vaccinated. Ten to 12 days post 18 month vaccination, the relative incidence was 1.25 (95%, 1.17–1.33) which represented at least one excess event for every 730 children vaccinated. The primary reason for increased events was statistically significant elevations in emergency room visits following all vaccinations. There were non-significant increases in hospital admissions. There were an additional 20 febrile seizures for every 100,000 vaccinated at 12 months. Conclusions There are significantly elevated risks of primarily emergency room visits approximately one to two weeks following 12 and 18 month vaccination. Future studies should examine whether these events could be predicted or prevented. PMID:22174753

  6. Enhancing the Effectiveness of Significant Event Analysis: Exploring Personal Impact and Applying Systems Thinking in Primary Care

    PubMed Central

    McNaughton, Elaine; Bruce, David; Holly, Deirdre; Forrest, Eleanor; Macleod, Marion; Kennedy, Susan; Power, Ailsa; Toppin, Denis; Black, Irene; Pooley, Janet; Taylor, Audrey; Swanson, Vivien; Kelly, Moya; Ferguson, Julie; Stirling, Suzanne; Wakeling, Judy; Inglis, Angela; McKay, John; Sargeant, Joan

    2016-01-01

    Introduction: Significant event analysis (SEA) is well established in many primary care settings but can be poorly implemented. Reasons include the emotional impact on clinicians and limited knowledge of systems thinking in establishing why events happen and formulating improvements. To enhance SEA effectiveness, we developed and tested “guiding tools” based on human factors principles. Methods: Mixed-methods development of guiding tools (Personal Booklet—to help with emotional demands and apply a human factors analysis at the individual level; Desk Pad—to guide a team-based systems analysis; and a written Report Format) by a multiprofessional “expert” group and testing with Scottish primary care practitioners who submitted completed enhanced SEA reports. Evaluation data were collected through questionnaire, telephone interviews, and thematic analysis of SEA reports. Results: Overall, 149/240 care practitioners tested the guiding tools and submitted completed SEA reports (62.1%). Reported understanding of how to undertake SEA improved postintervention (P < .001), while most agreed that the Personal Booklet was practical (88/123, 71.5%) and relevant to dealing with related emotions (93/123, 75.6%). The Desk Pad tool helped focus the SEA on systems issues (85/123, 69.1%), while most found the Report Format clear (94/123, 76.4%) and would recommend it (88/123, 71.5%). Most SEA reports adopted a systems approach to analyses (125/149, 83.9%), care improvement (74/149, 49.7), or planned actions (42/149, 28.2%). Discussion: Applying human factors principles to SEA potentially enables care teams to gain a systems-based understanding of why things go wrong, which may help with related emotional demands and with more effective learning and improvement. PMID:27583996

  7. Enhancing the Effectiveness of Significant Event Analysis: Exploring Personal Impact and Applying Systems Thinking in Primary Care.

    PubMed

    Bowie, Paul; McNaughton, Elaine; Bruce, David; Holly, Deirdre; Forrest, Eleanor; Macleod, Marion; Kennedy, Susan; Power, Ailsa; Toppin, Denis; Black, Irene; Pooley, Janet; Taylor, Audrey; Swanson, Vivien; Kelly, Moya; Ferguson, Julie; Stirling, Suzanne; Wakeling, Judy; Inglis, Angela; McKay, John; Sargeant, Joan

    2016-01-01

    Significant event analysis (SEA) is well established in many primary care settings but can be poorly implemented. Reasons include the emotional impact on clinicians and limited knowledge of systems thinking in establishing why events happen and formulating improvements. To enhance SEA effectiveness, we developed and tested "guiding tools" based on human factors principles. Mixed-methods development of guiding tools (Personal Booklet-to help with emotional demands and apply a human factors analysis at the individual level; Desk Pad-to guide a team-based systems analysis; and a written Report Format) by a multiprofessional "expert" group and testing with Scottish primary care practitioners who submitted completed enhanced SEA reports. Evaluation data were collected through questionnaire, telephone interviews, and thematic analysis of SEA reports. Overall, 149/240 care practitioners tested the guiding tools and submitted completed SEA reports (62.1%). Reported understanding of how to undertake SEA improved postintervention (P < .001), while most agreed that the Personal Booklet was practical (88/123, 71.5%) and relevant to dealing with related emotions (93/123, 75.6%). The Desk Pad tool helped focus the SEA on systems issues (85/123, 69.1%), while most found the Report Format clear (94/123, 76.4%) and would recommend it (88/123, 71.5%). Most SEA reports adopted a systems approach to analyses (125/149, 83.9%), care improvement (74/149, 49.7), or planned actions (42/149, 28.2%). Applying human factors principles to SEA potentially enables care teams to gain a systems-based understanding of why things go wrong, which may help with related emotional demands and with more effective learning and improvement.

  8. Human Rights Event Detection from Heterogeneous Social Media Graphs.

    PubMed

    Chen, Feng; Neill, Daniel B

    2015-03-01

    Human rights organizations are increasingly monitoring social media for identification, verification, and documentation of human rights violations. Since manual extraction of events from the massive amount of online social network data is difficult and time-consuming, we propose an approach for automated, large-scale discovery and analysis of human rights-related events. We apply our recently developed Non-Parametric Heterogeneous Graph Scan (NPHGS), which models social media data such as Twitter as a heterogeneous network (with multiple different node types, features, and relationships) and detects emerging patterns in the network, to identify and characterize human rights events. NPHGS efficiently maximizes a nonparametric scan statistic (an aggregate measure of anomalousness) over connected subgraphs of the heterogeneous network to identify the most anomalous network clusters. It summarizes each event with information such as type of event, geographical locations, time, and participants, and provides documentation such as links to videos and news reports. Building on our previous work that demonstrates the utility of NPHGS for civil unrest prediction and rare disease outbreak detection, we present an analysis of human rights events detected by NPHGS using two years of Twitter data from Mexico. NPHGS was able to accurately detect relevant clusters of human rights-related tweets prior to international news sources, and in some cases, prior to local news reports. Analysis of social media using NPHGS could enhance the information-gathering missions of human rights organizations by pinpointing specific abuses, revealing events and details that may be blocked from traditional media sources, and providing evidence of emerging patterns of human rights violations. This could lead to more timely, targeted, and effective advocacy, as well as other potential interventions.

  9. Event generators for address event representation transmitters

    NASA Astrophysics Data System (ADS)

    Serrano-Gotarredona, Rafael; Serrano-Gotarredona, Teresa; Linares Barranco, Bernabe

    2005-06-01

    Address Event Representation (AER) is an emergent neuromorphic interchip communication protocol that allows for real-time virtual massive connectivity between huge number neurons located on different chips. By exploiting high speed digital communication circuits (with nano-seconds timings), synaptic neural connections can be time multiplexed, while neural activity signals (with mili-seconds timings) are sampled at low frequencies. Also, neurons generate 'events' according to their activity levels. More active neurons generate more events per unit time, and access the interchip communication channel more frequently, while neurons with low activity consume less communication bandwidth. In a typical AER transmitter chip, there is an array of neurons that generate events. They send events to a peripheral circuitry (let's call it "AER Generator") that transforms those events to neurons coordinates (addresses) which are put sequentially on an interchip high speed digital bus. This bus includes a parallel multi-bit address word plus a Rqst (request) and Ack (acknowledge) handshaking signals for asynchronous data exchange. There have been two main approaches published in the literature for implementing such "AER Generator" circuits. They differ on the way of handling event collisions coming from the array of neurons. One approach is based on detecting and discarding collisions, while the other incorporates arbitration for sequencing colliding events . The first approach is supposed to be simpler and faster, while the second is able to handle much higher event traffic. In this article we will concentrate on the second arbiter-based approach. Boahen has been publishing several techniques for implementing and improving the arbiter based approach. Originally, he proposed an arbitration squeme by rows, followed by a column arbitration. In this scheme, while one neuron was selected by the arbiters to transmit his event out of the chip, the rest of neurons in the array were

  10. Columbia Classification Algorithm of Suicide Assessment (C-CASA): classification of suicidal events in the FDA's pediatric suicidal risk analysis of antidepressants.

    PubMed

    Posner, Kelly; Oquendo, Maria A; Gould, Madelyn; Stanley, Barbara; Davies, Mark

    2007-07-01

    To evaluate the link between antidepressants and suicidal behavior and ideation (suicidality) in youth, adverse events from pediatric clinical trials were classified in order to identify suicidal events. The authors describe the Columbia Classification Algorithm for Suicide Assessment (C-CASA), a standardized suicidal rating system that provided data for the pediatric suicidal risk analysis of antidepressants conducted by the Food and Drug Administration (FDA). Adverse events (N=427) from 25 pediatric antidepressant clinical trials were systematically identified by pharmaceutical companies. Randomly assigned adverse events were evaluated by three of nine independent expert suicidologists using the Columbia classification algorithm. Reliability of the C-CASA ratings and agreement with pharmaceutical company classification were estimated. Twenty-six new, possibly suicidal events (behavior and ideation) that were not originally identified by pharmaceutical companies were identified in the C-CASA, and 12 events originally labeled as suicidal by pharmaceutical companies were eliminated, which resulted in a total of 38 discrepant ratings. For the specific label of "suicide attempt," a relatively low level of agreement was observed between the C-CASA and pharmaceutical company ratings, with the C-CASA reporting a 50% reduction in ratings. Thus, although the C-CASA resulted in the identification of more suicidal events overall, fewer events were classified as suicide attempts. Additionally, the C-CASA ratings were highly reliable (intraclass correlation coefficient [ICC]=0.89). Utilizing a methodical, anchored approach to categorizing suicidality provides an accurate and comprehensive identification of suicidal events. The FDA's audit of the C-CASA demonstrated excellent transportability of this approach. The Columbia algorithm was used to classify suicidal adverse events in the recent FDA adult antidepressant safety analyses and has also been mandated to be applied to all

  11. Factor analysis of safety for visitors to a mega-event.

    PubMed

    Kwon, Young Guk; Park, Hyun Jee

    2002-01-01

    This paper investigated the safety factors considered by visitors to the Kwangju Biennale 2000 and analyzed the correlation between the safety factors and the demographic characteristics of the visitors. Global tourism increased throughout the 1990s, with the biggest surge occurring in the Asia-Pacific region. Long-distance travel is also increasing, and at a rate faster than the global average. The opportunities for event tourism appear to be strong almost everywhere, even though recessions may have an impact on these destinations. Along with this upward trend, competition for more desirable tourists is also surging (Getz, 1997). Therefore event tourism is appearing as a powerful method in the fierce competition around the tourism industry.

  12. Alcohol intoxication in the context of major public holidays, sporting and social events: a time-series analysis in Melbourne, Australia, 2000-2009.

    PubMed

    Lloyd, Belinda; Matthews, Sharon; Livingston, Michael; Jayasekara, Harindra; Smith, Karen

    2013-04-01

    To assess the relationship between ambulance attendances, emergency department (ED) presentations and hospital admissions for acute alcohol intoxication and the timing of public holidays, sporting and social events. Time-series analysis was used to explore trends in intoxication in the context of major events. Population of Melbourne, Victoria, Australia between 2000 and 2009. All patients attended by ambulance, presenting to hospital EDs, or admitted to hospital who were classified as acutely alcohol intoxicated. Analysis of daily numbers of presentations for acute alcohol intoxication associated with major events were undertaken, including lead and lag effects. Analyses controlled for day of week and month of year to address temporal and seasonal variations. Alcohol intoxication presentations were significantly elevated the day before all public holidays, with intoxication cases on the day of public holidays only higher on New Year's Day (ambulance 6.57, 95% confidence intervals (CI): 3.4-9.74; ED 3.34, 95% CI: 1.28-5.4) and ANZAC Day (ambulance 3.71, 95% CI: 0.68-6.75). The Australian Football League (AFL) Grand Final (ED 2.37, 95% CI: 0.55-4.19), Commonwealth Games (ED 2.45, 95% CI: 0.6-4.3) and Melbourne Cup Day (ambulance 6.14, 95% CI: 2.42-9.85) represented the sporting events with significant elevations in acute intoxication requiring medical attention. The last working day before Christmas was the only social event where a significant increase in acute intoxication occurred (ambulance 8.98, 95% CI: 6.8-11.15). Acute alcohol intoxication cases requiring ambulance, emergency department and hospital in-patient treatment increase substantially on the day preceding public holidays and other major social events. © 2012 The Authors, Addiction © 2012 Society for the Study of Addiction.

  13. Event Boundaries Trigger Rapid Memory Reinstatement of the Prior Events to Promote Their Representation in Long-Term Memory.

    PubMed

    Sols, Ignasi; DuBrow, Sarah; Davachi, Lila; Fuentemilla, Lluís

    2017-11-20

    Although everyday experiences unfold continuously over time, shifts in context, or event boundaries, can influence how those events come to be represented in memory [1-4]. Specifically, mnemonic binding across sequential representations is more challenging at context shifts, such that successful temporal associations are more likely to be formed within than across contexts [1, 2, 5-9]. However, in order to preserve a subjective sense of continuity, it is important that the memory system bridge temporally adjacent events, even if they occur in seemingly distinct contexts. Here, we used pattern similarity analysis to scalp electroencephalographic (EEG) recordings during a sequential learning task [2, 3] in humans and showed that the detection of event boundaries triggered a rapid memory reinstatement of the just-encoded sequence episode. Memory reactivation was detected rapidly (∼200-800 ms from the onset of the event boundary) and was specific to context shifts that were preceded by an event sequence with episodic content. Memory reinstatement was not observed during the sequential encoding of events within an episode, indicating that memory reactivation was induced specifically upon context shifts. Finally, the degree of neural similarity between neural responses elicited during sequence encoding and at event boundaries correlated positively with participants' ability to later link across sequences of events, suggesting a critical role in binding temporally adjacent events in long-term memory. Current results shed light onto the neural mechanisms that promote episodic encoding not only for information within the event, but also, importantly, in the ability to link across events to create a memory representation of continuous experience. Copyright © 2017 Elsevier Ltd. All rights reserved.

  14. Pulse Rate and Transit Time Analysis to Predict Hypotension Events After Spinal Anesthesia During Programmed Cesarean Labor.

    PubMed

    Bolea, Juan; Lázaro, Jesús; Gil, Eduardo; Rovira, Eva; Remartínez, José M; Laguna, Pablo; Pueyo, Esther; Navarro, Augusto; Bailón, Raquel

    2017-09-01

    Prophylactic treatment has been proved to reduce hypotension incidence after spinal anesthesia during cesarean labor. However, the use of pharmacological prophylaxis could carry out undesirable side-effects on mother and fetus. Thus, the prediction of hypotension becomes an important challenge. Hypotension events are hypothesized to be related to a malfunctioning of autonomic nervous system (ANS) regulation of blood pressure. In this work, ANS responses to positional changes of 51 pregnant women programmed for a cesarean labor were explored for hypotension prediction. Lateral and supine decubitus, and sitting position were considered while electrocardiographic and pulse photoplethysmographic signals were recorded. Features based on heart rate variability, pulse rate variability (PRV) and pulse transit time (PTT) analysis were used in a logistic regression classifier. The results showed that PRV irregularity changes, assessed by approximate entropy, from supine to lateral decubitus, and standard deviation of PTT in supine decubitus were found as the combination of features that achieved the best classification results sensitivity of 76%, specificity of 70% and accuracy of 72%, being normotensive the positive class. Peripheral regulation and blood pressure changes, measured by PRV and PTT analysis, could help to predict hypotension events reducing prophylactic side-effects in the low-risk population.

  15. Analysis and Prediction of West African Moist Events during the Boreal Spring of 2009

    NASA Astrophysics Data System (ADS)

    Mera, Roberto Javier

    Weather and climate in Sahelian West Africa are dominated by two major wind systems, the southwesterly West African Monsoon (WAM) and the northeasterly (Harmattan) trade winds. In addition to the agricultural benefit of the WAM, the public health sector is affected given the relationship between the onset of moisture and end of meningitis outbreaks. Knowledge and prediction of moisture distribution during the boreal spring is vital to the mitigation of meningitis by providing guidance for vaccine dissemination. The goal of the present study is to (a) develop a climatology and conceptual model of the moisture regime during the boreal spring, (b) investigate the role of extra-tropical and Convectively-coupled Equatorial Waves (CCEWs) on the modulation of westward moving synoptic waves and (c) determine the efficacy of a regional model as a tool for predicting moisture variability. Medical reports during 2009, along with continuous meteorological observations at Kano, Nigeria, showed that the advent of high humidity correlated with cessation of the disease. Further analysis of the 2009 boreal spring elucidated the presence of short-term moist events that modulated surface moisture on temporal scales relevant to the health sector. The May moist event (MME) provided insight into interplays among climate anomalies, extra-tropical systems, equatorially trapped waves and westward-propagating synoptic disturbances. The synoptic disturbance initiated 7 May and traveled westward to the coast by 12 May. There was a marked, semi-stationary moist anomaly in the precipitable water field (kg m-2) east of 10°E through late April and early May, that moved westward at the time of the MME. Further inspection revealed a mid-latitude system may have played a role in increasing the latitudinal amplitude of the MME. CCEWs were also found to have an impact on the MME. A coherent Kelvin wave propagated through the region, providing increased monsoonal flow and heightened convection. A

  16. Extreme precipitation events and related weather patterns over Iraq

    NASA Astrophysics Data System (ADS)

    raheem Al-nassar, Ali; Sangrà, Pablo; Alarcón, Marta

    2016-04-01

    This study aims to investigate the extreme precipitation events and the associated weather phenomena in the Middle East and particularly in Iraq. For this purpose we used Baghdad daily precipitation records from the Iraqi Meteorological and Seismology Organization combined with ECMWF (ERA-Interim) reanalysis data for the period from January 2002 to December 2013. Extreme events were found statistically at the 90% percentile of the recorded precipitation, and were highly correlated with hydrological flooding in some cities of Iraq. We identified fifteen extreme precipitation events. The analysis of the corresponding weather patterns (500 hPa and 250 hPa geopotential and velocity field distribution) indicated that 5 events were related with cut off low causing the highest precipitation (180 mm), 3 events related with rex block (158 mm), 3 events related with jet streak occurrence (130 mm) and 4 events related with troughs (107 mm). . Five of these events caused flash floods and in particular one of them related with a rex block was the most dramatic heavy rain event in Iraq in 30 years. We investigated for each case the convective instability and dynamical forcing together with humidity sources. For convective instability we explored the distribution of the K index and SWEAT index. For dynamical forcing we analyzed at several levels Q vector, divergence, potential and relative vorticity advection and omega vertical velocity. Source of humidity was investigated through humidity and convergence of specific humidity distribution. One triggering factor of all the events is the advection and convergence of humidity from the Red Sea and the Persian Gulf. Therefore a necessary condition for extreme precipitation in Iraq is the advection and convergence of humidity from the Red Sea and Persian Gulf. Our preliminary analysis also indicates that extreme precipitation events are primary dynamical forced playing convective instability a secondary role.

  17. Extreme Weather Events and Climate Change Attribution

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Thomas, Katherine

    A report from the National Academies of Sciences, Engineering, and Medicine concludes it is now possible to estimate the influence of climate change on some types of extreme events. The science of extreme event attribution has advanced rapidly in recent years, giving new insight to the ways that human-caused climate change can influence the magnitude or frequency of some extreme weather events. This report examines the current state of science of extreme weather attribution, and identifies ways to move the science forward to improve attribution capabilities. Confidence is strongest in attributing types of extreme events that are influenced by climatemore » change through a well-understood physical mechanism, such as, the more frequent heat waves that are closely connected to human-caused global temperature increases, the report finds. Confidence is lower for other types of events, such as hurricanes, whose relationship to climate change is more complex and less understood at present. For any extreme event, the results of attribution studies hinge on how questions about the event's causes are posed, and on the data, modeling approaches, and statistical tools chosen for the analysis.« less

  18. Nature of events and alcohol-related content in marketing materials at a university freshers' fair: a summative content analysis.

    PubMed

    Fuller, A; Fleming, K M; Szatkowski, L; Bains, M

    2017-12-15

    The transition to university is a potentially influential time upon students' drinking behaviour. This study explored the nature of activities and alcohol-related content in marketing materials from student-led societies and local businesses provided to students, at a university freshers' fair in the UK. All marketing materials handed out at the fair were collected across the 5-day event in September 2015. Written and visual content was analysed using a summative qualitative content analysis. Most student-led societies promoted social events they were hosting (n = 530), most of which took place in a drinking venue or referred to drinking (n = 335). Only four explicitly alcohol-free events were promoted. Student-led societies also promoted activities relating to their interest, e.g. sports training (n = 519), a small proportion of which had references to drinking and drinking venues (n = 54). Three societies provided promotional handouts from local bars or nightclubs. Local bars, pubs and nightclubs promoted events they hosted (n = 81) as well as alcoholic drink promotions (n = 79) and alcohol branded advertising (n = 22), albeit infrequently for the latter. In the first week of university, students are exposed to alcohol-related events, promotions and advertising, which may act as an incentive to participate in drinking. © The Author(s) 2017. Published by Oxford University Press on behalf of Faculty of Public Health. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com

  19. Estimate of neutrons event-by-event in DREAM

    NASA Astrophysics Data System (ADS)

    Hauptman, John; DREAM Collaboration

    2009-04-01

    We have measured the contribution of neutrons to hadronic showers in the DREAM module event-by-event as a means to estimate the event-by-event fluctuations in binding energy losses by hadrons as they break up nuclei of the Cu absorber. We make a preliminary assessment of the consequences for hadronic energy resolution in dual-readout calorimeters.

  20. A network of discrete events for the representation and analysis of diffusion dynamics.

    PubMed

    Pintus, Alberto M; Pazzona, Federico G; Demontis, Pierfranco; Suffritti, Giuseppe B

    2015-11-14

    We developed a coarse-grained description of the phenomenology of diffusive processes, in terms of a space of discrete events and its representation as a network. Once a proper classification of the discrete events underlying the diffusive process is carried out, their transition matrix is calculated on the basis of molecular dynamics data. This matrix can be represented as a directed, weighted network where nodes represent discrete events, and the weight of edges is given by the probability that one follows the other. The structure of this network reflects dynamical properties of the process of interest in such features as its modularity and the entropy rate of nodes. As an example of the applicability of this conceptual framework, we discuss here the physics of diffusion of small non-polar molecules in a microporous material, in terms of the structure of the corresponding network of events, and explain on this basis the diffusivity trends observed. A quantitative account of these trends is obtained by considering the contribution of the various events to the displacement autocorrelation function.

  1. Probabilistic Cross-identification of Cosmic Events

    NASA Astrophysics Data System (ADS)

    Budavári, Tamás

    2011-08-01

    I discuss a novel approach to identifying cosmic events in separate and independent observations. The focus is on the true events, such as supernova explosions, that happen once and, hence, whose measurements are not repeatable. Their classification and analysis must make the best use of all available data. Bayesian hypothesis testing is used to associate streams of events in space and time. Probabilities are assigned to the matches by studying their rates of occurrence. A case study of Type Ia supernovae illustrates how to use light curves in the cross-identification process. Constraints from realistic light curves happen to be well approximated by Gaussians in time, which makes the matching process very efficient. Model-dependent associations are computationally more demanding but can further boost one's confidence.

  2. Signals of New Physics in the Underlying Event

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Harnik, Roni; /Stanford U., ITP /SLAC; Wizansky, Tommer

    2010-06-11

    LHC searches for new physics focus on combinations of hard physics objects. In this work we propose a qualitatively different soft signal for new physics at the LHC - the 'anomalous underlying event'. Every hard LHC event will be accompanied by a soft underlying event due to QCD and pile-up effects. Though it is often used for QCD and monte carlo studies, here we propose the incorporation of an underlying event analysis in some searches for new physics. An excess of anomalous underlying events may be a smoking-gun signal for particular new physics scenarios such as 'quirks' or 'hidden valleys'more » in which large amounts of energy may be emitted by a large multiplicity of soft particles. We discuss possible search strategies for such soft diffuse signals in the tracking system and calorimetry of the LHC experiments. We present a detailed study of the calorimetric signal in a concrete example, a simple quirk model motivated by folded supersymmetry. In these models the production and radiative decay of highly excited quirk bound states leads to an 'antenna pattern' of soft unclustered energy. Using a dedicated simulation of a toy detector and a 'CMB-like' multipole analysis we compare the signal to the expected backgrounds.« less

  3. Using a Six Sigma Fishbone Analysis Approach To Evaluate the Effect of Extreme Weather Events on Salmonella Positives in Young Chicken Slaughter Establishments.

    PubMed

    Linville, John W; Schumann, Douglas; Aston, Christopher; Defibaugh-Chavez, Stephanie; Seebohm, Scott; Touhey, Lucy

    2016-12-01

    A six sigma fishbone analysis approach was used to develop a machine learning model in SAS, Version 9.4, by using stepwise linear regression. The model evaluated the effect of a wide variety of variables, including slaughter establishment operational measures, normal (30-year average) weather, and extreme weather events on the rate of Salmonella -positive carcasses in young chicken slaughter establishments. Food Safety and Inspection Service (FSIS) verification carcass sampling data, as well as corresponding data from the National Oceanographic and Atmospheric Administration and the Federal Emergency Management Agency, from September 2011 through April 2015, were included in the model. The results of the modeling show that in addition to basic establishment operations, normal weather patterns, differences from normal and disaster events, including time lag weather and disaster variables, played a role in explaining the Salmonella percent positive that varied by slaughter volume quartile. Findings show that weather and disaster events should be considered as explanatory variables when assessing pathogen-related prevalence analysis or research and slaughter operational controls. The apparent significance of time lag weather variables suggested that at least some of the impact on Salmonella rates occurred after the weather events, which may offer opportunities for FSIS or the poultry industry to implement interventions to mitigate those effects.

  4. Grain-size analysis and sediment dynamics of hurricane-induced event beds in a coastal New England pond

    NASA Astrophysics Data System (ADS)

    Castagno, K. A.; Ruehr, S. A.; Donnelly, J. P.; Woodruff, J. D.

    2017-12-01

    Coastal populations have grown increasingly susceptible to the impacts of tropical cyclone events as they grow in size, wealth, and infrastructure. Changes in tropical cyclone frequency and intensity, augmented by a changing climate, pose an increasing threat of property damage and loss of life. Reconstructions of intense-hurricane landfalls from a series of southeastern New England sediment cores identify a series of events spanning the past 2,000 years. Though the frequency of these landfalls is well constrained, the intensity of these storms, particularly those for which no historical record exists, is not. This study analyzes the grain-size distribution of major storm event beds along a transect of sediment cores from a kettle pond in Falmouth, MA. The grain-size distribution of each event is determined using an image processing, size, and shape analyzer. The depositional patterns and changes in grain-size distribution in these fine-grained systems may both spatially and temporally reveal characteristics of both storm intensity and the nature of sediment deposition. An inverse-modeling technique using this kind of grain-size analysis to determine past storm intensity has been explored in back-barrier lagoon systems in the Caribbean, but limited research has assessed its utility to assess deposits from back-barrier ponds in the northeastern United States. Increases in hurricane intensity may be closely tied to increases in sea surface temperature. As such, research into these prehistoric intervals of increased frequency and/or intensity provides important insight into the current and future hurricane risks facing coastal communities in New England.

  5. Biological Event Modeling for Response Planning

    NASA Astrophysics Data System (ADS)

    McGowan, Clement; Cecere, Fred; Darneille, Robert; Laverdure, Nate

    People worldwide continue to fear a naturally occurring or terrorist-initiated biological event. Responsible decision makers have begun to prepare for such a biological event, but critical policy and system questions remain: What are the best courses of action to prepare for and react to such an outbreak? Where resources should be stockpiled? How many hospital resources—doctors, nurses, intensive-care beds—will be required? Will quarantine be necessary? Decision analysis tools, particularly modeling and simulation, offer ways to address and help answer these questions.

  6. Multi-threaded Event Processing with DANA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    David Lawrence; Elliott Wolin

    2007-05-14

    The C++ data analysis framework DANA has been written to support the next generation of Nuclear Physics experiments at Jefferson Lab commensurate with the anticipated 12GeV upgrade. The DANA framework was designed to allow multi-threaded event processing with a minimal impact on developers of reconstruction software. This document describes how DANA implements multi-threaded event processing and compares it to simply running multiple instances of a program. Also presented are relative reconstruction rates for Pentium4, Xeon, and Opteron based machines.

  7. Single-cell analysis of pyroptosis dynamics reveals conserved GSDMD-mediated subcellular events that precede plasma membrane rupture.

    PubMed

    de Vasconcelos, Nathalia M; Van Opdenbosch, Nina; Van Gorp, Hanne; Parthoens, Eef; Lamkanfi, Mohamed

    2018-04-17

    Pyroptosis is rapidly emerging as a mechanism of anti-microbial host defense, and of extracellular release of the inflammasome-dependent cytokines interleukin (IL)-1β and IL-18, which contributes to autoinflammatory pathology. Caspases 1, 4, 5 and 11 trigger this regulated form of necrosis by cleaving the pyroptosis effector gasdermin D (GSDMD), causing its pore-forming amino-terminal domain to oligomerize and perforate the plasma membrane. However, the subcellular events that precede pyroptotic cell lysis are ill defined. In this study, we triggered primary macrophages to undergo pyroptosis from three inflammasome types and recorded their dynamics and morphology using high-resolution live-cell spinning disk confocal laser microscopy. Based on quantitative analysis of single-cell subcellular events, we propose a model of pyroptotic cell disintegration that is initiated by opening of GSDMD-dependent ion channels or pores that are more restrictive than recently proposed GSDMD pores, followed by osmotic cell swelling, commitment of mitochondria and other membrane-bound organelles prior to sudden rupture of the plasma membrane and full permeability to intracellular proteins. This study provides a dynamic framework for understanding cellular changes that occur during pyroptosis, and charts a chronological sequence of GSDMD-mediated subcellular events that define pyroptotic cell death at the single-cell level.

  8. Evidence of validity of the Stress-Producing Life Events (SPLE) instrument.

    PubMed

    Rizzini, Marta; Santos, Alcione Miranda Dos; Silva, Antônio Augusto Moura da

    2018-01-01

    OBJECTIVE Evaluate the construct validity of a list of eight Stressful Life Events in pregnant women. METHODS A cross-sectional study was conducted with 1,446 pregnant women in São Luís, MA, and 1,364 pregnant women in Ribeirão Preto, SP (BRISA cohort), from February 2010 to June 2011. In the exploratory factorial analysis, the promax oblique rotation was used and for the calculation of the internal consistency, we used the compound reliability. The construct validity was determined by means of the confirmatory factorial analysis with the method of estimation of weighted least squares adjusted by the mean and variance. RESULTS The model with the best fit in the exploratory analysis was the one that retained three factors with a cumulative variance of 61.1%. The one-factor model did not obtain a good fit in both samples in the confirmatory analysis. The three-factor model called Stress-Producing Life Events presented a good fit (RMSEA < 0.05; CFI/TLI > 0.90) for both samples. CONCLUSIONS The Stress-Producing Life Events constitute a second order construct with three dimensions related to health, personal and financial aspects and violence. This study found evidence that confirms the construct validity of a list of stressor events, entitled Stress-Producing Life Events Inventory.

  9. Using timed event sequential data in nursing research.

    PubMed

    Pecanac, Kristen E; Doherty-King, Barbara; Yoon, Ju Young; Brown, Roger; Schiefelbein, Tony

    2015-01-01

    Measuring behavior is important in nursing research, and innovative technologies are needed to capture the "real-life" complexity of behaviors and events. The purpose of this article is to describe the use of timed event sequential data in nursing research and to demonstrate the use of this data in a research study. Timed event sequencing allows the researcher to capture the frequency, duration, and sequence of behaviors as they occur in an observation period and to link the behaviors to contextual details. Timed event sequential data can easily be collected with handheld computers, loaded with a software program designed for capturing observations in real time. Timed event sequential data add considerable strength to analysis of any nursing behavior of interest, which can enhance understanding and lead to improvement in nursing practice.

  10. Impact of Azithromycin on Pregnancy Prolongation in Women at Risk of Preterm Labor: A Time-to-Event Analysis.

    PubMed

    Goyer, Isabelle; Ferland, Gabrielle; Ruo, Ni; Morin, Caroline; Brochet, Marie-Sophie; Morin, Lucie; Ferreira, Ema

    2016-09-13

    Since 2006, the empiric use of azithromycin in women at risk of premature birth has become prevalent in our institution without any evidence of its efficacy. Although antibiotics can prolong pregnancy in preterm prolonged rupture of membranes, no published data are available for women with intact membranes. To describe the purpose of adding azithromycin to the usual treatments (cerclage, tocolysis, rest, etc.) to prolong pregnancy in women with intact membranes who are at risk of or already in preterm labour. A retrospective observational cohort study was done at a Mother-Child University Hospital Centre. Patients admitted to obstetric ward who received azithromycin between January 1 st , 2006 and August 1 st , 2010 were included. A total of 127 exposed women were matched to 127 controls through medical records and pharmacy software. A time-to-event analysis was done to compare gestational age at the time of the recorded composite event (delivery, or rupture of membranes, or second intervention to prolong pregnancy). To compare proportions of composite event at different time points, χ 2 tests were used. Patients who received azithromycin had a more severe condition at presentation. Once adjusted for confounding factors, prolongation of pregnancy (HR =1.049; CI 95%: 0.774-1.421 [p=0.758]) and gestational age at the event (HR=1.200; CI 95%: 0.894-1.609 [p=0.225]) did not differ between the groups. The proportions of women with an event ≥7 days post-diagnosis or ≥37 gestational weeks were similar. Azithromycin was added to medical therapy in a more at-risk population and no clear benefit was measured.

  11. [Adverse events in general surgery. A prospective analysis of 13,950 consecutive patients].

    PubMed

    Rebasa, Pere; Mora, Laura; Vallverdú, Helena; Luna, Alexis; Montmany, Sandra; Romaguera, Andreu; Navarro, Salvador

    2011-11-01

    Adverse event (AE) rates in General Surgery vary, according to different authors and recording methods, between 2% and 30%. Six years ago we designed a prospective AE recording system to change patient safety culture in our Department. We present the results of this work after a 6 year follow-up. The AE, sequelae and health care errors in a University Hospital surgery department were recorded. An analysis of each incident recorded was performed by a reviewer. The data was entered into data base for rapid access and consultation. The results were routinely presented in Departmental morbidity-mortality sessions. A total of 13,950 patients had suffered 11,254 AE, which affected 5142 of them (36.9% of admissions). A total of 920 patients were subjected to at least one health care error (6.6% of admissions). This meant that 6.6% of our patients suffered an avoidable AE. The overall mortality at 5 years in our department was 2.72% (380 deaths). An adverse event was implicated in the death of the patient in 180 cases (1.29% of admissions). In 49 cases (0.35% of admissions), mortality could be attributed to an avoidable AE. After 6 years there tends to be an increasingly lower incidence of errors. The exhaustive and prospective recording of AE leads to changes in patient safety culture in a Surgery Department and helps decrease the incidence of health care errors. Copyright © 2011 AEC. Published by Elsevier Espana. All rights reserved.

  12. Seasonal prediction of extreme precipitation events and frequency of rainy days over Costa Rica, Central America, using Canonical Correlation Analysis

    NASA Astrophysics Data System (ADS)

    Maldonado, T.; Alfaro, E.; Fallas-López, B.; Alvarado, L.

    2013-04-01

    High mountains divide Costa Rica, Central America, into two main climate regions, the Pacific and Caribbean slopes, which are lee and windward, respectively, according to the North Atlantic trade winds - the dominant wind regime. The rain over the Pacific slope has a bimodal annual cycle, having two maxima, one in May-June and the other in August-September-October (ASO), separated by the mid-summer drought in July. A first maximum of deep convection activity, and hence a first maximum of precipitation, is reached when sea surface temperature (SST) exceeds 29 °C (around May). Then, the SST decreases to around 1 °C due to diminished downwelling solar radiation and stronger easterly winds (during July and August), resulting in a decrease in deep convection activity. Such a reduction in deep convection activity allows an increase in down welling solar radiation and a slight increase in SST (about 28.5 °C) by the end of August and early September, resulting once again in an enhanced deep convection activity, and, consequently, in a second maximum of precipitation. Most of the extreme events are found during ASO. Central American National Meteorological and Hydrological Services (NMHS) have periodic Regional Climate Outlook Fora (RCOF) to elaborate seasonal predictions. Recently, meetings after RCOF with different socioeconomic stakeholders took place to translate the probable climate impacts from predictions. From the feedback processes of these meetings has emerged that extreme event and rainy days seasonal predictions are necessary for different sectors. As is shown in this work, these predictions can be tailored using Canonical Correlation Analysis for rain during ASO, showing that extreme events and rainy days in Central America are influenced by interannual variability related to El Niño-Southern Oscillation and decadal variability associated mainly with Atlantic Multidecadal Oscillation. Analyzing the geographical distribution of the ASO-2010 disaster reports

  13. The impact of economic austerity and prosperity events on suicide in Greece: a 30-year interrupted time-series analysis.

    PubMed

    Branas, Charles C; Kastanaki, Anastasia E; Michalodimitrakis, Manolis; Tzougas, John; Kranioti, Elena F; Theodorakis, Pavlos N; Carr, Brendan G; Wiebe, Douglas J

    2015-02-02

    To complete a 30-year interrupted time-series analysis of the impact of austerity-related and prosperity-related events on the occurrence of suicide across Greece. Greece from 1 January 1983 to 31 December 2012. A total of 11 505 suicides, 9079 by men and 2426 by women, occurring in Greece over the study period. National data from the Hellenic Statistical Authority assembled as 360 monthly counts of: all suicides, male suicides, female suicides and all suicides plus potentially misclassified suicides. In 30 years, the highest months of suicide in Greece occurred in 2012. The passage of new austerity measures in June 2011 marked the beginning of significant, abrupt and sustained increases in total suicides (+35.7%, p<0.001) and male suicides (+18.5%, p<0.01). Sensitivity analyses that figured in undercounting of suicides also found a significant, abrupt and sustained increase in June 2011 (+20.5%, p<0.001). Suicides by men in Greece also underwent a significant, abrupt and sustained increase in October 2008 when the Greek recession began (+13.1%, p<0.01), and an abrupt but temporary increase in April 2012 following a public suicide committed in response to austerity conditions (+29.7%, p<0.05). Suicides by women in Greece also underwent an abrupt and sustained increase in May 2011 following austerity-related events (+35.8%, p<0.05). One prosperity-related event, the January 2002 launch of the Euro in Greece, marked an abrupt but temporary decrease in male suicides (-27.1%, p<0.05). This is the first multidecade, national analysis of suicide in Greece using monthly data. Select austerity-related events in Greece corresponded to statistically significant increases for suicides overall, as well as for suicides among men and women. The consideration of future austerity measures should give greater weight to the unintended mental health consequences that may follow and the public messaging of these policies and related events. Published by the BMJ Publishing Group

  14. Cloud structure evolution of heavy rain events from the East-West Pacific Ocean: a combined global observation analysis

    NASA Astrophysics Data System (ADS)

    Sekaranom, A. B.; Nurjani, E.; Pujiastuti, I.

    2018-04-01

    Heavy rain events are often associated with flood hazards as one of the most devastating events across the globe. It is therefore essential to identify the evolution of heavy rainfall cloud structures, primarily from global satellite observation, as a tool to provide better disaster early warning systems. To identify the mechanism of heavy rainfall systems and its relationship with cloud development, especially over The Pacific Ocean, we aim to study the westward evolution of the convective systems over this area. Several datasets from Tropical Rainfall Measuring Mission (TRMM), CloudSat GEOPROF product, and ECMWF-reanalysis (ERA) interim were utilized to characterize the evolution. Geolocation and orbital time-lag analysis of the three different datasets for more than 8 years (2006-2014) could provide information related to the evolution of cloud structures associated with heavy rain events. In the first step, a heavy rainfall database was generated from TRMM. The CloudSat coordinate and time position were then matched with TRMM coordinate and time position. All of the processes were programatically conducted in fortran programming language. The result shows a transition between East and West Pacific ocean for TMI data.

  15. Bayesian inference on risk differences: an application to multivariate meta-analysis of adverse events in clinical trials.

    PubMed

    Chen, Yong; Luo, Sheng; Chu, Haitao; Wei, Peng

    2013-05-01

    Multivariate meta-analysis is useful in combining evidence from independent studies which involve several comparisons among groups based on a single outcome. For binary outcomes, the commonly used statistical models for multivariate meta-analysis are multivariate generalized linear mixed effects models which assume risks, after some transformation, follow a multivariate normal distribution with possible correlations. In this article, we consider an alternative model for multivariate meta-analysis where the risks are modeled by the multivariate beta distribution proposed by Sarmanov (1966). This model have several attractive features compared to the conventional multivariate generalized linear mixed effects models, including simplicity of likelihood function, no need to specify a link function, and has a closed-form expression of distribution functions for study-specific risk differences. We investigate the finite sample performance of this model by simulation studies and illustrate its use with an application to multivariate meta-analysis of adverse events of tricyclic antidepressants treatment in clinical trials.

  16. New Perspectives on Long Run-out Rock Avalanches: A Dynamic Analysis of 20 Events in the Vaigat Strait, West Greenland

    NASA Astrophysics Data System (ADS)

    Benjamin, J.; Rosser, N. J.; Dunning, S.; Hardy, R. J.; Karim, K.; Szczucinski, W.; Norman, E. C.; Strzelecki, M.; Drewniak, M.

    2014-12-01

    simulate a series of events using a single set of parameters obtained by back-analysis of the Paatuut event alone. The results also hold important implications for our process understanding of rock avalanches in confined fjord settings, where correctly modelling material flux at the point of entry into the water is critical in tsunami generation.

  17. Adverse childhood events, substance abuse, and measures of affiliation.

    PubMed

    Zlotnick, Cheryl; Tam, Tammy; Robertson, Marjorie J

    2004-08-01

    Adverse childhood events may influence later behaviors, including adulthood substance use and social affiliation. Studies have noted high prevalence rates of adverse childhood experiences and adulthood substance abuse among homeless adults. Using an existing longitudinal, countywide probability sample of 397 homeless adults, we examine the relationships among adverse childhood events on adulthood substance use, and the relationship of these variables to affiliation. Almost 75% of the sample had experienced an adverse childhood event. Path analysis indicated adulthood substance abuse mediated the inverse relationship between adverse childhood events and two measures of adulthood affiliation. Thus, although there is a relationship between adverse childhood events and adulthood substance use, it is adulthood substance use that determines most aspects of affiliation.

  18. Genome-wide association analysis accounting for environmental factors through propensity-score matching: application to stressful live events in major depressive disorder.

    PubMed

    Power, Robert A; Cohen-Woods, Sarah; Ng, Mandy Y; Butler, Amy W; Craddock, Nick; Korszun, Ania; Jones, Lisa; Jones, Ian; Gill, Michael; Rice, John P; Maier, Wolfgang; Zobel, Astrid; Mors, Ole; Placentino, Anna; Rietschel, Marcella; Aitchison, Katherine J; Tozzi, Federica; Muglia, Pierandrea; Breen, Gerome; Farmer, Anne E; McGuffin, Peter; Lewis, Cathryn M; Uher, Rudolf

    2013-09-01

    Stressful life events are an established trigger for depression and may contribute to the heterogeneity within genome-wide association analyses. With depression cases showing an excess of exposure to stressful events compared to controls, there is difficulty in distinguishing between "true" cases and a "normal" response to a stressful environment. This potential contamination of cases, and that from genetically at risk controls that have not yet experienced environmental triggers for onset, may reduce the power of studies to detect causal variants. In the RADIANT sample of 3,690 European individuals, we used propensity score matching to pair cases and controls on exposure to stressful life events. In 805 case-control pairs matched on stressful life event, we tested the influence of 457,670 common genetic variants on the propensity to depression under comparable level of adversity with a sign test. While this analysis produced no significant findings after genome-wide correction for multiple testing, we outline a novel methodology and perspective for providing environmental context in genetic studies. We recommend contextualizing depression by incorporating environmental exposure into genome-wide analyses as a complementary approach to testing gene-environment interactions. Possible explanations for negative findings include a lack of statistical power due to small sample size and conditional effects, resulting from the low rate of adequate matching. Our findings underscore the importance of collecting information on environmental risk factors in studies of depression and other complex phenotypes, so that sufficient sample sizes are available to investigate their effect in genome-wide association analysis. Copyright © 2013 Wiley Periodicals, Inc.

  19. Standardized Analytical Methods for Environmental Restoration Following Homeland Security Events

    USDA-ARS?s Scientific Manuscript database

    Methodology was formulated for use in the event of a terrorist attack using a variety of chemical, radioactive, biological, and toxic agents. Standardized analysis procedures were determined for use should these events occur. This publication is annually updated....

  20. Propensity for Violence among Homeless and Runaway Adolescents: An Event History Analysis*

    PubMed Central

    Crawford, Devan M.; Whitbeck, Les B.; Hoyt, Dan R.

    2012-01-01

    Little is known about the prevalence of violent behaviors among homeless and runaway adolescents or the specific behavioral factors that influence violent behaviors across time. In this longitudinal study of 300 homeless and runaway adolescents aged 16–19 years at baseline, we use event history analysis to assess the factors associated with acts of violence over three years, controlling for individual propensities and time-varying behaviors. The results indicate that females, non-minorities, and non-heterosexuals were less likely to engage in violence across time. Those who met criteria for substance abuse disorders (i.e. alcohol abuse, alcohol dependence, drug abuse) were more likely to engage in violence. A history of caretaker abuse was associated with violent behaviors, as were street survival strategies such as selling drugs, participating in gang activity, and associating with deviant peers. Simply having spent time directly on the streets at any specific time point also increased the likelihood for violence. PMID:22865932

  1. [Spatiotemporal pattern analysis of event-related potentials elicited by emotional Stroop task].

    PubMed

    Liu, Qi; Liu, Ling; He, Hui; Zhou, Shu

    2007-05-01

    To investigate the spatiotemporal pattern of event-related potentials (ERPs) induced by emotional Stroop task. The ERPs of 19 channels were recorded from 13 healthy subjects while performing emotional Stroop task by pressing the buttons representing the colors in which the words denoting different emotions were displayed. A repeated-measures factorial design was adopted with three levels (word valence: positive, neutral and negative). The result of ERP analysis was presented in the form of statistical parametric mapping (SPM) of F value. No significant difference was found in either reaction time or accuracy. The SPM of ERPs suggested significant emotional valence effects in the occipital region (200-220 ms), the left and central frontal regions (270-300 ms), and the bilateral temporal and parietal cortex (560-580 and 620-630 ms, respectively). Processing of task-irrelevant emotional valence information involves the dynamic operation of extensive brain regions. The ERPs are more sensitive than the behavioral indices in emotional evaluation.

  2. Effects of Individual Nurse and Hospital Characteristics on Patient Adverse Events and Quality of Care: A Multilevel Analysis.

    PubMed

    Lee, Seung Eun; Vincent, Catherine; Dahinten, V Susan; Scott, Linda D; Park, Chang Gi; Dunn Lopez, Karen

    2018-06-14

    This study aimed to investigate effects of individual nurse and hospital characteristics on patient adverse events and quality of care using a multilevel approach. This is a secondary analysis of a combination of nurse survey data (N = 1,053 nurses) and facility data (N = 63 hospitals) in Canada. Multilevel ordinal logistic regression was employed to examine effects of individual nurse and hospital characteristics on patient adverse events. Multilevel linear regressions were used to investigate effects of individual nurse and hospital characteristics on quality of care. Organizational safety culture was associated with patient adverse events and quality of care. Controlling for effects of nurse and hospital characteristics, nurses in hospitals with a stronger safety culture were 64% less likely to report administration of wrong medication, time, or dose; 58% less likely to report patient falls with injury; and 60% less likely to report urinary tract infections; and were more likely to report higher levels of quality of care. Additionally, the effects of individual-level baccalaureate education and years of experience on quality of care differed across hospitals, and hospital-level nurse education interacted with individual-level baccalaureate education. This study makes significant contributions to existing knowledge regarding the positive effect of organizational safety culture on patient adverse events and quality of care. Healthcare organizations should strive to improve their safety culture by creating environments where healthcare providers trust each other, work collaboratively, and share accountability for patient safety and care quality. © 2018 Sigma Theta Tau International.

  3. Adverse events associated with deep brain stimulation for movement disorders: analysis of 510 consecutive cases.

    PubMed

    Patel, Daxa M; Walker, Harrison C; Brooks, Rebekah; Omar, Nidal; Ditty, Benjamin; Guthrie, Barton L

    2015-03-01

    Although numerous studies have focused on the efficacy of deep brain stimulation (DBS) for movement disorders, less is known about surgical adverse events, especially over longer time intervals. Here, we analyze adverse events in 510 consecutive cases from a tertiary movement disorders center at up to 10 years postoperatively. We conducted a retrospective review of adverse events from craniotomies between January 2003 and March 2013. The adverse events were categorized into 2 broad categories--immediate perioperative and time-dependent postoperative events. Across all targets, perioperative mental status change occurred in 18 (3.5%) cases, and symptomatic intracranial hemorrhage occurred in 4 (0.78%) cases. The most common hardware-related event was skin erosion in 13 (2.5%) cases. The most frequent stimulation-related event was speech disturbance in 16 (3.1%) cases. There were no significant differences among surgical targets with respect to the incidence of these events. Time-dependent postoperative events leading to the revision of a given DBS electrode for any reason occurred in 4.7% ± 1.0%, 9.3% ± 1.4%, and 12.4% ± 1.5% of electrodes at 1, 4, and 7 years postoperatively, respectively. Staged bilateral DBS was associated with approximately twice the risk of repeat surgery for electrode replacement vs unilateral surgery (P = .020). These data provide low incidences for adverse events in a large series of DBS surgeries for movement disorders at up to 10 years follow-up. Accurate estimates of adverse events will better inform patients and caregivers about the potential risks and benefits of surgery and provide normative data for process improvement.

  4. [Adverse events in patients from a pediatric hospital.

    PubMed

    Ornelas-Aguirre, José Manuel; Arriaga-Dávila, José de Jesús; Domínguez-Serrano, María Isabel; Guzmán-Bihouet, Beatriz Filomena; Navarrete-Navarro, Susana

    2013-01-01

    Background: detection of adverse events is part of the safety management in hospitalized patients. The objective of this study was to describe the incidence of adverse events that occurred in a pediatric hospital. Methods: cross-sectional study of the adverse events occurred in a pediatric hospital from 2007 to 2009. Factors associated with their developmental causes were identified. The statistical analysis was descriptive and bivariate, with contingency tables to estimate the relationship between those factors. A p value = 0.05 was considered significant. Results: a total of 177 adverse events were registered. When they began, human factor occurred in 23 cases (13 %, OR = 1.41, p = 0.001), organizational factor was present in 71 cases (40 %, OR = 1.91, p = 0.236) and technical factor in 46 cases (26 %, OR = 0.87, p = 0.01). Blows or bruises from falls as a result of adverse events occurred in 71 cases (40 %, 95 % CI = 64-78). Conclusions: we found 1.84 events per 100 hospital discharges during the study period. The fall of patients ranked first of the adverse events identified.

  5. Climate change impacts on extreme events in the United States: an uncertainty analysis

    EPA Science Inventory

    Extreme weather and climate events, such as heat waves, droughts and severe precipitation events, have substantial impacts on ecosystems and the economy. However, future climate simulations display large uncertainty in mean changes. As a result, the uncertainty in future changes ...

  6. Columbia Classification Algorithm of Suicide Assessment (C-CASA): Classification of Suicidal Events in the FDA’s Pediatric Suicidal Risk Analysis of Antidepressants

    PubMed Central

    Posner, Kelly; Oquendo, Maria A.; Gould, Madelyn; Stanley, Barbara; Davies, Mark

    2013-01-01

    Objective To evaluate the link between antidepressants and suicidal behavior and ideation (suicidality) in youth, adverse events from pediatric clinical trials were classified in order to identify suicidal events. The authors describe the Columbia Classification Algorithm for Suicide Assessment (C-CASA), a standardized suicidal rating system that provided data for the pediatric suicidal risk analysis of antide-pressants conducted by the Food and Drug Administration (FDA). Method Adverse events (N=427) from 25 pediatric antidepressant clinical trials were systematically identified by pharmaceutical companies. Randomly assigned adverse events were evaluated by three of nine independent expert suicidologists using the Columbia classification algorithm. Reliability of the C-CASA ratings and agreement with pharmaceutical company classification were estimated. Results Twenty-six new, possibly suicidal events (behavior and ideation) that were not originally identified by pharmaceutical companies were identified in the C-CASA, and 12 events originally labeled as suicidal by pharmaceutical companies were eliminated, which resulted in a total of 38 discrepant ratings. For the specific label of “suicide attempt,” a relatively low level of agreement was observed between the C-CASA and pharmaceutical company ratings, with the C-CASA reporting a 50% reduction in ratings. Thus, although the C-CASA resulted in the identification of more suicidal events overall, fewer events were classified as suicide attempts. Additionally, the C-CASA ratings were highly reliable (intraclass correlation coefficient [ICC]=0.89). Conclusions Utilizing a methodical, anchored approach to categorizing suicidality provides an accurate and comprehensive identification of suicidal events. The FDA’s audit of the C-CASA demonstrated excellent transportability of this approach. The Columbia algorithm was used to classify suicidal adverse events in the recent FDA adult antidepressant safety analyses and

  7. Automated Electroglottographic Inflection Events Detection. A Pilot Study.

    PubMed

    Codino, Juliana; Torres, María Eugenia; Rubin, Adam; Jackson-Menaldi, Cristina

    2016-11-01

    Vocal-fold vibration can be analyzed in a noninvasive way by registering impedance changes within the glottis, through electroglottography. The morphology of the electroglottographic (EGG) signal is related to different vibratory patterns. In the literature, a characteristic knee in the descending portion of the signal has been reported. Some EGG signals do not exhibit this particular knee and have other types of events (inflection events) throughout the ascending and/or descending portion of the vibratory cycle. The goal of this work is to propose an automatic method to identify and classify these events. A computational algorithm was developed based on the mathematical properties of the EGG signal, which detects and reports events throughout the contact phase. Retrospective analysis of EGG signals obtained during routine voice evaluation of adult individuals with a variety of voice disorders was performed using the algorithm as well as human raters. Two judges, both experts in clinical voice analysis, and three general speech pathologists performed manual and visual evaluation of the sample set. The results obtained by the automatic method were compared with those of the human raters. Statistical analysis revealed a significant level of agreement. This automatic tool could allow professionals in the clinical setting to obtain an automatic quantitative and qualitative report of such events present in a voice sample, without having to manually analyze the whole EGG signal. In addition, it might provide the speech pathologist with more information that would complement the standard voice evaluation. It could also be a valuable tool in voice research. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  8. Behaviorism, Private Events, and the Molar View of Behavior

    PubMed Central

    Baum, William M

    2011-01-01

    Viewing the science of behavior (behavior analysis) to be a natural science, radical behaviorism rejects any form of dualism, including subjective–objective or inner–outer dualism. Yet radical behaviorists often claim that treating private events as covert behavior and internal stimuli is necessary and important to behavior analysis. To the contrary, this paper argues that, compared with the rejection of dualism, private events constitute a trivial idea and are irrelevant to accounts of behavior. Viewed in the framework of evolutionary theory or for any practical purpose, behavior is commerce with the environment. By its very nature, behavior is extended in time. The temptation to posit private events arises when an activity is viewed in too small a time frame, obscuring what the activity does. When activities are viewed in an appropriately extended time frame, private events become irrelevant to the account. This insight provides the answer to many philosophical questions about thinking, sensing, and feeling. Confusion about private events arises in large part from failure to appreciate fully the radical implications of replacing mentalistic ideas about language with the concept of verbal behavior. Like other operant behavior, verbal behavior involves no agent and no hidden causes; like all natural events, it is caused by other natural events. In a science of behavior grounded in evolutionary theory, the same set of principles applies to verbal and nonverbal behavior and to human and nonhuman organisms. PMID:22532740

  9. Dietary Supplement Adverse Event Report Data From the FDA Center for Food Safety and Applied Nutrition Adverse Event Reporting System (CAERS), 2004-2013.

    PubMed

    Timbo, Babgaleh B; Chirtel, Stuart J; Ihrie, John; Oladipo, Taiye; Velez-Suarez, Loy; Brewer, Vickery; Mozersky, Robert

    2018-05-01

    The Food and Drug Administration (FDA)'s Center for Food Safety and Applied Nutrition (CFSAN) oversees the safety of the nation's foods, dietary supplements, and cosmetic products. To present a descriptive analysis of the 2004-2013 dietary supplement adverse event report (AER) data from CAERS and evaluate the 2006 Dietary Supplements and Nonprescription Drug Consumer Protection Act as pertaining to dietary supplements adverse events reporting. We queried CAERS for data from the 2004-2013 AERs specifying at least 1 suspected dietary supplement product. We extracted the product name(s), the symptom(s) reported, age, sex, and serious adverse event outcomes. We examined time trends for mandatory and voluntary reporting and performed analysis using SAS v9.4 and R v3.3.0 software. Of the total AERs (n = 15 430) received from January 1, 2004, through December 31, 2013, indicating at least 1 suspected dietary supplement product, 66.9% were mandatory, 32.2% were voluntary, and 0.9% were both mandatory and voluntary. Reported serious outcomes included death, life-threatening conditions, hospitalizations, congenital anomalies/birth defects and events requiring interventions to prevent permanent impairments (5.1%). The dietary supplement adverse event reporting rate in the United States was estimated at ~2% based on CAERS data. This study characterizes CAERS dietary supplement adverse event data for the 2004-2013 period and estimates a reporting rate of 2% for dietary supplement adverse events based on CAERS data. The findings show that the 2006 Dietary Supplements and Nonprescription Drug Consumer Protection Act had a substantial impact on the reporting of adverse events.

  10. Exercise-induced hypertension, cardiovascular events, and mortality in patients undergoing exercise stress testing: a systematic review and meta-analysis.

    PubMed

    Schultz, Martin G; Otahal, Petr; Cleland, Verity J; Blizzard, Leigh; Marwick, Thomas H; Sharman, James E

    2013-03-01

    The prognostic relevance of a hypertensive response to exercise (HRE) is ill-defined in individuals undergoing exercise stress testing. The study described here was intended to provide a systematic review and meta-analysis of published literature to determine the value of exercise-related blood pressure (BP) (independent of office BP) for predicting cardiovascular (CV) events and mortality. Online databases were searched for published longitudinal studies reporting exercise-related BP and CV events and mortality rates. We identified for review 12 longitudinal studies with a total of 46,314 individuals without significant coronary artery disease, with total CV event and mortality rates recorded over a mean follow-up of 15.2±4.0 years. After adjustment for age, office BP, and CV risk factors, an HRE at moderate exercise intensity carried a 36% greater rate of CV events and mortality (95% CI, 1.02-1.83, P = 0.039) than that of subjects without an HRE. Additionally, each 10mm Hg increase in systolic BP during exercise at moderate intensity was accompanied by a 4% increase in CV events and mortality, independent of office BP, age, or CV risk factors (95% CI, 1.01-1.07, P = 0.02). Systolic BP at maximal workload was not significantly associated with the outcome of an increased rate of CV, whether analyzed as a categorical (HR=1.49, 95% CI, 0.90-2.46, P = 0.12) or a continuous (HR=1.01, 95% CI, 0.98-1.04, P = 0.53) variable. An HRE at moderate exercise intensity during exercise stress testing is an independent risk factor for CV events and mortality. This highlights the need to determine underlying pathophysiological mechanisms of exercise-induced hypertension.

  11. The Monitoring Erosion of Agricultural Land and spatial database of erosion events

    NASA Astrophysics Data System (ADS)

    Kapicka, Jiri; Zizala, Daniel

    2013-04-01

    In 2011 originated in The Czech Republic The Monitoring Erosion of Agricultural Land as joint project of State Land Office (SLO) and Research Institute for Soil and Water Conservation (RISWC). The aim of the project is collecting and record keeping information about erosion events on agricultural land and their evaluation. The main idea is a creation of a spatial database that will be source of data and information for evaluation and modeling erosion process, for proposal of preventive measures and measures to reduce negative impacts of erosion events. A subject of monitoring is the manifestations of water erosion, wind erosion and slope deformation in which cause damaged agriculture land. A website, available on http://me.vumop.cz, is used as a tool for keeping and browsing information about monitored events. SLO employees carry out record keeping. RISWC is specialist institute in the Monitoring Erosion of Agricultural Land that performs keeping the spatial database, running the website, managing the record keeping of events, analysis the cause of origins events and statistical evaluations of keeping events and proposed measures. Records are inserted into the database using the user interface of the website which has map server as a component. Website is based on database technology PostgreSQL with superstructure PostGIS and MapServer UMN. Each record is in the database spatial localized by a drawing and it contains description information about character of event (data, situation description etc.) then there are recorded information about land cover and about grown crops. A part of database is photodocumentation which is taken in field reconnaissance which is performed within two days after notify of event. Another part of database are information about precipitations from accessible precipitation gauges. Website allows to do simple spatial analysis as are area calculation, slope calculation, percentage representation of GAEC etc.. Database structure was designed

  12. Analysis of adverse events of renal impairment related to platinum-based compounds using the Japanese Adverse Drug Event Report database.

    PubMed

    Naganuma, Misa; Motooka, Yumi; Sasaoka, Sayaka; Hatahira, Haruna; Hasegawa, Shiori; Fukuda, Akiho; Nakao, Satoshi; Shimada, Kazuyo; Hirade, Koseki; Mori, Takayuki; Yoshimura, Tomoaki; Kato, Takeshi; Nakamura, Mitsuhiro

    2018-01-01

    Platinum compounds cause several adverse events, such as nephrotoxicity, gastrointestinal toxicity, myelosuppression, ototoxicity, and neurotoxicity. We evaluated the incidence of renal impairment as adverse events are related to the administration of platinum compounds using the Japanese Adverse Drug Event Report database. We analyzed adverse events associated with the use of platinum compounds reported from April 2004 to November 2016. The reporting odds ratio at 95% confidence interval was used to detect the signal for each renal impairment incidence. We evaluated the time-to-onset profile of renal impairment and assessed the hazard type using Weibull shape parameter and used the applied association rule mining technique to discover undetected relationships such as possible risk factor. In total, 430,587 reports in the Japanese Adverse Drug Event Report database were analyzed. The reporting odds ratios (95% confidence interval) for renal impairment resulting from the use of cisplatin, oxaliplatin, carboplatin, and nedaplatin were 2.7 (2.5-3.0), 0.6 (0.5-0.7), 0.8 (0.7-1.0), and 1.3 (0.8-2.1), respectively. The lower limit of the reporting odds ratio (95% confidence interval) for cisplatin was >1. The median (lower-upper quartile) onset time of renal impairment following the use of platinum-based compounds was 6.0-8.0 days. The Weibull shape parameter β and 95% confidence interval upper limit of oxaliplatin were <1. In the association rule mining, the score of lift for patients who were treated with cisplatin and co-administered furosemide, loxoprofen, or pemetrexed was high. Similarly, the scores for patients with hypertension or diabetes mellitus were high. Our findings suggest a potential risk of renal impairment during cisplatin use in real-world setting. The present findings demonstrate that the incidence of renal impairment following cisplatin use should be closely monitored when patients are hypertensive or diabetic, or when they are co

  13. Binary Microlensing Events from the MACHO Project

    NASA Astrophysics Data System (ADS)

    Alcock, C.; Allsman, R. A.; Alves, D.; Axelrod, T. S.; Baines, D.; Becker, A. C.; Bennett, D. P.; Bourke, A.; Brakel, A.; Cook, K. H.; Crook, B.; Crouch, A.; Dan, J.; Drake, A. J.; Fragile, P. C.; Freeman, K. C.; Gal-Yam, A.; Geha, M.; Gray, J.; Griest, K.; Gurtierrez, A.; Heller, A.; Howard, J.; Johnson, B. R.; Kaspi, S.; Keane, M.; Kovo, O.; Leach, C.; Leach, T.; Leibowitz, E. M.; Lehner, M. J.; Lipkin, Y.; Maoz, D.; Marshall, S. L.; McDowell, D.; McKeown, S.; Mendelson, H.; Messenger, B.; Minniti, D.; Nelson, C.; Peterson, B. A.; Popowski, P.; Pozza, E.; Purcell, P.; Pratt, M. R.; Quinn, J.; Quinn, P. J.; Rhie, S. H.; Rodgers, A. W.; Salmon, A.; Shemmer, O.; Stetson, P.; Stubbs, C. W.; Sutherland, W.; Thomson, S.; Tomaney, A.; Vandehei, T.; Walker, A.; Ward, K.; Wyper, G.

    2000-09-01

    We present the light curves of 21 gravitational microlensing events from the first six years of the MACHO Project gravitational microlensing survey that are likely examples of lensing by binary systems. These events were manually selected from a total sample of ~350 candidate microlensing events that were either detected by the MACHO Alert System or discovered through retrospective analyses of the MACHO database. At least 14 of these 21 events exhibit strong (caustic) features, and four of the events are well fit with lensing by large mass ratio (brown dwarf or planetary) systems, although these fits are not necessarily unique. The total binary event rate is roughly consistent with predictions based upon our knowledge of the properties of binary stars, but a precise comparison cannot be made without a determination of our binary lens event detection efficiency. Toward the Galactic bulge, we find a ratio of caustic crossing to noncaustic crossing binary lensing events of 12:4, excluding one event for which we present two fits. This suggests significant incompleteness in our ability to detect and characterize noncaustic crossing binary lensing. The distribution of mass ratios, N(q), for these binary lenses appears relatively flat. We are also able to reliably measure source-face crossing times in four of the bulge caustic crossing events, and recover from them a distribution of lens proper motions, masses, and distances consistent with a population of Galactic bulge lenses at a distance of 7+/-1 kpc. This analysis yields two systems with companions of ~0.05 Msolar.

  14. Experimental Characterization of a Grid-Loss Event on a 2.5-MW Dynamometer Using Advanced Operational Modal Analysis: Preprint

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Helsen, J.; Weijtjens, W.; Guo, Y.

    2015-02-01

    This paper experimentally investigates a worst case grid loss event conducted on the National Renewable Energy Laboratory (NREL) Gearbox Reliability Collaborative (GRC) drivetrain mounted on the 2.5MW NREL dynamic nacelle test-rig. The GRC drivetrain has a directly grid-coupled, fixed speed asynchronous generator. The main goal is the assessment of the dynamic content driving this particular assess the dynamic content of the high-speed stage of the GRC gearbox. In addition to external accelerometers, high frequency sampled measurements of strain gauges were used to assess torque fluctuations and bending moments both at the nacelle main shaft and gearbox high-speed shaft (HSS) throughmore » the entire duration of the event. Modal analysis was conducted using a polyreference Least Squares Complex Frequency-domain (pLSCF) modal identification estimator. The event driving the torsional resonance was identified. Moreover, the pLSCF estimator identified main drivetrain resonances based on a combination of acceleration and strain measurements. Without external action during the grid-loss event, a mode shape characterized by counter phase rotation of the rotor and generator rotor determined by the drivetrain flexibility and rotor inertias was the main driver of the event. This behavior resulted in significant torque oscillations with large amplitude negative torque periods. Based on tooth strain measurements of the HSS pinion, this work showed that at each zero-crossing, the teeth lost contact and came into contact with the backside flank. In addition, dynamic nontorque loads between the gearbox and generator at the HSS played an important role, as indicated by strain gauge-measurements.« less

  15. Tool for Constructing Data Albums for Significant Weather Events

    NASA Astrophysics Data System (ADS)

    Kulkarni, A.; Ramachandran, R.; Conover, H.; McEniry, M.; Goodman, H.; Zavodsky, B. T.; Braun, S. A.; Wilson, B. D.

    2012-12-01

    Case study analysis and climatology studies are common approaches used in Atmospheric Science research. Research based on case studies involves a detailed description of specific weather events using data from different sources, to characterize physical processes in play for a given event. Climatology-based research tends to focus on the representativeness of a given event, by studying the characteristics and distribution of a large number of events. To gather relevant data and information for case studies and climatology analysis is tedious and time consuming; current Earth Science data systems are not suited to assemble multi-instrument, multi mission datasets around specific events. For example, in hurricane science, finding airborne or satellite data relevant to a given storm requires searching through web pages and data archives. Background information related to damages, deaths, and injuries requires extensive online searches for news reports and official storm summaries. We will present a knowledge synthesis engine to create curated "Data Albums" to support case study analysis and climatology studies. The technological challenges in building such a reusable and scalable knowledge synthesis engine are several. First, how to encode domain knowledge in a machine usable form? This knowledge must capture what information and data resources are relevant and the semantic relationships between the various fragments of information and data. Second, how to extract semantic information from various heterogeneous sources including unstructured texts using the encoded knowledge? Finally, how to design a structured database from the encoded knowledge to store all information and to support querying? The structured database must allow both knowledge overviews of an event as well as drill down capability needed for detailed analysis. An application ontology driven framework is being used to design the knowledge synthesis engine. The knowledge synthesis engine is being

  16. Top-down and bottom-up definitions of human failure events in human reliability analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Boring, Ronald Laurids

    2014-10-01

    In the probabilistic risk assessments (PRAs) used in the nuclear industry, human failure events (HFEs) are determined as a subset of hardware failures, namely those hardware failures that could be triggered by human action or inaction. This approach is top-down, starting with hardware faults and deducing human contributions to those faults. Elsewhere, more traditionally human factors driven approaches would tend to look at opportunities for human errors first in a task analysis and then identify which of those errors is risk significant. The intersection of top-down and bottom-up approaches to defining HFEs has not been carefully studied. Ideally, both approachesmore » should arrive at the same set of HFEs. This question is crucial, however, as human reliability analysis (HRA) methods are generalized to new domains like oil and gas. The HFEs used in nuclear PRAs tend to be top-down—defined as a subset of the PRA—whereas the HFEs used in petroleum quantitative risk assessments (QRAs) often tend to be bottom-up—derived from a task analysis conducted by human factors experts. The marriage of these approaches is necessary in order to ensure that HRA methods developed for top-down HFEs are also sufficient for bottom-up applications.« less

  17. HIGH-PRECISION BIOLOGICAL EVENT EXTRACTION: EFFECTS OF SYSTEM AND OF DATA

    PubMed Central

    Cohen, K. Bretonnel; Verspoor, Karin; Johnson, Helen L.; Roeder, Chris; Ogren, Philip V.; Baumgartner, William A.; White, Elizabeth; Tipney, Hannah; Hunter, Lawrence

    2013-01-01

    We approached the problems of event detection, argument identification, and negation and speculation detection in the BioNLP’09 information extraction challenge through concept recognition and analysis. Our methodology involved using the OpenDMAP semantic parser with manually written rules. The original OpenDMAP system was updated for this challenge with a broad ontology defined for the events of interest, new linguistic patterns for those events, and specialized coordination handling. We achieved state-of-the-art precision for two of the three tasks, scoring the highest of 24 teams at precision of 71.81 on Task 1 and the highest of 6 teams at precision of 70.97 on Task 2. We provide a detailed analysis of the training data and show that a number of trigger words were ambiguous as to event type, even when their arguments are constrained by semantic class. The data is also shown to have a number of missing annotations. Analysis of a sampling of the comparatively small number of false positives returned by our system shows that major causes of this type of error were failing to recognize second themes in two-theme events, failing to recognize events when they were the arguments to other events, failure to recognize nontheme arguments, and sentence segmentation errors. We show that specifically handling coordination had a small but important impact on the overall performance of the system. The OpenDMAP system and the rule set are available at http://bionlp.sourceforge.net. PMID:25937701

  18. Creating Reality: How TV News Distorts Events.

    ERIC Educational Resources Information Center

    Altheide, David L.

    A three-year research project, including more than one year in a network affiliate station, provided the material for an analysis of current practices in television news programming. Based on the thesis that the organization of news encourages the oversimplification of events, this analysis traces the foundation of the bias called the "news…

  19. The problem of extreme events in paired-watershed studies

    Treesearch

    James W. Hornbeck

    1973-01-01

    In paired-watershed studies, the occurrence of an extreme event during the after-treatment period presents a problem: the effects of treatment must be determined by using greatly extrapolated regression statistics. Several steps are presented to help insure careful handling of extreme events during analysis and reporting of research results.

  20. Local-scale analysis of temperature patterns over Poland during heatwave events

    NASA Astrophysics Data System (ADS)

    Krzyżewska, Agnieszka; Dyer, Jamie

    2018-01-01

    Heatwaves are predicted to increase in frequency, duration, and severity in the future, including over Central Europe where populations are sensitive to extreme temperature. This paper studies six recent major heatwave events over Poland from 2006 through 2015 using regional-scale simulations (10-km grid spacing, hourly frequency) from the Weather Research and Forecast (WRF) model to define local-scale 2-m temperature patterns. For this purpose, a heatwave is defined as at least three consecutive days with maximum 2-m air temperature exceeding 30 °C. The WRF simulations were validated using maximum daily 2-m temperature observations from 12 meteorological stations in select Polish cities, which were selected to have even spatial coverage across the study area. Synoptic analysis of the six study events shows that the inflow of tropical air masses from the south is the primary driver of heatwave onset and maintenance, the highest temperatures (and most vulnerable areas) occur over arable land and artificial surfaces in central and western Poland, while coastal areas in the north, mountain areas in the south, and forested and mosaic areas of smaller fields and pastures of the northwest, northeast, and southeast are less affected by prolonged periods of elevated temperatures. In general, regional differences in 2-m temperature between the hottest and coolest areas is about 2-4 °C. Large urban areas like Warsaw, or the large complex of artificial areas in the conurbation of Silesian cities, are also generally warmer than surrounding areas by roughly 2-4 °C, and even up to 6 °C, especially during the night. Additionally, hot air from the south of Poland flows through a low-lying area between two mountain ranges (Sudetes and Carpathian Mountains)—the so-called Moravian Gate—hitting densely populated urban areas (Silesian cities) and Cracow. These patterns occur only during high-pressure synoptic conditions with low cloudiness and wind and without any active fronts

  1. Measuring the impact of major life events upon happiness.

    PubMed

    Ballas, Dimitris; Dorling, Danny

    2007-12-01

    In recent years there have been numerous attempts to define and measure happiness in various contexts and pertaining to a wide range of disciplines, ranging from neuroscience and psychology to philosophy, economics and social policy. This article builds on recent work by economists who attempt to estimate happiness regressions using large random samples of individuals in order to calculate monetary 'compensating amounts' for different life 'events'. We estimate happiness regressions using the 'major life event' and 'happiness' data from the British Household Panel Survey. The data and methods used in this article suggest that in contrast to living states such as 'being married', it is more events such as 'starting a new relationship' that have the highest positive effect on happiness. This is closely followed by 'employment-related gains' (in contrast to employment status). Also, women who become pregnant on average report higher than average levels of subjective happiness (in contrast to 'being a parent'). Other events that appear to be associated with happiness according to our analysis include 'personal education-related events' (e.g. starting a new course, graduating from University, passing exams) and 'finance/house related events' (e.g. buying a new house). On the other hand, the event that has the highest negative impact upon happiness according to our analysis is 'the end of my relationship' closely followed by 'death of a parent'. Adverse health events pertaining to the parents of the respondents also have a high negative coefficient and so does an employment-related loss. The analysis presented in this article suggests that what matters the most in people's lives in Britain is to have good dynamic interpersonal relationships and to be respected at work with that respect being constantly renewed. These 'goods' are as much reflected through dynamic events as static situations. Relationships at work appear to be of a similar order of importance to those at

  2. Changes in "hotter and wetter" events across China

    NASA Astrophysics Data System (ADS)

    Liu, C.; Deng, H.; Lu, Y.; Qiu, X.; Wang, D.

    2017-12-01

    As global warming intensifies, efforts to understand the changes in extreme climate events have increased in recent years. A combined analysis of the changes in extreme temperature and precipitation events is presented in this paper. Using observational data from 1961 to 2015, a set of hotter and wetter (HW) events is defined, and we examine the changes in these events across China. The results show that more HW events occur in Central and Eastern China than in other subregions, especially in South China (SC). The rate of increase in HW events is 2.7 and 1.9 per decade in SC and East China (EC), respectively. In China, most HW events occurred in the last 20 years of the study period, indicating that China entered a period of high-frequency HW events. Indeed, the range in anomalies in the torrential rain days is greater than that of the high-temperature days in Northwest China (NWC), Central China (CC), and EC after the mid- to late 1990s. The opposite pattern is found in Northeast China (NEC), Southwest China-region 1 (SWC1), Southwest China-region 2 (SWC2), and SC. Finally, the increase in HW events in most regions of China is closely associated with warming.

  3. Big Data Tools as Applied to ATLAS Event Data

    NASA Astrophysics Data System (ADS)

    Vukotic, I.; Gardner, R. W.; Bryant, L. A.

    2017-10-01

    Big Data technologies have proven to be very useful for storage, processing and visualization of derived metrics associated with ATLAS distributed computing (ADC) services. Logfiles, database records, and metadata from a diversity of systems have been aggregated and indexed to create an analytics platform for ATLAS ADC operations analysis. Dashboards, wide area data access cost metrics, user analysis patterns, and resource utilization efficiency charts are produced flexibly through queries against a powerful analytics cluster. Here we explore whether these techniques and associated analytics ecosystem can be applied to add new modes of open, quick, and pervasive access to ATLAS event data. Such modes would simplify access and broaden the reach of ATLAS public data to new communities of users. An ability to efficiently store, filter, search and deliver ATLAS data at the event and/or sub-event level in a widely supported format would enable or significantly simplify usage of machine learning environments and tools like Spark, Jupyter, R, SciPy, Caffe, TensorFlow, etc. Machine learning challenges such as the Higgs Boson Machine Learning Challenge, the Tracking challenge, Event viewers (VP1, ATLANTIS, ATLASrift), and still to be developed educational and outreach tools would be able to access the data through a simple REST API. In this preliminary investigation we focus on derived xAOD data sets. These are much smaller than the primary xAODs having containers, variables, and events of interest to a particular analysis. Being encouraged with the performance of Elasticsearch for the ADC analytics platform, we developed an algorithm for indexing derived xAOD event data. We have made an appropriate document mapping and have imported a full set of standard model W/Z datasets. We compare the disk space efficiency of this approach to that of standard ROOT files, the performance in simple cut flow type of data analysis, and will present preliminary results on its scaling

  4. Using Wavelet Analysis To Assist in Identification of Significant Events in Molecular Dynamics Simulations.

    PubMed

    Heidari, Zahra; Roe, Daniel R; Galindo-Murillo, Rodrigo; Ghasemi, Jahan B; Cheatham, Thomas E

    2016-07-25

    Long time scale molecular dynamics (MD) simulations of biological systems are becoming increasingly commonplace due to the availability of both large-scale computational resources and significant advances in the underlying simulation methodologies. Therefore, it is useful to investigate and develop data mining and analysis techniques to quickly and efficiently extract the biologically relevant information from the incredible amount of generated data. Wavelet analysis (WA) is a technique that can quickly reveal significant motions during an MD simulation. Here, the application of WA on well-converged long time scale (tens of μs) simulations of a DNA helix is described. We show how WA combined with a simple clustering method can be used to identify both the physical and temporal locations of events with significant motion in MD trajectories. We also show that WA can not only distinguish and quantify the locations and time scales of significant motions, but by changing the maximum time scale of WA a more complete characterization of these motions can be obtained. This allows motions of different time scales to be identified or ignored as desired.

  5. Analysis of low-frequency seismic signals generated during a multiple-iceberg calving event at Jakobshavn Isbræ, Greenland

    USGS Publications Warehouse

    Walter, Fabian; Amundson, Jason M.; O'Neel, Shad; Truffer, Martin; Fahnestock, Mark; Fricker, Helen A.

    2012-01-01

    We investigated seismic signals generated during a large-scale, multiple iceberg calving event that occurred at Jakobshavn Isbræ, Greenland, on 21 August 2009. The event was recorded by a high-rate time-lapse camera and five broadband seismic stations located within a few hundred kilometers of the terminus. During the event two full-glacier-thickness icebergs calved from the grounded (or nearly grounded) terminus and immediately capsized; the second iceberg to calve was two to three times smaller than the first. The individual calving and capsize events were well-correlated with the radiation of low-frequency seismic signals (<0.1 Hz) dominated by Love and Rayleigh waves. In agreement with regional records from previously published ‘glacial earthquakes’, these low-frequency seismic signals had maximum power and/or signal-to-noise ratios in the 0.05–0.1 Hz band. Similarly, full waveform inversions indicate that these signals were also generated by horizontal single forces acting at the glacier terminus. The signals therefore appear to be local manifestations of glacial earthquakes, although the magnitudes of the signals (twice-time integrated force histories) were considerably smaller than previously reported glacial earthquakes. We thus speculate that such earthquakes may be a common, if not pervasive, feature of all full-glacier-thickness calving events from grounded termini. Finally, a key result from our study is that waveform inversions performed on low-frequency, calving-generated seismic signals may have only limited ability to quantitatively estimate mass losses from calving. In particular, the choice of source time function has little impact on the inversion but dramatically changes the earthquake magnitude. Accordingly, in our analysis, it is unclear whether the smaller or larger of the two calving icebergs generated a larger seismic signal.

  6. Retrospective Analysis of Communication Events - Understanding the Dynamics of Collaborative Multi-Party Discourse

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cowell, Andrew J.; Haack, Jereme N.; McColgin, Dave W.

    2006-06-08

    This research is aimed at understanding the dynamics of collaborative multi-party discourse across multiple communication modalities. Before we can truly make sig-nificant strides in devising collaborative communication systems, there is a need to understand how typical users utilize com-putationally supported communications mechanisms such as email, instant mes-saging, video conferencing, chat rooms, etc., both singularly and in conjunction with traditional means of communication such as face-to-face meetings, telephone calls and postal mail. Attempting to un-derstand an individual’s communications profile with access to only a single modal-ity is challenging at best and often futile. Here, we discuss the development of RACE –more » Retrospective Analysis of Com-munications Events – a test-bed prototype to investigate issues relating to multi-modal multi-party discourse.« less

  7. Joint Modeling Approach for Semicompeting Risks Data with Missing Nonterminal Event Status

    PubMed Central

    Hu, Chen; Tsodikov, Alex

    2014-01-01

    Semicompeting risks data, where a subject may experience sequential non-terminal and terminal events, and the terminal event may censor the non-terminal event but not vice versa, are widely available in many biomedical studies. We consider the situation when a proportion of subjects’ non-terminal events is missing, such that the observed data become a mixture of “true” semicompeting risks data and partially observed terminal event only data. An illness-death multistate model with proportional hazards assumptions is proposed to study the relationship between non-terminal and terminal events, and provide covariate-specific global and local association measures. Maximum likelihood estimation based on semiparametric regression analysis is used for statistical inference, and asymptotic properties of proposed estimators are studied using empirical process and martingale arguments. We illustrate the proposed method with simulation studies and data analysis of a follicular cell lymphoma study. PMID:24430204

  8. Geohazard assessment through the analysis of historical alluvial events in Southern Italy

    NASA Astrophysics Data System (ADS)

    Esposito, Eliana; Violante, Crescenzo

    2015-04-01

    The risk associated with extreme water events such as flash floods, results from a combination of overflows and landslides hazards. A multi-hazard approach have been utilized to analyze the 1773 flood that occurred in conjunction with heavy rainfall, causing major damage in terms of lost lives and economic cost over an area of 200 km2, including both the coastal strip between Salerno and Maiori and the Apennine hinterland, Campania region - Southern Italy. This area has been affected by a total of 40 flood events over the last five centuries, 26 of them occurred between 1900 and 2000. Streamflow events have produced severe impacts on Cava de' Tirreni (SA) and its territory and in particular four catastrophic floods in 1581, 1773, 1899 and 1954, caused a pervasive pattern of destruction. In the study area, rainstorm events typically occur in small and medium-sized fluvial system, characterized by small catchment areas and high-elevation drainage basins, causing the detachment of large amount of volcaniclastic and siliciclastic covers from the carbonate bedrock. The mobilization of these deposits (slope debris) mixed with rising floodwaters along the water paths can produce fast-moving streamflows of large proportion with significant hazardous implications (Violante et al., 2009). In this context the study of 1773 historical flood allows the detection and the definition of those areas where catastrophic events repeatedly took place over the time. Moreover, it improves the understanding of the phenomena themselves, including some key elements in the management of risk mitigation, such as the restoration of the damage suffered by the buildings and/or the environmental effects caused by the floods.

  9. Transient Volcano Deformation Event Detection over Variable Spatial Scales in Alaska

    NASA Astrophysics Data System (ADS)

    Li, J. D.; Rude, C. M.; Gowanlock, M.; Herring, T.; Pankratius, V.

    2016-12-01

    Transient deformation events driven by volcanic activity can be monitored using increasingly dense networks of continuous Global Positioning System (GPS) ground stations. The wide spatial extent of GPS networks, the large number of GPS stations, and the spatially and temporally varying scale of deformation events result in the mixing of signals from multiple sources. Typical analysis then necessitates manual identification of times and regions of volcanic activity for further study and the careful tuning of algorithmic parameters to extract possible transient events. Here we present a computer-aided discovery system that facilitates the discovery of potential transient deformation events at volcanoes by providing a framework for selecting varying spatial regions of interest and for tuning the analysis parameters. This site specification step in the framework reduces the spatial mixing of signals from different volcanic sources before applying filters to remove interfering signals originating from other geophysical processes. We analyze GPS data recorded by the Plate Boundary Observatory network and volcanic activity logs from the Alaska Volcano Observatory to search for and characterize transient inflation events in Alaska. We find 3 transient inflation events between 2008 and 2015 at the Akutan, Westdahl, and Shishaldin volcanoes in the Aleutian Islands. The inflation event detected in the first half of 2008 at Akutan is validated other studies, while the inflation events observed in early 2011 at Westdahl and in early 2013 at Shishaldin are previously unreported. Our analysis framework also incorporates modelling of the transient inflation events and enables a comparison of different magma chamber inversion models. Here, we also estimate the magma sources that best describe the deformation observed by the GPS stations at Akutan, Westdahl, and Shishaldin. We acknowledge support from NASA AIST-NNX15AG84G (PI: V. Pankratius).

  10. Extreme events in total ozone over the Northern mid-latitudes: an analysis based on long-term data sets from five European ground-based stations

    NASA Astrophysics Data System (ADS)

    Rieder, Harald E.; Jancso, Leonhardt M.; Rocco, Stefania Di; Staehelin, Johannes; Maeder, Joerg A.; Peter, Thomas; Ribatet, Mathieu; Davison, Anthony C.; de Backer, Hugo; Koehler, Ulf; Krzyścin, Janusz; Vaníček, Karel

    2011-11-01

    We apply methods from extreme value theory to identify extreme events in high (termed EHOs) and low (termed ELOs) total ozone and to describe the distribution tails (i.e. very high and very low values) of five long-term European ground-based total ozone time series. The influence of these extreme events on observed mean values, long-term trends and changes is analysed. The results show a decrease in EHOs and an increase in ELOs during the last decades, and establish that the observed downward trend in column ozone during the 1970-1990s is strongly dominated by changes in the frequency of extreme events. Furthermore, it is shown that clear ‘fingerprints’ of atmospheric dynamics (NAO, ENSO) and chemistry [ozone depleting substances (ODSs), polar vortex ozone loss] can be found in the frequency distribution of ozone extremes, even if no attribution is possible from standard metrics (e.g. annual mean values). The analysis complements earlier analysis for the world's longest total ozone record at Arosa, Switzerland, confirming and revealing the strong influence of atmospheric dynamics on observed ozone changes. The results provide clear evidence that in addition to ODS, volcanic eruptions and strong/moderate ENSO and NAO events had significant influence on column ozone in the European sector.

  11. Event Display for the Visualization of CMS Events

    NASA Astrophysics Data System (ADS)

    Bauerdick, L. A. T.; Eulisse, G.; Jones, C. D.; Kovalskyi, D.; McCauley, T.; Mrak Tadel, A.; Muelmenstaedt, J.; Osborne, I.; Tadel, M.; Tu, Y.; Yagil, A.

    2011-12-01

    During the last year the CMS experiment engaged in consolidation of its existing event display programs. The core of the new system is based on the Fireworks event display program which was by-design directly integrated with the CMS Event Data Model (EDM) and the light version of the software framework (FWLite). The Event Visualization Environment (EVE) of the ROOT framework is used to manage a consistent set of 3D and 2D views, selection, user-feedback and user-interaction with the graphics windows; several EVE components were developed by CMS in collaboration with the ROOT project. In event display operation simple plugins are registered into the system to perform conversion from EDM collections into their visual representations which are then managed by the application. Full event navigation and filtering as well as collection-level filtering is supported. The same data-extraction principle can also be applied when Fireworks will eventually operate as a service within the full software framework.

  12. Metagenomic Analysis of Airborne Bacterial Community and Diversity in Seoul, Korea, during December 2014, Asian Dust Event.

    PubMed

    Cha, Seho; Srinivasan, Sathiyaraj; Jang, Jun Hyeong; Lee, Dongwook; Lim, Sora; Kim, Kyung Sang; Jheong, Weonhwa; Lee, Dong-Won; Park, Eung-Roh; Chung, Hyun-Mi; Choe, Joonho; Kim, Myung Kyum; Seo, Taegun

    2017-01-01

    Asian dust or yellow sand events in East Asia are a major issue of environmental contamination and human health, causing increasing concern. A high amount of dust particles, especially called as particulate matter 10 (PM10), is transported by the wind from the arid and semi-arid tracks to the Korean peninsula, bringing a bacterial population that alters the terrestrial and atmospheric microbial communities. In this study, we aimed to explore the bacterial populations of Asian dust samples collected during November-December 2014. The dust samples were collected using the impinger method, and the hypervariable regions of the 16S rRNA gene were amplified using PCR followed by pyrosequencing. Analysis of the sequencing data were performed using Mothur software. The data showed that the number of operational taxonomic units and diversity index during Asian dust events were higher than those during non-Asian dust events. At the phylum level, the proportions of Proteobacteria, Actinobacteria, and Firmicutes were different between Asian dust and non-Asian dust samples. At the genus level, the proportions of the genus Bacillus (6.9%), Arthrobacter (3.6%), Blastocatella (2%), Planomicrobium (1.4%) were increased during Asian dust compared to those in non-Asian dust samples. This study showed that the significant relationship between bacterial populations of Asian dust samples and non-Asian dust samples in Korea, which could significantly affect the microbial population in the environment.

  13. Compilation and Analysis of 20 and 30 GHz Rain Fade Events at the ACTS NASA Ground Station: Statistics and Model Assessment

    NASA Technical Reports Server (NTRS)

    Manning, Robert M.

    1996-01-01

    The purpose of the propagation studies within the ACTS Project Office is to acquire 20 and 30 GHz rain fade statistics using the ACTS beacon links received at the NGS (NASA Ground Station) in Cleveland. Other than the raw, statistically unprocessed rain fade events that occur in real time, relevant rain fade statistics derived from such events are the cumulative rain fade statistics as well as fade duration statistics (beyond given fade thresholds) over monthly and yearly time intervals. Concurrent with the data logging exercise, monthly maximum rainfall levels recorded at the US Weather Service at Hopkins Airport are appended to the database to facilitate comparison of observed fade statistics with those predicted by the ACTS Rain Attenuation Model. Also, the raw fade data will be in a format, complete with documentation, for use by other investigators who require realistic fade event evolution in time for simulation purposes or further analysis for comparisons with other rain fade prediction models, etc. The raw time series data from the 20 and 30 GHz beacon signals is purged of non relevant data intervals where no rain fading has occurred. All other data intervals which contain rain fade events are archived with the accompanying time stamps. The definition of just what constitutes a rain fade event will be discussed later. The archived data serves two purposes. First, all rain fade event data is recombined into a contiguous data series every month and every year; this will represent an uninterrupted record of the actual (i.e., not statistically processed) temporal evolution of rain fade at 20 and 30 GHz at the location of the NGS. The second purpose of the data in such a format is to enable a statistical analysis of prevailing propagation parameters such as cumulative distributions of attenuation on a monthly and yearly basis as well as fade duration probabilities below given fade thresholds, also on a monthly and yearly basis. In addition, various subsidiary

  14. Rapid depressurization event analysis in BWR/6 using RELAP5 and contain

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mueftueoglu, A.K.; Feltus, M.A.

    1995-09-01

    Noncondensable gases may become dissolved in Boiling Water Reactor (BWR) water level instrumentation during normal operations. Any dissolved noncondensable gases inside these water columns may come out of solution during rapid depressurization events, and displace water from the reference leg piping resulting in a false high level. These water level errors may cause a delay or failure in actuation, or premature shutdown of the Emergency Core Cooling System. (ECCS). If a rapid depressurization causes an erroneously high water level, preventing automatic ECCS actuation, it becomes important to determine if there would be other adequate indications for operator response and othermore » signals for automatic actuation such as high drywell pressure. It is also important to determine the effect of the level signal on ECCS operation after it is being actuated. The objective of this study is to determine the detailed coupled containment/NSSS response during this rapid depressurization events in BWR/6. The selected scenarios involve: (a) inadvertent opening of all ADS valves, (b) design basis (DB) large break loss of coolant accident (LOCA), and (c) main steam line break (MSLB). The transient behaviors are evaluated in terms of: (a) vessel pressure and collapsed water level response, (b) specific transient boundary conditions, (e.g., scram, MSIV closure timing, feedwater flow, and break blowdown rates), (c) ECCS initiation timing, (d) impact of operator actions, (e) whether indications besides low-low water level were available. The results of the analysis had shown that there would be signals to actuate ECCS other than low reactor level, such as high drywell pressure, low vessel pressure, high suppression pool temperature, and that the plant operators would have significant indications to actuate ECCS.« less

  15. Integrated survival analysis using an event-time approach in a Bayesian framework

    USGS Publications Warehouse

    Walsh, Daniel P.; Dreitz, VJ; Heisey, Dennis M.

    2015-01-01

    Event-time or continuous-time statistical approaches have been applied throughout the biostatistical literature and have led to numerous scientific advances. However, these techniques have traditionally relied on knowing failure times. This has limited application of these analyses, particularly, within the ecological field where fates of marked animals may be unknown. To address these limitations, we developed an integrated approach within a Bayesian framework to estimate hazard rates in the face of unknown fates. We combine failure/survival times from individuals whose fates are known and times of which are interval-censored with information from those whose fates are unknown, and model the process of detecting animals with unknown fates. This provides the foundation for our integrated model and permits necessary parameter estimation. We provide the Bayesian model, its derivation, and use simulation techniques to investigate the properties and performance of our approach under several scenarios. Lastly, we apply our estimation technique using a piece-wise constant hazard function to investigate the effects of year, age, chick size and sex, sex of the tending adult, and nesting habitat on mortality hazard rates of the endangered mountain plover (Charadrius montanus) chicks. Traditional models were inappropriate for this analysis because fates of some individual chicks were unknown due to failed radio transmitters. Simulations revealed biases of posterior mean estimates were minimal (≤ 4.95%), and posterior distributions behaved as expected with RMSE of the estimates decreasing as sample sizes, detection probability, and survival increased. We determined mortality hazard rates for plover chicks were highest at <5 days old and were lower for chicks with larger birth weights and/or whose nest was within agricultural habitats. Based on its performance, our approach greatly expands the range of problems for which event-time analyses can be used by eliminating the

  16. Integrated survival analysis using an event-time approach in a Bayesian framework.

    PubMed

    Walsh, Daniel P; Dreitz, Victoria J; Heisey, Dennis M

    2015-02-01

    Event-time or continuous-time statistical approaches have been applied throughout the biostatistical literature and have led to numerous scientific advances. However, these techniques have traditionally relied on knowing failure times. This has limited application of these analyses, particularly, within the ecological field where fates of marked animals may be unknown. To address these limitations, we developed an integrated approach within a Bayesian framework to estimate hazard rates in the face of unknown fates. We combine failure/survival times from individuals whose fates are known and times of which are interval-censored with information from those whose fates are unknown, and model the process of detecting animals with unknown fates. This provides the foundation for our integrated model and permits necessary parameter estimation. We provide the Bayesian model, its derivation, and use simulation techniques to investigate the properties and performance of our approach under several scenarios. Lastly, we apply our estimation technique using a piece-wise constant hazard function to investigate the effects of year, age, chick size and sex, sex of the tending adult, and nesting habitat on mortality hazard rates of the endangered mountain plover (Charadrius montanus) chicks. Traditional models were inappropriate for this analysis because fates of some individual chicks were unknown due to failed radio transmitters. Simulations revealed biases of posterior mean estimates were minimal (≤ 4.95%), and posterior distributions behaved as expected with RMSE of the estimates decreasing as sample sizes, detection probability, and survival increased. We determined mortality hazard rates for plover chicks were highest at <5 days old and were lower for chicks with larger birth weights and/or whose nest was within agricultural habitats. Based on its performance, our approach greatly expands the range of problems for which event-time analyses can be used by eliminating the

  17. Accelerated increase in the Arctic tropospheric warming events surpassing stratospheric warming events during winter

    NASA Astrophysics Data System (ADS)

    Wang, S.-Y. Simon; Lin, Yen-Heng; Lee, Ming-Ying; Yoon, Jin-Ho; Meyer, Jonathan D. D.; Rasch, Philip J.

    2017-04-01

    In January 2016, a robust reversal of the Arctic Oscillation took place associated with a rapid tropospheric warming in the Arctic region; this was followed by the occurrence of a classic sudden stratospheric warming in March. The succession of these two distinct Arctic warming events provides a stimulating opportunity to examine their characteristics in terms of similarities and differences. Historical cases of these two types of Arctic warming were identified and validated based upon tropical linkages with the Madden-Julian Oscillation and El Niño as documented in previous studies. The analysis indicates a recent and seemingly accelerated increase in the tropospheric warming type versus a flat trend in stratospheric warming type. The shorter duration and more rapid transition of tropospheric warming events may connect to the documented increase in midlatitude weather extremes, more so than the route of stratospheric warming type. Forced simulations with an atmospheric general circulation model suggest that the reduced Arctic sea ice contributes to the observed increase in the tropospheric warming events and associated remarkable strengthening of the cold Siberian high manifest in 2016.

  18. One Size Does Not Fit All: Human Failure Event Decomposition and Task Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ronald Laurids Boring, PhD

    2014-09-01

    In the probabilistic safety assessments (PSAs) used in the nuclear industry, human failure events (HFEs) are determined as a subset of hardware failures, namely those hardware failures that could be triggered or exacerbated by human action or inaction. This approach is top-down, starting with hardware faults and deducing human contributions to those faults. Elsewhere, more traditionally human factors driven approaches would tend to look at opportunities for human errors first in a task analysis and then identify which of those errors is risk significant. The intersection of top-down and bottom-up approaches to defining HFEs has not been carefully studied. Ideally,more » both approaches should arrive at the same set of HFEs. This question remains central as human reliability analysis (HRA) methods are generalized to new domains like oil and gas. The HFEs used in nuclear PSAs tend to be top-down—defined as a subset of the PSA—whereas the HFEs used in petroleum quantitative risk assessments (QRAs) are more likely to be bottom-up—derived from a task analysis conducted by human factors experts. The marriage of these approaches is necessary in order to ensure that HRA methods developed for top-down HFEs are also sufficient for bottom-up applications. In this paper, I first review top-down and bottom-up approaches for defining HFEs and then present a seven-step guideline to ensure a task analysis completed as part of human error identification decomposes to a level suitable for use as HFEs. This guideline illustrates an effective way to bridge the bottom-up approach with top-down requirements.« less

  19. Evaluation of Visual Field Progression in Glaucoma: Quasar Regression Program and Event Analysis.

    PubMed

    Díaz-Alemán, Valentín T; González-Hernández, Marta; Perera-Sanz, Daniel; Armas-Domínguez, Karintia

    2016-01-01

    To determine the sensitivity, specificity and agreement between the Quasar program, glaucoma progression analysis (GPA II) event analysis and expert opinion in the detection of glaucomatous progression. The Quasar program is based on linear regression analysis of both mean defect (MD) and pattern standard deviation (PSD). Each series of visual fields was evaluated by three methods; Quasar, GPA II and four experts. The sensitivity, specificity and agreement (kappa) for each method was calculated, using expert opinion as the reference standard. The study included 439 SITA Standard visual fields of 56 eyes of 42 patients, with a mean of 7.8 ± 0.8 visual fields per eye. When suspected cases of progression were considered stable, sensitivity and specificity of Quasar, GPA II and the experts were 86.6% and 70.7%, 26.6% and 95.1%, and 86.6% and 92.6% respectively. When suspected cases of progression were considered as progressing, sensitivity and specificity of Quasar, GPA II and the experts were 79.1% and 81.2%, 45.8% and 90.6%, and 85.4% and 90.6% respectively. The agreement between Quasar and GPA II when suspected cases were considered stable or progressing was 0.03 and 0.28 respectively. The degree of agreement between Quasar and the experts when suspected cases were considered stable or progressing was 0.472 and 0.507. The degree of agreement between GPA II and the experts when suspected cases were considered stable or progressing was 0.262 and 0.342. The combination of MD and PSD regression analysis in the Quasar program showed better agreement with the experts and higher sensitivity than GPA II.

  20. Exploring Evolving Media Discourse Through Event Cueing.

    PubMed

    Lu, Yafeng; Steptoe, Michael; Burke, Sarah; Wang, Hong; Tsai, Jiun-Yi; Davulcu, Hasan; Montgomery, Douglas; Corman, Steven R; Maciejewski, Ross

    2016-01-01

    Online news, microblogs and other media documents all contain valuable insight regarding events and responses to events. Underlying these documents is the concept of framing, a process in which communicators act (consciously or unconsciously) to construct a point of view that encourages facts to be interpreted by others in a particular manner. As media discourse evolves, how topics and documents are framed can undergo change, shifting the discussion to different viewpoints or rhetoric. What causes these shifts can be difficult to determine directly; however, by linking secondary datasets and enabling visual exploration, we can enhance the hypothesis generation process. In this paper, we present a visual analytics framework for event cueing using media data. As discourse develops over time, our framework applies a time series intervention model which tests to see if the level of framing is different before or after a given date. If the model indicates that the times before and after are statistically significantly different, this cues an analyst to explore related datasets to help enhance their understanding of what (if any) events may have triggered these changes in discourse. Our framework consists of entity extraction and sentiment analysis as lenses for data exploration and uses two different models for intervention analysis. To demonstrate the usage of our framework, we present a case study on exploring potential relationships between climate change framing and conflicts in Africa.

  1. Security Event Recognition for Visual Surveillance

    NASA Astrophysics Data System (ADS)

    Liao, W.; Yang, C.; Yang, M. Ying; Rosenhahn, B.

    2017-05-01

    With rapidly increasing deployment of surveillance cameras, the reliable methods for automatically analyzing the surveillance video and recognizing special events are demanded by different practical applications. This paper proposes a novel effective framework for security event analysis in surveillance videos. First, convolutional neural network (CNN) framework is used to detect objects of interest in the given videos. Second, the owners of the objects are recognized and monitored in real-time as well. If anyone moves any object, this person will be verified whether he/she is its owner. If not, this event will be further analyzed and distinguished between two different scenes: moving the object away or stealing it. To validate the proposed approach, a new video dataset consisting of various scenarios is constructed for more complex tasks. For comparison purpose, the experiments are also carried out on the benchmark databases related to the task on abandoned luggage detection. The experimental results show that the proposed approach outperforms the state-of-the-art methods and effective in recognizing complex security events.

  2. Event shapes and azimuthal correlations in Z +jets events in pp collisions at √{ s} = 7 TeV

    NASA Astrophysics Data System (ADS)

    Chatrchyan, S.; Khachatryan, V.; Sirunyan, A. M.; Tumasyan, A.; Adam, W.; Aguilo, E.; Bergauer, T.; Dragicevic, M.; Erö, J.; Fabjan, C.; Friedl, M.; Frühwirth, R.; Ghete, V. M.; Hörmann, N.; Hrubec, J.; Jeitler, M.; Kiesenhofer, W.; Knünz, V.; Krammer, M.; Krätschmer, I.; Liko, D.; Mikulec, I.; Pernicka, M.; Rabady, D.; Rahbaran, B.; Rohringer, C.; Rohringer, H.; Schöfbeck, R.; Strauss, J.; Taurok, A.; Waltenberger, W.; Wulz, C.-E.; Mossolov, V.; Shumeiko, N.; Suarez Gonzalez, J.; Bansal, M.; Bansal, S.; Cornelis, T.; De Wolf, E. A.; Janssen, X.; Luyckx, S.; Mucibello, L.; Ochesanu, S.; Roland, B.; Rougny, R.; Selvaggi, M.; Van Haevermaet, H.; Van Mechelen, P.; Van Remortel, N.; Van Spilbeeck, A.; Blekman, F.; Blyweert, S.; D'Hondt, J.; Gonzalez Suarez, R.; Kalogeropoulos, A.; Maes, M.; Olbrechts, A.; Tavernier, S.; Van Doninck, W.; Van Mulders, P.; Van Onsem, G. P.; Villella, I.; Clerbaux, B.; De Lentdecker, G.; Dero, V.; Gay, A. P. R.; Hreus, T.; Léonard, A.; Marage, P. E.; Mohammadi, A.; Reis, T.; Thomas, L.; Vander Velde, C.; Vanlaer, P.; Wang, J.; Adler, V.; Beernaert, K.; Cimmino, A.; Costantini, S.; Garcia, G.; Grunewald, M.; Klein, B.; Lellouch, J.; Marinov, A.; Mccartin, J.; Ocampo Rios, A. A.; Ryckbosch, D.; Sigamani, M.; Strobbe, N.; Thyssen, F.; Tytgat, M.; Walsh, S.; Yazgan, E.; Zaganidis, N.; Basegmez, S.; Bruno, G.; Castello, R.; Ceard, L.; Delaere, C.; du Pree, T.; Favart, D.; Forthomme, L.; Giammanco, A.; Hollar, J.; Lemaitre, V.; Liao, J.; Militaru, O.; Nuttens, C.; Pagano, D.; Pin, A.; Piotrzkowski, K.; Vizan Garcia, J. M.; Beliy, N.; Caebergs, T.; Daubie, E.; Hammad, G. H.; Alves, G. A.; Correa Martins Junior, M.; Martins, T.; Pol, M. E.; Souza, M. H. G.; Aldá Júnior, W. L.; Carvalho, W.; Custódio, A.; Da Costa, E. M.; De Jesus Damiao, D.; De Oliveira Martins, C.; Fonseca De Souza, S.; Malbouisson, H.; Malek, M.; Matos Figueiredo, D.; Mundim, L.; Nogima, H.; Prado Da Silva, W. L.; Santoro, A.; Soares Jorge, L.; Sznajder, A.; Vilela Pereira, A.; Anjos, T. S.; Bernardes, C. A.; Dias, F. A.; Fernandez Perez Tomei, T. R.; Gregores, E. M.; Lagana, C.; Marinho, F.; Mercadante, P. G.; Novaes, S. F.; Padula, Sandra S.; Genchev, V.; Iaydjiev, P.; Piperov, S.; Rodozov, M.; Stoykova, S.; Sultanov, G.; Tcholakov, V.; Trayanov, R.; Vutova, M.; Dimitrov, A.; Hadjiiska, R.; Kozhuharov, V.; Litov, L.; Pavlov, B.; Petkov, P.; Bian, J. G.; Chen, G. M.; Chen, H. S.; Jiang, C. H.; Liang, D.; Liang, S.; Meng, X.; Tao, J.; Wang, J.; Wang, X.; Wang, Z.; Xiao, H.; Xu, M.; Zang, J.; Zhang, Z.; Asawatangtrakuldee, C.; Ban, Y.; Guo, Y.; Li, W.; Liu, S.; Mao, Y.; Qian, S. J.; Teng, H.; Wang, D.; Zhang, L.; Zou, W.; Avila, C.; Carrillo Montoya, C. A.; Gomez, J. P.; Gomez Moreno, B.; Osorio Oliveros, A. F.; Sanabria, J. C.; Godinovic, N.; Lelas, D.; Plestina, R.; Polic, D.; Puljak, I.; Antunovic, Z.; Kovac, M.; Brigljevic, V.; Duric, S.; Kadija, K.; Luetic, J.; Mekterovic, D.; Morovic, S.; Tikvica, L.; Attikis, A.; Galanti, M.; Mavromanolakis, G.; Mousa, J.; Nicolaou, C.; Ptochos, F.; Razis, P. A.; Finger, M.; Finger, M.; Assran, Y.; Elgammal, S.; Ellithi Kamel, A.; Mahmoud, M. A.; Mahrous, A.; Radi, A.; Kadastik, M.; Müntel, M.; Murumaa, M.; Raidal, M.; Rebane, L.; Tiko, A.; Eerola, P.; Fedi, G.; Voutilainen, M.; Härkönen, J.; Heikkinen, A.; Karimäki, V.; Kinnunen, R.; Kortelainen, M. J.; Lampén, T.; Lassila-Perini, K.; Lehti, S.; Lindén, T.; Luukka, P.; Mäenpää, T.; Peltola, T.; Tuominen, E.; Tuominiemi, J.; Tuovinen, E.; Ungaro, D.; Wendland, L.; Korpela, A.; Tuuva, T.; Besancon, M.; Choudhury, S.; Dejardin, M.; Denegri, D.; Fabbro, B.; Faure, J. L.; Ferri, F.; Ganjour, S.; Givernaud, A.; Gras, P.; Hamel de Monchenault, G.; Jarry, P.; Locci, E.; Malcles, J.; Millischer, L.; Nayak, A.; Rander, J.; Rosowsky, A.; Titov, M.; Baffioni, S.; Beaudette, F.; Benhabib, L.; Bianchini, L.; Bluj, M.; Busson, P.; Charlot, C.; Daci, N.; Dahms, T.; Dalchenko, M.; Dobrzynski, L.; Florent, A.; Granier de Cassagnac, R.; Haguenauer, M.; Miné, P.; Mironov, C.; Naranjo, I. N.; Nguyen, M.; Ochando, C.; Paganini, P.; Sabes, D.; Salerno, R.; Sirois, Y.; Veelken, C.; Zabi, A.; Agram, J.-L.; Andrea, J.; Bloch, D.; Bodin, D.; Brom, J.-M.; Cardaci, M.; Chabert, E. C.; Collard, C.; Conte, E.; Drouhin, F.; Fontaine, J.-C.; Gelé, D.; Goerlach, U.; Juillot, P.; Le Bihan, A.-C.; Van Hove, P.; Beauceron, S.; Beaupere, N.; Bondu, O.; Boudoul, G.; Brochet, S.; Chasserat, J.; Chierici, R.; Contardo, D.; Depasse, P.; El Mamouni, H.; Fay, J.; Gascon, S.; Gouzevitch, M.; Ille, B.; Kurca, T.; Lethuillier, M.; Mirabito, L.; Perries, S.; Sgandurra, L.; Sordini, V.; Tschudi, Y.; Verdier, P.; Viret, S.; Tsamalaidze, Z.; Autermann, C.; Beranek, S.; Calpas, B.; Edelhoff, M.; Feld, L.; Heracleous, N.; Hindrichs, O.; Jussen, R.; Klein, K.; Merz, J.; Ostapchuk, A.; Perieanu, A.; Raupach, F.; Sammet, J.; Schael, S.; Sprenger, D.; Weber, H.; Wittmer, B.; Zhukov, V.; Ata, M.; Caudron, J.; Dietz-Laursonn, E.; Duchardt, D.; Erdmann, M.; Fischer, R.; Güth, A.; Hebbeker, T.; Heidemann, C.; Hoepfner, K.; Klingebiel, D.; Kreuzer, P.; Merschmeyer, M.; Meyer, A.; Olschewski, M.; Papacz, P.; Pieta, H.; Reithler, H.; Schmitz, S. A.; Sonnenschein, L.; Steggemann, J.; Teyssier, D.; Thüer, S.; Weber, M.; Bontenackels, M.; Cherepanov, V.; Erdogan, Y.; Flügge, G.; Geenen, H.; Geisler, M.; Haj Ahmad, W.; Hoehle, F.; Kargoll, B.; Kress, T.; Kuessel, Y.; Lingemann, J.; Nowack, A.; Perchalla, L.; Pooth, O.; Sauerland, P.; Stahl, A.; Aldaya Martin, M.; Behr, J.; Behrenhoff, W.; Behrens, U.; Bergholz, M.; Bethani, A.; Borras, K.; Burgmeier, A.; Cakir, A.; Calligaris, L.; Campbell, A.; Castro, E.; Costanza, F.; Dammann, D.; Diez Pardos, C.; Dorland, T.; Eckerlin, G.; Eckstein, D.; Flucke, G.; Geiser, A.; Glushkov, I.; Gunnellini, P.; Habib, S.; Hauk, J.; Hellwig, G.; Jung, H.; Kasemann, M.; Katsas, P.; Kleinwort, C.; Kluge, H.; Knutsson, A.; Krämer, M.; Krücker, D.; Kuznetsova, E.; Lange, W.; Leonard, J.; Lohmann, W.; Lutz, B.; Mankel, R.; Marfin, I.; Marienfeld, M.; Melzer-Pellmann, I.-A.; Meyer, A. B.; Mnich, J.; Mussgiller, A.; Naumann-Emme, S.; Novgorodova, O.; Nowak, F.; Olzem, J.; Perrey, H.; Petrukhin, A.; Pitzl, D.; Raspereza, A.; Ribeiro Cipriano, P. M.; Riedl, C.; Ron, E.; Rosin, M.; Salfeld-Nebgen, J.; Schmidt, R.; Schoerner-Sadenius, T.; Sen, N.; Spiridonov, A.; Stein, M.; Walsh, R.; Wissing, C.; Blobel, V.; Enderle, H.; Erfle, J.; Gebbert, U.; Görner, M.; Gosselink, M.; Haller, J.; Hermanns, T.; Höing, R. S.; Kaschube, K.; Kaussen, G.; Kirschenmann, H.; Klanner, R.; Lange, J.; Peiffer, T.; Pietsch, N.; Rathjens, D.; Sander, C.; Schettler, H.; Schleper, P.; Schlieckau, E.; Schmidt, A.; Schröder, M.; Schum, T.; Seidel, M.; Sibille, J.; Sola, V.; Stadie, H.; Steinbrück, G.; Thomsen, J.; Vanelderen, L.; Barth, C.; Berger, J.; Böser, C.; Chwalek, T.; De Boer, W.; Descroix, A.; Dierlamm, A.; Feindt, M.; Guthoff, M.; Hackstein, C.; Hartmann, F.; Hauth, T.; Heinrich, M.; Held, H.; Hoffmann, K. H.; Husemann, U.; Katkov, I.; Komaragiri, J. R.; Lobelle Pardo, P.; Martschei, D.; Mueller, S.; Müller, Th.; Niegel, M.; Nürnberg, A.; Oberst, O.; Oehler, A.; Ott, J.; Quast, G.; Rabbertz, K.; Ratnikov, F.; Ratnikova, N.; Röcker, S.; Schilling, F.-P.; Schott, G.; Simonis, H. J.; Stober, F. M.; Troendle, D.; Ulrich, R.; Wagner-Kuhr, J.; Wayand, S.; Weiler, T.; Zeise, M.; Anagnostou, G.; Daskalakis, G.; Geralis, T.; Kesisoglou, S.; Kyriakis, A.; Loukas, D.; Manolakos, I.; Markou, A.; Markou, C.; Ntomari, E.; Gouskos, L.; Mertzimekis, T. J.; Panagiotou, A.; Saoulidou, N.; Evangelou, I.; Foudas, C.; Kokkas, P.; Manthos, N.; Papadopoulos, I.; Bencze, G.; Hajdu, C.; Hidas, P.; Horvath, D.; Sikler, F.; Veszpremi, V.; Vesztergombi, G.; Zsigmond, A. J.; Beni, N.; Czellar, S.; Molnar, J.; Palinkas, J.; Szillasi, Z.; Karancsi, J.; Raics, P.; Trocsanyi, Z. L.; Ujvari, B.; Beri, S. B.; Bhatnagar, V.; Dhingra, N.; Gupta, R.; Kaur, M.; Mehta, M. Z.; Mittal, M.; Nishu, N.; Saini, L. K.; Sharma, A.; Singh, J. B.; Kumar, Ashok; Kumar, Arun; Ahuja, S.; Bhardwaj, A.; Choudhary, B. C.; Malhotra, S.; Naimuddin, M.; Ranjan, K.; Sharma, V.; Shivpuri, R. K.; Banerjee, S.; Bhattacharya, S.; Chatterjee, K.; Dutta, S.; Gomber, B.; Jain, Sa.; Jain, Sh.; Khurana, R.; Modak, A.; Mukherjee, S.; Roy, D.; Sarkar, S.; Sharan, M.; Abdulsalam, A.; Dutta, D.; Kailas, S.; Kumar, V.; Mohanty, A. K.; Pant, L. M.; Shukla, P.; Aziz, T.; Chatterjee, R. M.; Ganguly, S.; Guchait, M.; Gurtu, A.; Maity, M.; Majumder, G.; Mazumdar, K.; Mohanty, G. B.; Parida, B.; Sudhakar, K.; Wickramage, N.; Banerjee, S.; Dugad, S.; Arfaei, H.; Bakhshiansohi, H.; Etesami, S. M.; Fahim, A.; Hashemi, M.; Hesari, H.; Jafari, A.; Khakzad, M.; Mohammadi Najafabadi, M.; Paktinat Mehdiabadi, S.; Safarzadeh, B.; Zeinali, M.; Abbrescia, M.; Barbone, L.; Calabria, C.; Chhibra, S. S.; Colaleo, A.; Creanza, D.; De Filippis, N.; De Palma, M.; Fiore, L.; Iaselli, G.; Maggi, G.; Maggi, M.; Marangelli, B.; My, S.; Nuzzo, S.; Pacifico, N.; Pompili, A.; Pugliese, G.; Selvaggi, G.; Silvestris, L.; Singh, G.; Venditti, R.; Verwilligen, P.; Zito, G.; Abbiendi, G.; Benvenuti, A. C.; Bonacorsi, D.; Braibant-Giacomelli, S.; Brigliadori, L.; Capiluppi, P.; Castro, A.; Cavallo, F. R.; Cuffiani, M.; Dallavalle, G. M.; Fabbri, F.; Fanfani, A.; Fasanella, D.; Giacomelli, P.; Grandi, C.; Guiducci, L.; Marcellini, S.; Masetti, G.; Meneghelli, M.; Montanari, A.; Navarria, F. L.; Odorici, F.; Perrotta, A.; Primavera, F.; Rossi, A. M.; Rovelli, T.; Siroli, G. P.; Tosi, N.; Travaglini, R.; Albergo, S.; Cappello, G.; Chiorboli, M.; Costa, S.; Potenza, R.; Tricomi, A.; Tuve, C.; Barbagli, G.; Ciulli, V.; Civinini, C.; D'Alessandro, R.; Focardi, E.; Frosali, S.; Gallo, E.; Gonzi, S.; Meschini, M.; Paoletti, S.; Sguazzoni, G.; Tropiano, A.; Benussi, L.; Bianco, S.; Colafranceschi, S.; Fabbri, F.; Piccolo, D.; Fabbricatore, P.; Musenich, R.; Tosi, S.; Benaglia, A.; De Guio, F.; Di Matteo, L.; Fiorendi, S.; Gennai, S.; Ghezzi, A.; Malvezzi, S.; Manzoni, R. A.; Martelli, A.; Massironi, A.; Menasce, D.; Moroni, L.; Paganoni, M.; Pedrini, D.; Ragazzi, S.; Redaelli, N.; Tabarelli de Fatis, T.; Buontempo, S.; Cavallo, N.; De Cosa, A.; Dogangun, O.; Fabozzi, F.; Iorio, A. O. M.; Lista, L.; Meola, S.; Merola, M.; Paolucci, P.; Azzi, P.; Bacchetta, N.; Bellato, M.; Bisello, D.; Branca, A.; Carlin, R.; Checchia, P.; Dorigo, T.; Gasparini, F.; Gozzelino, A.; Kanishchev, K.; Lacaprara, S.; Lazzizzera, I.; Margoni, M.; Meneguzzo, A. T.; Pazzini, J.; Pozzobon, N.; Ronchese, P.; Simonetto, F.; Torassa, E.; Tosi, M.; Vanini, S.; Zotto, P.; Zucchetta, A.; Zumerle, G.; Gabusi, M.; Ratti, S. P.; Riccardi, C.; Torre, P.; Vitulo, P.; Biasini, M.; Bilei, G. M.; Fanò, L.; Lariccia, P.; Mantovani, G.; Menichelli, M.; Nappi, A.; Romeo, F.; Saha, A.; Santocchia, A.; Spiezia, A.; Taroni, S.; Azzurri, P.; Bagliesi, G.; Bernardini, J.; Boccali, T.; Broccolo, G.; Castaldi, R.; D'Agnolo, R. T.; Dell'Orso, R.; Fiori, F.; Foà, L.; Giassi, A.; Kraan, A.; Ligabue, F.; Lomtadze, T.; Martini, L.; Messineo, A.; Palla, F.; Rizzi, A.; Serban, A. T.; Spagnolo, P.; Squillacioti, P.; Tenchini, R.; Tonelli, G.; Venturi, A.; Verdini, P. G.; Barone, L.; Cavallari, F.; Del Re, D.; Diemoz, M.; Fanelli, C.; Grassi, M.; Longo, E.; Meridiani, P.; Micheli, F.; Nourbakhsh, S.; Organtini, G.; Paramatti, R.; Rahatlou, S.; Soffi, L.; Amapane, N.; Arcidiacono, R.; Argiro, S.; Arneodo, M.; Biino, C.; Cartiglia, N.; Casasso, S.; Costa, M.; Demaria, N.; Mariotti, C.; Maselli, S.; Migliore, E.; Monaco, V.; Musich, M.; Obertino, M. M.; Pastrone, N.; Pelliccioni, M.; Potenza, A.; Romero, A.; Ruspa, M.; Sacchi, R.; Solano, A.; Staiano, A.; Belforte, S.; Candelise, V.; Casarsa, M.; Cossutti, F.; Della Ricca, G.; Gobbo, B.; Marone, M.; Montanino, D.; Penzo, A.; Schizzi, A.; Kim, T. Y.; Nam, S. K.; Chang, S.; Kim, D. H.; Kim, G. N.; Kong, D. J.; Park, H.; Son, D. C.; Son, T.; Kim, J. Y.; Kim, Zero J.; Song, S.; Choi, S.; Gyun, D.; Hong, B.; Jo, M.; Kim, H.; Kim, T. J.; Lee, K. S.; Moon, D. H.; Park, S. K.; Roh, Y.; Choi, M.; Kim, J. H.; Park, C.; Park, I. C.; Park, S.; Ryu, G.; Choi, Y.; Choi, Y. K.; Goh, J.; Kim, M. S.; Kwon, E.; Lee, B.; Lee, J.; Lee, S.; Seo, H.; Yu, I.; Bilinskas, M. J.; Grigelionis, I.; Janulis, M.; Juodagalvis, A.; Castilla-Valdez, H.; De La Cruz-Burelo, E.; Heredia-de La Cruz, I.; Lopez-Fernandez, R.; Martínez-Ortega, J.; Sanchez-Hernandez, A.; Villasenor-Cendejas, L. M.; Carrillo Moreno, S.; Vazquez Valencia, F.; Salazar Ibarguen, H. A.; Casimiro Linares, E.; Morelos Pineda, A.; Reyes-Santos, M. A.; Krofcheck, D.; Bell, A. J.; Butler, P. H.; Doesburg, R.; Reucroft, S.; Silverwood, H.; Ahmad, M.; Asghar, M. I.; Butt, J.; Hoorani, H. R.; Khalid, S.; Khan, W. A.; Khurshid, T.; Qazi, S.; Shah, M. A.; Shoaib, M.; Bialkowska, H.; Boimska, B.; Frueboes, T.; Górski, M.; Kazana, M.; Nawrocki, K.; Romanowska-Rybinska, K.; Szleper, M.; Wrochna, G.; Zalewski, P.; Brona, G.; Bunkowski, K.; Cwiok, M.; Dominik, W.; Doroba, K.; Kalinowski, A.; Konecki, M.; Krolikowski, J.; Misiura, M.; Almeida, N.; Bargassa, P.; David, A.; Faccioli, P.; Ferreira Parracho, P. G.; Gallinaro, M.; Seixas, J.; Varela, J.; Vischia, P.; Belotelov, I.; Bunin, P.; Gavrilenko, M.; Golutvin, I.; Gorbunov, I.; Kamenev, A.; Karjavin, V.; Kozlov, G.; Lanev, A.; Malakhov, A.; Moisenz, P.; Palichik, V.; Perelygin, V.; Shmatov, S.; Smirnov, V.; Volodko, A.; Zarubin, A.; Evstyukhin, S.; Golovtsov, V.; Ivanov, Y.; Kim, V.; Levchenko, P.; Murzin, V.; Oreshkin, V.; Smirnov, I.; Sulimov, V.; Uvarov, L.; Vavilov, S.; Vorobyev, A.; Vorobyev, An.; Andreev, Yu.; Dermenev, A.; Gninenko, S.; Golubev, N.; Kirsanov, M.; Krasnikov, N.; Matveev, V.; Pashenkov, A.; Tlisov, D.; Toropin, A.; Epshteyn, V.; Erofeeva, M.; Gavrilov, V.; Kossov, M.; Lychkovskaya, N.; Popov, V.; Safronov, G.; Semenov, S.; Shreyber, I.; Stolin, V.; Vlasov, E.; Zhokin, A.; Belyaev, A.; Boos, E.; Dubinin, M.; Dudko, L.; Ershov, A.; Gribushin, A.; Klyukhin, V.; Kodolova, O.; Lokhtin, I.; Markina, A.; Obraztsov, S.; Perfilov, M.; Petrushanko, S.; Popov, A.; Sarycheva, L.; Savrin, V.; Snigirev, A.; Andreev, V.; Azarkin, M.; Dremin, I.; Kirakosyan, M.; Leonidov, A.; Mesyats, G.; Rusakov, S. V.; Vinogradov, A.; Azhgirey, I.; Bayshev, I.; Bitioukov, S.; Grishin, V.; Kachanov, V.; Konstantinov, D.; Krychkine, V.; Petrov, V.; Ryutin, R.; Sobol, A.; Tourtchanovitch, L.; Troshin, S.; Tyurin, N.; Uzunian, A.; Volkov, A.; Adzic, P.; Djordjevic, M.; Ekmedzic, M.; Krpic, D.; Milosevic, J.; Aguilar-Benitez, M.; Alcaraz Maestre, J.; Arce, P.; Battilana, C.; Calvo, E.; Cerrada, M.; Chamizo Llatas, M.; Colino, N.; De La Cruz, B.; Delgado Peris, A.; Domínguez Vázquez, D.; Fernandez Bedoya, C.; Fernández Ramos, J. P.; Ferrando, A.; Flix, J.; Fouz, M. C.; Garcia-Abia, P.; Gonzalez Lopez, O.; Goy Lopez, S.; Hernandez, J. M.; Josa, M. I.; Merino, G.; Puerta Pelayo, J.; Quintario Olmeda, A.; Redondo, I.; Romero, L.; Santaolalla, J.; Soares, M. S.; Willmott, C.; Albajar, C.; Codispoti, G.; de Trocóniz, J. F.; Brun, H.; Cuevas, J.; Fernandez Menendez, J.; Folgueras, S.; Gonzalez Caballero, I.; Lloret Iglesias, L.; Piedra Gomez, J.; Brochero Cifuentes, J. A.; Cabrillo, I. J.; Calderon, A.; Chuang, S. H.; Duarte Campderros, J.; Felcini, M.; Fernandez, M.; Gomez, G.; Gonzalez Sanchez, J.; Graziano, A.; Jorda, C.; Lopez Virto, A.; Marco, J.; Marco, R.; Martinez Rivero, C.; Matorras, F.; Munoz Sanchez, F. J.; Rodrigo, T.; Rodríguez-Marrero, A. Y.; Ruiz-Jimeno, A.; Scodellaro, L.; Vila, I.; Vilar Cortabitarte, R.; Abbaneo, D.; Auffray, E.; Auzinger, G.; Bachtis, M.; Baillon, P.; Ball, A. H.; Barney, D.; Benitez, J. F.; Bernet, C.; Bianchi, G.; Bloch, P.; Bocci, A.; Bonato, A.; Botta, C.; Breuker, H.; Camporesi, T.; Cerminara, G.; Christiansen, T.; Coarasa Perez, J. A.; D'Enterria, D.; Dabrowski, A.; De Roeck, A.; Di Guida, S.; Dobson, M.; Dupont-Sagorin, N.; Elliott-Peisert, A.; Frisch, B.; Funk, W.; Georgiou, G.; Giffels, M.; Gigi, D.; Gill, K.; Giordano, D.; Girone, M.; Giunta, M.; Glege, F.; Gomez-Reino Garrido, R.; Govoni, P.; Gowdy, S.; Guida, R.; Gundacker, S.; Hammer, J.; Hansen, M.; Harris, P.; Hartl, C.; Harvey, J.; Hegner, B.; Hinzmann, A.; Innocente, V.; Janot, P.; Kaadze, K.; Karavakis, E.; Kousouris, K.; Lecoq, P.; Lee, Y.-J.; Lenzi, P.; Lourenço, C.; Magini, N.; Mäki, T.; Malberti, M.; Malgeri, L.; Mannelli, M.; Masetti, L.; Meijers, F.; Mersi, S.; Meschi, E.; Moser, R.; Mulders, M.; Musella, P.; Nesvold, E.; Orsini, L.; Palencia Cortezon, E.; Perez, E.; Perrozzi, L.; Petrilli, A.; Pfeiffer, A.; Pierini, M.; Pimiä, M.; Piparo, D.; Polese, G.; Quertenmont, L.; Racz, A.; Reece, W.; Rodrigues Antunes, J.; Rolandi, G.; Rovelli, C.; Rovere, M.; Sakulin, H.; Santanastasio, F.; Schäfer, C.; Schwick, C.; Segoni, I.; Sekmen, S.; Sharma, A.; Siegrist, P.; Silva, P.; Simon, M.; Sphicas, P.; Spiga, D.; Tsirou, A.; Veres, G. I.; Vlimant, J. R.; Wöhri, H. K.; Worm, S. D.; Zeuner, W. D.; Bertl, W.; Deiters, K.; Erdmann, W.; Gabathuler, K.; Horisberger, R.; Ingram, Q.; Kaestli, H. C.; König, S.; Kotlinski, D.; Langenegger, U.; Meier, F.; Renker, D.; Rohe, T.; Bäni, L.; Bortignon, P.; Buchmann, M. A.; Casal, B.; Chanon, N.; Deisher, A.; Dissertori, G.; Dittmar, M.; Donegà, M.; Dünser, M.; Eller, P.; Eugster, J.; Freudenreich, K.; Grab, C.; Hits, D.; Lecomte, P.; Lustermann, W.; Marini, A. C.; Martinez Ruiz del Arbol, P.; Mohr, N.; Moortgat, F.; Nägeli, C.; Nef, P.; Nessi-Tedaldi, F.; Pandolfi, F.; Pape, L.; Pauss, F.; Peruzzi, M.; Ronga, F. J.; Rossini, M.; Sala, L.; Sanchez, A. K.; Starodumov, A.; Stieger, B.; Takahashi, M.; Tauscher, L.; Thea, A.; Theofilatos, K.; Treille, D.; Urscheler, C.; Wallny, R.; Weber, H. A.; Wehrli, L.; Amsler, C.; Chiochia, V.; De Visscher, S.; Favaro, C.; Ivova Rikova, M.; Kilminster, B.; Millan Mejias, B.; Otiougova, P.; Robmann, P.; Snoek, H.; Tupputi, S.; Verzetti, M.; Chang, Y. H.; Chen, K. H.; Ferro, C.; Kuo, C. M.; Li, S. W.; Lin, W.; Lu, Y. J.; Singh, A. P.; Volpe, R.; Yu, S. S.; Bartalini, P.; Chang, P.; Chang, Y. H.; Chang, Y. W.; Chao, Y.; Chen, K. F.; Dietz, C.; Grundler, U.; Hou, W.-S.; Hsiung, Y.; Kao, K. Y.; Lei, Y. J.; Lu, R.-S.; Majumder, D.; Petrakou, E.; Shi, X.; Shiu, J. G.; Tzeng, Y. M.; Wan, X.; Wang, M.; Asavapibhop, B.; Srimanobhas, N.; Suwonjandee, N.; Adiguzel, A.; Bakirci, M. N.; Cerci, S.; Dozen, C.; Dumanoglu, I.; Eskut, E.; Girgis, S.; Gokbulut, G.; Gurpinar, E.; Hos, I.; Kangal, E. E.; Karaman, T.; Karapinar, G.; Kayis Topaksu, A.; Onengut, G.; Ozdemir, K.; Ozturk, S.; Polatoz, A.; Sogut, K.; Sunar Cerci, D.; Tali, B.; Topakli, H.; Vergili, L. N.; Vergili, M.; Akin, I. V.; Aliev, T.; Bilin, B.; Bilmis, S.; Deniz, M.; Gamsizkan, H.; Guler, A. M.; Ocalan, K.; Ozpineci, A.; Serin, M.; Sever, R.; Surat, U. E.; Yalvac, M.; Yildirim, E.; Zeyrek, M.; Gülmez, E.; Isildak, B.; Kaya, M.; Kaya, O.; Ozkorucuklu, S.; Sonmez, N.; Bahtiyar, H.; Barlas, E.; Cankocak, K.; Günaydin, Y. O.; Vardarlı, F. I.; Yücel, M.; Levchuk, L.; Brooke, J. J.; Clement, E.; Cussans, D.; Flacher, H.; Frazier, R.; Goldstein, J.; Grimes, M.; Heath, G. P.; Heath, H. F.; Kreczko, L.; Metson, S.; Newbold, D. M.; Nirunpong, K.; Poll, A.; Senkin, S.; Smith, V. J.; Williams, T.; Basso, L.; Bell, K. W.; Belyaev, A.; Brew, C.; Brown, R. M.; Cockerill, D. J. A.; Coughlan, J. A.; Harder, K.; Harper, S.; Jackson, J.; Kennedy, B. W.; Olaiya, E.; Petyt, D.; Radburn-Smith, B. C.; Shepherd-Themistocleous, C. H.; Tomalin, I. R.; Womersley, W. J.; Bainbridge, R.; Ball, G.; Beuselinck, R.; Buchmuller, O.; Colling, D.; Cripps, N.; Cutajar, M.; Dauncey, P.; Davies, G.; Della Negra, M.; Ferguson, W.; Fulcher, J.; Futyan, D.; Gilbert, A.; Guneratne Bryer, A.; Hall, G.; Hatherell, Z.; Hays, J.; Iles, G.; Jarvis, M.; Karapostoli, G.; Lyons, L.; Magnan, A.-M.; Marrouche, J.; Mathias, B.; Nandi, R.; Nash, J.; Nikitenko, A.; Pela, J.; Pesaresi, M.; Petridis, K.; Pioppi, M.; Raymond, D. M.; Rogerson, S.; Rose, A.; Seez, C.; Sharp, P.; Sparrow, A.; Stoye, M.; Tapper, A.; Vazquez Acosta, M.; Virdee, T.; Wakefield, S.; Wardle, N.; Whyntie, T.; Chadwick, M.; Cole, J. E.; Hobson, P. R.; Khan, A.; Kyberd, P.; Leggat, D.; Leslie, D.; Martin, W.; Reid, I. D.; Symonds, P.; Teodorescu, L.; Turner, M.; Hatakeyama, K.; Liu, H.; Scarborough, T.; Charaf, O.; Henderson, C.; Rumerio, P.; Avetisyan, A.; Bose, T.; Fantasia, C.; Heister, A.; St. John, J.; Lawson, P.; Lazic, D.; Rohlf, J.; Sperka, D.; Sulak, L.; Alimena, J.; Bhattacharya, S.; Christopher, G.; Cutts, D.; Demiragli, Z.; Ferapontov, A.; Garabedian, A.; Heintz, U.; Jabeen, S.; Kukartsev, G.; Laird, E.; Landsberg, G.; Luk, M.; Narain, M.; Segala, M.; Sinthuprasith, T.; Speer, T.; Breedon, R.; Breto, G.; Calderon De La Barca Sanchez, M.; Chauhan, S.; Chertok, M.; Conway, J.; Conway, R.; Cox, P. T.; Dolen, J.; Erbacher, R.; Gardner, M.; Houtz, R.; Ko, W.; Kopecky, A.; Lander, R.; Mall, O.; Miceli, T.; Pellett, D.; Ricci-Tam, F.; Rutherford, B.; Searle, M.; Smith, J.; Squires, M.; Tripathi, M.; Vasquez Sierra, R.; Yohay, R.; Andreev, V.; Cline, D.; Cousins, R.; Duris, J.; Erhan, S.; Everaerts, P.; Farrell, C.; Hauser, J.; Ignatenko, M.; Jarvis, C.; Rakness, G.; Schlein, P.; Traczyk, P.; Valuev, V.; Weber, M.; Babb, J.; Clare, R.; Dinardo, M. E.; Ellison, J.; Gary, J. W.; Giordano, F.; Hanson, G.; Liu, H.; Long, O. R.; Luthra, A.; Nguyen, H.; Paramesvaran, S.; Sturdy, J.; Sumowidagdo, S.; Wilken, R.; Wimpenny, S.; Andrews, W.; Branson, J. G.; Cerati, G. B.; Cittolin, S.; Evans, D.; Holzner, A.; Kelley, R.; Lebourgeois, M.; Letts, J.; Macneill, I.; Mangano, B.; Padhi, S.; Palmer, C.; Petrucciani, G.; Pieri, M.; Sani, M.; Sharma, V.; Simon, S.; Sudano, E.; Tadel, M.; Tu, Y.; Vartak, A.; Wasserbaech, S.; Würthwein, F.; Yagil, A.; Yoo, J.; Barge, D.; Bellan, R.; Campagnari, C.; D'Alfonso, M.; Danielson, T.; Flowers, K.; Geffert, P.; George, C.; Golf, F.; Incandela, J.; Justus, C.; Kalavase, P.; Kovalskyi, D.; Krutelyov, V.; Lowette, S.; Magaña Villalba, R.; Mccoll, N.; Pavlunin, V.; Ribnik, J.; Richman, J.; Rossin, R.; Stuart, D.; To, W.; West, C.; Apresyan, A.; Bornheim, A.; Chen, Y.; Di Marco, E.; Duarte, J.; Gataullin, M.; Ma, Y.; Mott, A.; Newman, H. B.; Rogan, C.; Spiropulu, M.; Timciuc, V.; Veverka, J.; Wilkinson, R.; Xie, S.; Yang, Y.; Zhu, R. Y.; Azzolini, V.; Calamba, A.; Carroll, R.; Ferguson, T.; Iiyama, Y.; Jang, D. W.; Liu, Y. F.; Paulini, M.; Vogel, H.; Vorobiev, I.; Cumalat, J. P.; Drell, B. R.; Ford, W. T.; Gaz, A.; Luiggi Lopez, E.; Smith, J. G.; Stenson, K.; Ulmer, K. A.; Wagner, S. R.; Alexander, J.; Chatterjee, A.; Eggert, N.; Gibbons, L. K.; Heltsley, B.; Hopkins, W.; Khukhunaishvili, A.; Kreis, B.; Mirman, N.; Nicolas Kaufman, G.; Patterson, J. R.; Ryd, A.; Salvati, E.; Sun, W.; Teo, W. D.; Thom, J.; Thompson, J.; Tucker, J.; Vaughan, J.; Weng, Y.; Winstrom, L.; Wittich, P.; Winn, D.; Abdullin, S.; Albrow, M.; Anderson, J.; Apollinari, G.; Bauerdick, L. A. T.; Beretvas, A.; Berryhill, J.; Bhat, P. C.; Burkett, K.; Butler, J. N.; Chetluru, V.; Cheung, H. W. K.; Chlebana, F.; Elvira, V. D.; Fisk, I.; Freeman, J.; Gao, Y.; Green, D.; Gutsche, O.; Hanlon, J.; Harris, R. M.; Hirschauer, J.; Hooberman, B.; Jindariani, S.; Johnson, M.; Joshi, U.; Klima, B.; Kunori, S.; Kwan, S.; Leonidopoulos, C.; Linacre, J.; Lincoln, D.; Lipton, R.; Lykken, J.; Maeshima, K.; Marraffino, J. M.; Martinez Outschoorn, V. I.; Maruyama, S.; Mason, D.; McBride, P.; Mishra, K.; Mrenna, S.; Musienko, Y.; Newman-Holmes, C.; O'Dell, V.; Sexton-Kennedy, E.; Sharma, S.; Spalding, W. J.; Spiegel, L.; Taylor, L.; Tkaczyk, S.; Tran, N. V.; Uplegger, L.; Vaandering, E. W.; Vidal, R.; Whitmore, J.; Wu, W.; Yang, F.; Yun, J. C.; Acosta, D.; Avery, P.; Bourilkov, D.; Chen, M.; Cheng, T.; Das, S.; De Gruttola, M.; Di Giovanni, G. P.; Dobur, D.; Drozdetskiy, A.; Field, R. D.; Fisher, M.; Fu, Y.; Furic, I. K.; Gartner, J.; Hugon, J.; Kim, B.; Konigsberg, J.; Korytov, A.; Kropivnitskaya, A.; Kypreos, T.; Low, J. F.; Matchev, K.; Milenovic, P.; Mitselmakher, G.; Muniz, L.; Park, M.; Remington, R.; Rinkevicius, A.; Sellers, P.; Skhirtladze, N.; Snowball, M.; Yelton, J.; Zakaria, M.; Gaultney, V.; Hewamanage, S.; Lebolo, L. M.; Linn, S.; Markowitz, P.; Martinez, G.; Rodriguez, J. L.; Adams, T.; Askew, A.; Bochenek, J.; Chen, J.; Diamond, B.; Gleyzer, S. V.; Haas, J.; Hagopian, S.; Hagopian, V.; Jenkins, M.; Johnson, K. F.; Prosper, H.; Veeraraghavan, V.; Weinberg, M.; Baarmand, M. M.; Dorney, B.; Hohlmann, M.; Kalakhety, H.; Vodopiyanov, I.; Yumiceva, F.; Adams, M. R.; Anghel, I. M.; Apanasevich, L.; Bai, Y.; Bazterra, V. E.; Betts, R. R.; Bucinskaite, I.; Callner, J.; Cavanaugh, R.; Evdokimov, O.; Gauthier, L.; Gerber, C. E.; Hofman, D. J.; Khalatyan, S.; Lacroix, F.; O'Brien, C.; Silkworth, C.; Strom, D.; Turner, P.; Varelas, N.; Akgun, U.; Albayrak, E. A.; Bilki, B.; Clarida, W.; Duru, F.; Griffiths, S.; Merlo, J.-P.; Mermerkaya, H.; Mestvirishvili, A.; Moeller, A.; Nachtman, J.; Newsom, C. R.; Norbeck, E.; Onel, Y.; Ozok, F.; Sen, S.; Tan, P.; Tiras, E.; Wetzel, J.; Yetkin, T.; Yi, K.; Barnett, B. A.; Blumenfeld, B.; Bolognesi, S.; Fehling, D.; Giurgiu, G.; Gritsan, A. V.; Guo, Z. J.; Hu, G.; Maksimovic, P.; Swartz, M.; Whitbeck, A.; Baringer, P.; Bean, A.; Benelli, G.; Kenny, R. P.; Murray, M.; Noonan, D.; Sanders, S.; Stringer, R.; Tinti, G.; Wood, J. S.; Barfuss, A. F.; Bolton, T.; Chakaberia, I.; Ivanov, A.; Khalil, S.; Makouski, M.; Maravin, Y.; Shrestha, S.; Svintradze, I.; Gronberg, J.; Lange, D.; Rebassoo, F.; Wright, D.; Baden, A.; Calvert, B.; Eno, S. C.; Gomez, J. A.; Hadley, N. J.; Kellogg, R. G.; Kirn, M.; Kolberg, T.; Lu, Y.; Marionneau, M.; Mignerey, A. C.; Pedro, K.; Peterman, A.; Skuja, A.; Temple, J.; Tonjes, M. B.; Tonwar, S. C.; Apyan, A.; Bauer, G.; Bendavid, J.; Busza, W.; Butz, E.; Cali, I. A.; Chan, M.; Dutta, V.; Gomez Ceballos, G.; Goncharov, M.; Kim, Y.; Klute, M.; Krajczar, K.; Levin, A.; Luckey, P. D.; Ma, T.; Nahn, S.; Paus, C.; Ralph, D.; Roland, C.; Roland, G.; Rudolph, M.; Stephans, G. S. F.; Stöckli, F.; Sumorok, K.; Sung, K.; Velicanu, D.; Wenger, E. A.; Wolf, R.; Wyslouch, B.; Yang, M.; Yilmaz, Y.; Yoon, A. S.; Zanetti, M.; Zhukova, V.; Cooper, S. I.; Dahmes, B.; De Benedetti, A.; Franzoni, G.; Gude, A.; Kao, S. C.; Klapoetke, K.; Kubota, Y.; Mans, J.; Pastika, N.; Rusack, R.; Sasseville, M.; Singovsky, A.; Tambe, N.; Turkewitz, J.; Cremaldi, L. M.; Kroeger, R.; Perera, L.; Rahmat, R.; Sanders, D. A.; Avdeeva, E.; Bloom, K.; Bose, S.; Claes, D. R.; Dominguez, A.; Eads, M.; Keller, J.; Kravchenko, I.; Lazo-Flores, J.; Malik, S.; Snow, G. R.; Godshalk, A.; Iashvili, I.; Jain, S.; Kharchilava, A.; Kumar, A.; Rappoccio, S.; Wan, Z.; Alverson, G.; Barberis, E.; Baumgartel, D.; Chasco, M.; Haley, J.; Nash, D.; Orimoto, T.; Trocino, D.; Wood, D.; Zhang, J.; Anastassov, A.; Hahn, K. A.; Kubik, A.; Lusito, L.; Mucia, N.; Odell, N.; Ofierzynski, R. A.; Pollack, B.; Pozdnyakov, A.; Schmitt, M.; Stoynev, S.; Velasco, M.; Won, S.; Berry, D.; Brinkerhoff, A.; Chan, K. M.; Hildreth, M.; Jessop, C.; Karmgard, D. J.; Kolb, J.; Lannon, K.; Luo, W.; Lynch, S.; Marinelli, N.; Morse, D. M.; Pearson, T.; Planer, M.; Ruchti, R.; Slaunwhite, J.; Valls, N.; Wayne, M.; Wolf, M.; Antonelli, L.; Bylsma, B.; Durkin, L. S.; Hill, C.; Hughes, R.; Kotov, K.; Ling, T. Y.; Puigh, D.; Rodenburg, M.; Vuosalo, C.; Williams, G.; Winer, B. L.; Berry, E.; Elmer, P.; Halyo, V.; Hebda, P.; Hegeman, J.; Hunt, A.; Jindal, P.; Koay, S. A.; Lopes Pegna, D.; Lujan, P.; Marlow, D.; Medvedeva, T.; Mooney, M.; Olsen, J.; Piroué, P.; Quan, X.; Raval, A.; Saka, H.; Stickland, D.; Tully, C.; Werner, J. S.; Zenz, S. C.; Zuranski, A.; Brownson, E.; Lopez, A.; Mendez, H.; Ramirez Vargas, J. E.; Alagoz, E.; Barnes, V. E.; Benedetti, D.; Bolla, G.; Bortoletto, D.; De Mattia, M.; Everett, A.; Hu, Z.; Jones, M.; Koybasi, O.; Kress, M.; Laasanen, A. T.; Leonardo, N.; Maroussov, V.; Merkel, P.; Miller, D. H.; Neumeister, N.; Shipsey, I.; Silvers, D.; Svyatkovskiy, A.; Vidal Marono, M.; Yoo, H. D.; Zablocki, J.; Zheng, Y.; Guragain, S.; Parashar, N.; Adair, A.; Akgun, B.; Boulahouache, C.; Ecklund, K. M.; Geurts, F. J. M.; Li, W.; Padley, B. P.; Redjimi, R.; Roberts, J.; Zabel, J.; Betchart, B.; Bodek, A.; Chung, Y. S.; Covarelli, R.; de Barbaro, P.; Demina, R.; Eshaq, Y.; Ferbel, T.; Garcia-Bellido, A.; Goldenzweig, P.; Han, J.; Harel, A.; Miner, D. C.; Vishnevskiy, D.; Zielinski, M.; Bhatti, A.; Ciesielski, R.; Demortier, L.; Goulianos, K.; Lungu, G.; Malik, S.; Mesropian, C.; Arora, S.; Barker, A.; Chou, J. P.; Contreras-Campana, C.; Contreras-Campana, E.; Duggan, D.; Ferencek, D.; Gershtein, Y.; Gray, R.; Halkiadakis, E.; Hidas, D.; Lath, A.; Panwalkar, S.; Park, M.; Patel, R.; Rekovic, V.; Robles, J.; Rose, K.; Salur, S.; Schnetzer, S.; Seitz, C.; Somalwar, S.; Stone, R.; Thomas, S.; Walker, M.; Cerizza, G.; Hollingsworth, M.; Spanier, S.; Yang, Z. C.; York, A.; Eusebi, R.; Flanagan, W.; Gilmore, J.; Kamon, T.; Khotilovich, V.; Montalvo, R.; Osipenkov, I.; Pakhotin, Y.; Perloff, A.; Roe, J.; Safonov, A.; Sakuma, T.; Sengupta, S.; Suarez, I.; Tatarinov, A.; Toback, D.; Akchurin, N.; Damgov, J.; Dragoiu, C.; Dudero, P. R.; Jeong, C.; Kovitanggoon, K.; Lee, S. W.; Libeiro, T.; Volobouev, I.; Appelt, E.; Delannoy, A. G.; Florez, C.; Greene, S.; Gurrola, A.; Johns, W.; Kurt, P.; Maguire, C.; Melo, A.; Sharma, M.; Sheldon, P.; Snook, B.; Tuo, S.; Velkovska, J.; Arenton, M. W.; Balazs, M.; Boutle, S.; Cox, B.; Francis, B.; Goodell, J.; Hirosky, R.; Ledovskoy, A.; Lin, C.; Neu, C.; Wood, J.; Gollapinni, S.; Harr, R.; Karchin, P. E.; Kottachchi Kankanamge Don, C.; Lamichhane, P.; Sakharov, A.; Anderson, M.; Belknap, D. A.; Borrello, L.; Carlsmith, D.; Cepeda, M.; Dasu, S.; Friis, E.; Gray, L.; Grogg, K. S.; Grothe, M.; Hall-Wilton, R.; Herndon, M.; Hervé, A.; Klabbers, P.; Klukas, J.; Lanaro, A.; Lazaridis, C.; Loveless, R.; Mohapatra, A.; Mozer, M. U.; Ojalvo, I.; Palmonari, F.; Pierro, G. A.; Ross, I.; Savin, A.; Smith, W. H.; Swanson, J.

    2013-05-01

    Measurements of event shapes and azimuthal correlations are presented for events where a Z boson is produced in association with jets in proton-proton collisions. The data collected with the CMS detector at the CERN LHC at √{ s} = 7 TeV correspond to an integrated luminosity of 5.0 fb-1. The analysis provides a test of predictions from perturbative QCD for a process that represents a substantial background to many physics channels. Results are presented as a function of jet multiplicity, for inclusive Z boson production and for Z bosons with transverse momenta greater than 150 GeV, and compared to predictions from Monte Carlo event generators that include leading-order multiparton matrix-element (with up to four hard partons in the final state) and next-to-leading-order simulations of Z +1-jet events. The experimental results are corrected for detector effects, and can be compared directly with other QCD models.

  3. Event shapes and azimuthal correlations in Z + jets events in pp collisions at s = 7   TeV

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chatrchyan, S.; Khachatryan, V.; Sirunyan, A. M.

    Measurements of event shapes and azimuthal correlations are presented for events where a Z boson is produced in association with jets in proton-proton collisions. The data collected with the CMS detector at the CERN LHC at sqrt(s) = 7 TeV correspond to an integrated luminosity of 5.0 inverse femtobarns. The analysis provides a test of predictions from perturbative QCD for a process that represents a substantial background to many physics channels. Results are presented as a function of jet multiplicity, for inclusive Z boson production and for Z bosons with transverse momenta greater than 150 GeV, and compared to predictionsmore » from Monte Carlo event generators that include leading-order multiparton matrix-element (with up to four hard partons in the final state) and next-to-leading-order simulations of Z + 1-jet events. The experimental results are corrected for detector effects, and can be compared directly with other QCD models.« less

  4. Analysis of the variation of atmospheric electric field during solar events

    NASA Astrophysics Data System (ADS)

    Tacza, J.; Raulin, J. P.

    2016-12-01

    We present the capability of a new network of electric field mill sensors to monitor the atmospheric electric field at various locations in South America. The first task is to obtain a diurnal curve of atmospheric electric field variations under fair weather conditions, which we will consider as a reference curve. To accomplish this, we made daily, monthly, seasonal and annual averages. For all sensor location, the results show significant similarities with the Carnegie curve. The Carnegie curve is the characteristic curve in universal time of atmospheric electric field in fair weather and one thinks it is related to the currents flowing in the global atmospheric electric circuit. Ultimately, we pretend to study departures of the daily observations from the standard curve. This difference can be caused by solar, geophysical and atmospheric phenomena such as the solar activity cycle, solar flares and energetic charged particles, galactic cosmic rays, seismic activity and/or specific meteorological events. As an illustration we investigate solar effects on the atmospheric electric field observed at CASLEO (Lat. 31.798°S, Long. 69.295°W, Altitude: 2552 masl) by the method of superposed epoch analysis, between January 2010 and December 2015.

  5. Semantic Context Detection Using Audio Event Fusion

    NASA Astrophysics Data System (ADS)

    Chu, Wei-Ta; Cheng, Wen-Huang; Wu, Ja-Ling

    2006-12-01

    Semantic-level content analysis is a crucial issue in achieving efficient content retrieval and management. We propose a hierarchical approach that models audio events over a time series in order to accomplish semantic context detection. Two levels of modeling, audio event and semantic context modeling, are devised to bridge the gap between physical audio features and semantic concepts. In this work, hidden Markov models (HMMs) are used to model four representative audio events, that is, gunshot, explosion, engine, and car braking, in action movies. At the semantic context level, generative (ergodic hidden Markov model) and discriminative (support vector machine (SVM)) approaches are investigated to fuse the characteristics and correlations among audio events, which provide cues for detecting gunplay and car-chasing scenes. The experimental results demonstrate the effectiveness of the proposed approaches and provide a preliminary framework for information mining by using audio characteristics.

  6. A computer aided treatment event recognition system in radiation therapy.

    PubMed

    Xia, Junyi; Mart, Christopher; Bayouth, John

    2014-01-01

    To develop an automated system to safeguard radiation therapy treatments by analyzing electronic treatment records and reporting treatment events. CATERS (Computer Aided Treatment Event Recognition System) was developed to detect treatment events by retrieving and analyzing electronic treatment records. CATERS is designed to make the treatment monitoring process more efficient by automating the search of the electronic record for possible deviations from physician's intention, such as logical inconsistencies as well as aberrant treatment parameters (e.g., beam energy, dose, table position, prescription change, treatment overrides, etc). Over a 5 month period (July 2012-November 2012), physicists were assisted by the CATERS software in conducting normal weekly chart checks with the aims of (a) determining the relative frequency of particular events in the authors' clinic and (b) incorporating these checks into the CATERS. During this study period, 491 patients were treated at the University of Iowa Hospitals and Clinics for a total of 7692 fractions. All treatment records from the 5 month analysis period were evaluated using all the checks incorporated into CATERS after the training period. About 553 events were detected as being exceptions, although none of them had significant dosimetric impact on patient treatments. These events included every known event type that was discovered during the trial period. A frequency analysis of the events showed that the top three types of detected events were couch position override (3.2%), extra cone beam imaging (1.85%), and significant couch position deviation (1.31%). The significant couch deviation is defined as the number of treatments where couch vertical exceeded two times standard deviation of all couch verticals, or couch lateral/longitudinal exceeded three times standard deviation of all couch laterals and longitudinals. On average, the application takes about 1 s per patient when executed on either a desktop computer

  7. Implementation of a patient safety program at a tertiary health system: A longitudinal analysis of interventions and serious safety events.

    PubMed

    Cropper, Douglas P; Harb, Nidal H; Said, Patricia A; Lemke, Jon H; Shammas, Nicolas W

    2018-04-01

    We hypothesize that implementation of a safety program based on high reliability organization principles will reduce serious safety events (SSE). The safety program focused on 7 essential elements: (a) safety rounding, (b) safety oversight teams, (c) safety huddles, (d) safety coaches, (e) good catches/safety heroes, (f) safety education, and (g) red rule. An educational curriculum was implemented focusing on changing high-risk behaviors and implementing critical safety policies. All unusual occurrences were captured in the Midas system and investigated by risk specialists, the safety officer, and the chief medical officer. A multidepartmental committee evaluated these events, and a root cause analysis (RCA) was performed. Events were tabulated and serious safety event (SSE) recorded and plotted over time. Safety success stories (SSSs) were also evaluated over time. A steady drop in SSEs was seen over 9 years. Also a rise in SSSs was evident, reflecting on staff engagement in the program. The parallel change in SSEs, SSSs, and the implementation of various safety interventions highly suggest that the program was successful in achieving its goals. A safety program based on high-reliability organization principles and made a core value of the institution can have a significant positive impact on reducing SSEs. © 2018 American Society for Healthcare Risk Management of the American Hospital Association.

  8. Analysis of extreme rain and flood events using a regional hydrologically enhanced hydrometeorological system

    NASA Astrophysics Data System (ADS)

    Yucel, Ismail; Onen, Alper

    2013-04-01

    Evidence is showing that global warming or climate change has a direct influence on changes in precipitation and the hydrological cycle. Extreme weather events such as heavy rainfall and flooding are projected to become much more frequent as climate warms. Regional hydrometeorological system model which couples the atmosphere with physical and gridded based surface hydrology provide efficient predictions for extreme hydrological events. This modeling system can be used for flood forecasting and warning issues as they provide continuous monitoring of precipitation over large areas at high spatial resolution. This study examines the performance of the Weather Research and Forecasting (WRF-Hydro) model that performs the terrain, sub-terrain, and channel routing in producing streamflow from WRF-derived forcing of extreme precipitation events. The capability of the system with different options such as data assimilation is tested for number of flood events observed in basins of western Black Sea Region in Turkey. Rainfall event structures and associated flood responses are evaluated with gauge and satellite-derived precipitation and measured streamflow values. The modeling system shows skills in capturing the spatial and temporal structure of extreme rainfall events and resulted flood hydrographs. High-resolution routing modules activated in the model enhance the simulated discharges.

  9. Analysis of Effects of Sensor Multithreading to Generate Local System Event Timelines

    DTIC Science & Technology

    2014-03-27

    works on logs highlights the importance of logs [17, 18]. The two aforementioned works both reference the same 2009 Data Breach Investigations Report...the data breaches report on, the logs contained evidence of events leading up to 82% of those data breaches . This means that preventing 82% of the data ...report states that of the data breaches reported on, the logs contained evidence of events leading up to 66% of those data breaches . • The 2010 DBIR

  10. Integrating Remote Sensing Data, Hybrid-Cloud Computing, and Event Notifications for Advanced Rapid Imaging & Analysis (Invited)

    NASA Astrophysics Data System (ADS)

    Hua, H.; Owen, S. E.; Yun, S.; Lundgren, P.; Fielding, E. J.; Agram, P.; Manipon, G.; Stough, T. M.; Simons, M.; Rosen, P. A.; Wilson, B. D.; Poland, M. P.; Cervelli, P. F.; Cruz, J.

    2013-12-01

    Space-based geodetic measurement techniques such as Interferometric Synthetic Aperture Radar (InSAR) and Continuous Global Positioning System (CGPS) are now important elements in our toolset for monitoring earthquake-generating faults, volcanic eruptions, hurricane damage, landslides, reservoir subsidence, and other natural and man-made hazards. Geodetic imaging's unique ability to capture surface deformation with high spatial and temporal resolution has revolutionized both earthquake science and volcanology. Continuous monitoring of surface deformation and surface change before, during, and after natural hazards improves decision-making from better forecasts, increased situational awareness, and more informed recovery. However, analyses of InSAR and GPS data sets are currently handcrafted following events and are not generated rapidly and reliably enough for use in operational response to natural disasters. Additionally, the sheer data volumes needed to handle a continuous stream of InSAR data sets also presents a bottleneck. It has been estimated that continuous processing of InSAR coverage of California alone over 3-years would reach PB-scale data volumes. Our Advanced Rapid Imaging and Analysis for Monitoring Hazards (ARIA-MH) science data system enables both science and decision-making communities to monitor areas of interest with derived geodetic data products via seamless data preparation, processing, discovery, and access. We will present our findings on the use of hybrid-cloud computing to improve the timely processing and delivery of geodetic data products, integrating event notifications from USGS to improve the timely processing for response, as well as providing browse results for quick looks with other tools for integrative analysis.

  11. Adverse Event extraction from Structured Product Labels using the Event-based Text-mining of Health Electronic Records (ETHER)system.

    PubMed

    Pandey, Abhishek; Kreimeyer, Kory; Foster, Matthew; Botsis, Taxiarchis; Dang, Oanh; Ly, Thomas; Wang, Wei; Forshee, Richard

    2018-01-01

    Structured Product Labels follow an XML-based document markup standard approved by the Health Level Seven organization and adopted by the US Food and Drug Administration as a mechanism for exchanging medical products information. Their current organization makes their secondary use rather challenging. We used the Side Effect Resource database and DailyMed to generate a comparison dataset of 1159 Structured Product Labels. We processed the Adverse Reaction section of these Structured Product Labels with the Event-based Text-mining of Health Electronic Records system and evaluated its ability to extract and encode Adverse Event terms to Medical Dictionary for Regulatory Activities Preferred Terms. A small sample of 100 labels was then selected for further analysis. Of the 100 labels, Event-based Text-mining of Health Electronic Records achieved a precision and recall of 81 percent and 92 percent, respectively. This study demonstrated Event-based Text-mining of Health Electronic Record's ability to extract and encode Adverse Event terms from Structured Product Labels which may potentially support multiple pharmacoepidemiological tasks.

  12. Multi-voxel pattern classification differentiates personally experienced event memories from secondhand event knowledge.

    PubMed

    Chow, Tiffany E; Westphal, Andrew J; Rissman, Jesse

    2018-04-11

    Studies of autobiographical memory retrieval often use photographs to probe participants' memories for past events. Recent neuroimaging work has shown that viewing photographs depicting events from one's own life evokes a characteristic pattern of brain activity across a network of frontal, parietal, and medial temporal lobe regions that can be readily distinguished from brain activity associated with viewing photographs from someone else's life (Rissman, Chow, Reggente, and Wagner, 2016). However, it is unclear whether the neural signatures associated with remembering a personally experienced event are distinct from those associated with recognizing previously encountered photographs of an event. The present experiment used a novel functional magnetic resonance imaging (fMRI) paradigm to investigate putative differences in brain activity patterns associated with these distinct expressions of memory retrieval. Eighteen participants wore necklace-mounted digital cameras to capture events from their everyday lives over the course of three weeks. One week later, participants underwent fMRI scanning, where on each trial they viewed a sequence of photographs depicting either an event from their own life or from another participant's life and judged their memory for this event. Importantly, half of the trials featured photographic sequences that had been shown to participants during a laboratory session administered the previous day. Multi-voxel pattern analyses assessed the sensitivity of two brain networks of interest-as identified by a meta-analysis of prior autobiographical and laboratory-based memory retrieval studies-to the original source of the photographs (own life or other's life) and their experiential history as stimuli (previewed or non-previewed). The classification analyses revealed a striking dissociation: activity patterns within the autobiographical memory network were significantly more diagnostic than those within the laboratory-based network as to

  13. Creating Special Events

    ERIC Educational Resources Information Center

    deLisle, Lee

    2009-01-01

    "Creating Special Events" is organized as a systematic approach to festivals and events for students who seek a career in event management. This book looks at the evolution and history of festivals and events and proceeds to the nuts and bolts of event management. The book presents event management as the means of planning, organizing, directing,…

  14. On the analysis of an extreme Bora wind event over the northern Adriatic Sea

    NASA Astrophysics Data System (ADS)

    Colucci, R. R.; Pucillo, A.

    2010-09-01

    On 10th March 2010 a severe Bora wind event affected the Friuli Venezia Giulia region, northeastern Italy, in particular the gulf of Trieste area (northern Adriatic Sea). Such event has been driven by a widespread westward moving cold pool aloft, coming from the Western Asia, that brought an intense potential vorticity anomaly over the western Mediterranean Sea. It determined a deep cyclogenesis involving all the troposphere. The pressure gradient force in the lowest layers forced a northeastern wind to blow with noticeable strength over the gulf of Trieste area and the Karstic region. The mean ground wind velocity has reached values above 27 m/s (about 100 km/h) for several hours, and maximum gusts exceeded 42 m/s (about 150 km/h) over Trieste town. The northeastern sector of the Adriatic Sea is frequently affected by strong Bora events in particular during the winter semester. This is a characteristic local wind mostly influenced by the orography of the Karstic relieves to the east of Trieste town. The target of this work is to assess the climatological relevance of such an event by comparing it with the most representative events of the past. It has been possible thanks to the long term archive of meteorological observations in Trieste site (I.R. Accademia di Commercio e Nautica, Regio Comitato Talassografico Italiano, Ministero dell'Agricoltura e Foreste, Consiglio Nazionale delle Ricerche): we have found out that this is one of the ten strongest Bora event along the 1871-2010 period. Considerations about the trend and frequency of severe Bora events have been proposed.

  15. Shifts in mortality during a hot weather event in Vancouver, British Columbia: rapid assessment with case-only analysis.

    PubMed

    Kosatsky, Tom; Henderson, Sarah B; Pollock, Sue L

    2012-12-01

    We assessed shifts in patterns of mortality during a hot weather event in greater Vancouver, British Columbia. We used a case-only analysis to compare characteristics of individuals who died during the hottest week of 2009 with those who died (1) during earlier summer weeks in 2009 and (2) during the same calendar weeks in the summers of 2001 through 2008. Compared with the 8 previous weeks of 2009, odds of mortality during the summer's hottest week were highest in the 65 to 74 years age category, compared with the 85 years and older category (odds ratio [OR] = 1.47; 95% confidence interval [CI] = 1.06, 2.03). The number of deaths at home increased over deaths in hospitals or institutions (OR = 1.43; 95% CI = 1.10, 1.86). Densely populated administrative health areas were more affected. A shift toward deaths at home suggests that in-home-based protective measures should be part of planning for hot weather events in greater Vancouver. Targeting should be considered for those aged 65 to 74 years. The case-only approach is quick and easy to apply and can provide useful information about localized, time-limited events.

  16. Long-term Outcomes After Stepping Down Asthma Controller Medications: A Claims-Based, Time-to-Event Analysis.

    PubMed

    Rank, Matthew A; Johnson, Ryan; Branda, Megan; Herrin, Jeph; van Houten, Holly; Gionfriddo, Michael R; Shah, Nilay D

    2015-09-01

    Long-term outcomes after stepping down asthma medications are not well described. This study was a retrospective time-to-event analysis of individuals diagnosed with asthma who stepped down their asthma controller medications using a US claims database spanning 2000 to 2012. Four-month intervals were established and a step-down event was defined by a ≥ 50% decrease in days-supplied of controller medications from one interval to the next; this definition is inclusive of step-down that occurred without health-care provider guidance or as a consequence of a medication adherence lapse. Asthma stability in the period prior to step-down was defined by not having an asthma exacerbation (inpatient visit, ED visit, or dispensing of a systemic corticosteroid linked to an asthma visit) and having fewer than two rescue inhaler claims in a 4-month period. The primary outcome in the period following step-down was time-to-first asthma exacerbation. Thirty-two percent of the 26,292 included individuals had an asthma exacerbation in the 24-month period following step-down of asthma controller medication, though only 7% had an ED visit or hospitalization for asthma. The length of asthma stability prior to stepping down asthma medication was strongly associated with the risk of an asthma exacerbation in the subsequent 24-month period: < 4 months' stability, 44%; 4 to 7 months, 34%; 8 to 11 months, 30%; and ≥ 12 months, 21% (P < .001). In a large, claims-based, real-world study setting, 32% of individuals have an asthma exacerbation in the 2 years following a step-down event.

  17. Transition Region Explosive Events in He II 304Å: Observation and Analysis

    NASA Astrophysics Data System (ADS)

    Rust, Thomas; Kankelborg, Charles C.

    2016-05-01

    We present examples of transition region explosive events observed in the He II 304Å spectral line with the Multi Order Solar EUV Spectrograph (MOSES). With small (<5000 km) spatial scale and large non-thermal (100-150 km/s) velocities these events satisfy the observational signatures of transition region explosive events. Derived line profiles show distinct blue and red velocity components with very little broadening of either component. We observe little to no emission from low velocity plasma, making the plasmoid instability reconnection model unlikely as the plasma acceleration mechanism for these events. Rather, the single speed, bi-directional jet characteristics suggested by these data are consistent with acceleration via Petschek reconnection.Observations were made during the first sounding rocket flight of MOSES in 2006. MOSES forms images in 3 orders of a concave diffraction grating. Multilayer coatings largely restrict the passband to the He II 303.8Å and Si XI 303.3Å spectral lines. The angular field of view is about 8.5'x17', or about 20% of the solar disk. These images constitute projections of the volume I(x,y,λ), the intensity as a function of sky plane position and wavelength. Spectral line profiles are recovered via tomographic inversion of these projections. Inversion is carried out using a multiplicative algebraic reconstruction technique.

  18. Fractal analysis of GPS time series for early detection of disastrous seismic events

    NASA Astrophysics Data System (ADS)

    Filatov, Denis M.; Lyubushin, Alexey A.

    2017-03-01

    A new method of fractal analysis of time series for estimating the chaoticity of behaviour of open stochastic dynamical systems is developed. The method is a modification of the conventional detrended fluctuation analysis (DFA) technique. We start from analysing both methods from the physical point of view and demonstrate the difference between them which results in a higher accuracy of the new method compared to the conventional DFA. Then, applying the developed method to estimate the measure of chaoticity of a real dynamical system - the Earth's crust, we reveal that the latter exhibits two distinct mechanisms of transition to a critical state: while the first mechanism has already been known due to numerous studies of other dynamical systems, the second one is new and has not previously been described. Using GPS time series, we demonstrate efficiency of the developed method in identification of critical states of the Earth's crust. Finally we employ the method to solve a practically important task: we show how the developed measure of chaoticity can be used for early detection of disastrous seismic events and provide a detailed discussion of the numerical results, which are shown to be consistent with outcomes of other researches on the topic.

  19. Higher moments of net-proton multiplicity distributions in a heavy-ion event pile-up scenario

    NASA Astrophysics Data System (ADS)

    Garg, P.; Mishra, D. K.

    2017-10-01

    High-luminosity modern accelerators, like the Relativistic Heavy Ion Collider (RHIC) at Brookhaven National Laboratory (BNL) and Large Hadron Collider (LHC) at European Organization for Nuclear Research (CERN), inherently have event pile-up scenarios which significantly contribute to physics events as a background. While state-of-the-art tracking algorithms and detector concepts take care of these event pile-up scenarios, several offline analytical techniques are used to remove such events from the physics analysis. It is still difficult to identify the remaining pile-up events in an event sample for physics analysis. Since the fraction of these events is significantly small, it may not be as serious of an issue for other analyses as it would be for an event-by-event analysis. Particularly when the characteristics of the multiplicity distribution are observable, one needs to be very careful. In the present work, we demonstrate how a small fraction of residual pile-up events can change the moments and their ratios of an event-by-event net-proton multiplicity distribution, which are sensitive to the dynamical fluctuations due to the QCD critical point. For this study, we assume that the individual event-by-event proton and antiproton multiplicity distributions follow Poisson, negative binomial, or binomial distributions. We observe a significant effect in cumulants and their ratios of net-proton multiplicity distributions due to pile-up events, particularly at lower energies. It might be crucial to estimate the fraction of pile-up events in the data sample while interpreting the experimental observable for the critical point.

  20. Dynamic Event Tree advancements and control logic improvements

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Alfonsi, Andrea; Rabiti, Cristian; Mandelli, Diego

    The RAVEN code has been under development at the Idaho National Laboratory since 2012. Its main goal is to create a multi-purpose platform for the deploying of all the capabilities needed for Probabilistic Risk Assessment, uncertainty quantification, data mining analysis and optimization studies. RAVEN is currently equipped with three different sampling categories: Forward samplers (Monte Carlo, Latin Hyper Cube, Stratified, Grid Sampler, Factorials, etc.), Adaptive Samplers (Limit Surface search, Adaptive Polynomial Chaos, etc.) and Dynamic Event Tree (DET) samplers (Deterministic and Adaptive Dynamic Event Trees). The main subject of this document is to report the activities that have been donemore » in order to: start the migration of the RAVEN/RELAP-7 control logic system into MOOSE, and develop advanced dynamic sampling capabilities based on the Dynamic Event Tree approach. In order to provide to all MOOSE-based applications a control logic capability, in this Fiscal Year an initial migration activity has been initiated, moving the control logic system, designed for RELAP-7 by the RAVEN team, into the MOOSE framework. In this document, a brief explanation of what has been done is going to be reported. The second and most important subject of this report is about the development of a Dynamic Event Tree (DET) sampler named “Hybrid Dynamic Event Tree” (HDET) and its Adaptive variant “Adaptive Hybrid Dynamic Event Tree” (AHDET). As other authors have already reported, among the different types of uncertainties, it is possible to discern two principle types: aleatory and epistemic uncertainties. The classical Dynamic Event Tree is in charge of treating the first class (aleatory) uncertainties; the dependence of the probabilistic risk assessment and analysis on the epistemic uncertainties are treated by an initial Monte Carlo sampling (MCDET). From each Monte Carlo sample, a DET analysis is run (in total, N trees). The Monte Carlo employs a pre-sampling of