Rotation in the Dynamic Factor Modeling of Multivariate Stationary Time Series.
ERIC Educational Resources Information Center
Molenaar, Peter C. M.; Nesselroade, John R.
2001-01-01
Proposes a special rotation procedure for the exploratory dynamic factor model for stationary multivariate time series. The rotation procedure applies separately to each univariate component series of a q-variate latent factor series and transforms such a component, initially represented as white noise, into a univariate moving-average.…
Nonstationary frequency analysis for the trivariate flood series of the Weihe River
NASA Astrophysics Data System (ADS)
Jiang, Cong; Xiong, Lihua
2016-04-01
Some intensive human activities such as water-soil conservation can significantly alter the natural hydrological processes of rivers. In this study, the effect of the water-soil conservation on the trivariate flood series from the Weihe River located in the Northwest China is investigated. The annual maxima daily discharge, annual maxima 3-day flood volume and annual maxima 5-day flood volume are chosen as the study data and used to compose the trivariate flood series. The nonstationarities in both the individual univariate flood series and the corresponding antecedent precipitation series generating the flood events are examined by the Mann-Kendall trend test. It is found that all individual univariate flood series present significant decreasing trend, while the antecedent precipitation series can be treated as stationary. It indicates that the increase of the water-soil conservation land area has altered the rainfall-runoff relationship of the Weihe basin, and induced the nonstationarities in the three individual univariate flood series. The time-varying moments model based on the Pearson type III distribution is applied to capture the nonstationarities in the flood frequency distribution with the water-soil conservation land area introduced as the explanatory variable of the flood distribution parameters. Based on the analysis for each individual univariate flood series, the dependence structure among the three univariate flood series are investigated by the time-varying copula model also with the water-soil conservation land area as the explanatory variable of copula parameters. The results indicate that the dependence among the trivariate flood series is enhanced by the increase of water-soil conservation land area.
Multivariate stochastic analysis for Monthly hydrological time series at Cuyahoga River Basin
NASA Astrophysics Data System (ADS)
zhang, L.
2011-12-01
Copula has become a very powerful statistic and stochastic methodology in case of the multivariate analysis in Environmental and Water resources Engineering. In recent years, the popular one-parameter Archimedean copulas, e.g. Gumbel-Houggard copula, Cook-Johnson copula, Frank copula, the meta-elliptical copula, e.g. Gaussian Copula, Student-T copula, etc. have been applied in multivariate hydrological analyses, e.g. multivariate rainfall (rainfall intensity, duration and depth), flood (peak discharge, duration and volume), and drought analyses (drought length, mean and minimum SPI values, and drought mean areal extent). Copula has also been applied in the flood frequency analysis at the confluences of river systems by taking into account the dependence among upstream gauge stations rather than by using the hydrological routing technique. In most of the studies above, the annual time series have been considered as stationary signal which the time series have been assumed as independent identically distributed (i.i.d.) random variables. But in reality, hydrological time series, especially the daily and monthly hydrological time series, cannot be considered as i.i.d. random variables due to the periodicity existed in the data structure. Also, the stationary assumption is also under question due to the Climate Change and Land Use and Land Cover (LULC) change in the fast years. To this end, it is necessary to revaluate the classic approach for the study of hydrological time series by relaxing the stationary assumption by the use of nonstationary approach. Also as to the study of the dependence structure for the hydrological time series, the assumption of same type of univariate distribution also needs to be relaxed by adopting the copula theory. In this paper, the univariate monthly hydrological time series will be studied through the nonstationary time series analysis approach. The dependence structure of the multivariate monthly hydrological time series will be studied through the copula theory. As to the parameter estimation, the maximum likelihood estimation (MLE) will be applied. To illustrate the method, the univariate time series model and the dependence structure will be determined and tested using the monthly discharge time series of Cuyahoga River Basin.
Time Series Modelling of Syphilis Incidence in China from 2005 to 2012
Zhang, Xingyu; Zhang, Tao; Pei, Jiao; Liu, Yuanyuan; Li, Xiaosong; Medrano-Gracia, Pau
2016-01-01
Background The infection rate of syphilis in China has increased dramatically in recent decades, becoming a serious public health concern. Early prediction of syphilis is therefore of great importance for heath planning and management. Methods In this paper, we analyzed surveillance time series data for primary, secondary, tertiary, congenital and latent syphilis in mainland China from 2005 to 2012. Seasonality and long-term trend were explored with decomposition methods. Autoregressive integrated moving average (ARIMA) was used to fit a univariate time series model of syphilis incidence. A separate multi-variable time series for each syphilis type was also tested using an autoregressive integrated moving average model with exogenous variables (ARIMAX). Results The syphilis incidence rates have increased three-fold from 2005 to 2012. All syphilis time series showed strong seasonality and increasing long-term trend. Both ARIMA and ARIMAX models fitted and estimated syphilis incidence well. All univariate time series showed highest goodness-of-fit results with the ARIMA(0,0,1)×(0,1,1) model. Conclusion Time series analysis was an effective tool for modelling the historical and future incidence of syphilis in China. The ARIMAX model showed superior performance than the ARIMA model for the modelling of syphilis incidence. Time series correlations existed between the models for primary, secondary, tertiary, congenital and latent syphilis. PMID:26901682
Time Series Modelling of Syphilis Incidence in China from 2005 to 2012.
Zhang, Xingyu; Zhang, Tao; Pei, Jiao; Liu, Yuanyuan; Li, Xiaosong; Medrano-Gracia, Pau
2016-01-01
The infection rate of syphilis in China has increased dramatically in recent decades, becoming a serious public health concern. Early prediction of syphilis is therefore of great importance for heath planning and management. In this paper, we analyzed surveillance time series data for primary, secondary, tertiary, congenital and latent syphilis in mainland China from 2005 to 2012. Seasonality and long-term trend were explored with decomposition methods. Autoregressive integrated moving average (ARIMA) was used to fit a univariate time series model of syphilis incidence. A separate multi-variable time series for each syphilis type was also tested using an autoregressive integrated moving average model with exogenous variables (ARIMAX). The syphilis incidence rates have increased three-fold from 2005 to 2012. All syphilis time series showed strong seasonality and increasing long-term trend. Both ARIMA and ARIMAX models fitted and estimated syphilis incidence well. All univariate time series showed highest goodness-of-fit results with the ARIMA(0,0,1)×(0,1,1) model. Time series analysis was an effective tool for modelling the historical and future incidence of syphilis in China. The ARIMAX model showed superior performance than the ARIMA model for the modelling of syphilis incidence. Time series correlations existed between the models for primary, secondary, tertiary, congenital and latent syphilis.
2009-12-18
cannot be detected with univariate techniques, but require multivariate analysis instead (Kamitani and Tong [2005]). Two other time series analysis ...learning for time series analysis . The historical record of DBNs can be traced back to Dean and Kanazawa [1988] and Dean and Wellman [1991], with...Rev. 8-98) Prescribed by ANSI Std Z39-18 Keywords: Hidden Process Models, probabilistic time series modeling, functional Magnetic Resonance Imaging
Why didn't Box-Jenkins win (again)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pack, D.J.; Downing, D.J.
This paper focuses on the forecasting performance of the Box-Jenkins methodology applied to the 111 time series of the Makridakis competition. It considers the influence of the following factors: (1) time series length, (2) time-series information (autocorrelation) content, (3) time-series outliers or structural changes, (4) averaging results over time series, and (5) forecast time origin choice. It is found that the 111 time series contain substantial numbers of very short series, series with obvious structural change, and series whose histories are relatively uninformative. If these series are typical of those that one must face in practice, the real message ofmore » the competition is that univariate time series extrapolations will frequently fail regardless of the methodology employed to produce them.« less
NASA Astrophysics Data System (ADS)
Nahar, Jannatun; Johnson, Fiona; Sharma, Ashish
2018-02-01
Conventional bias correction is usually applied on a grid-by-grid basis, meaning that the resulting corrections cannot address biases in the spatial distribution of climate variables. To solve this problem, a two-step bias correction method is proposed here to correct time series at multiple locations conjointly. The first step transforms the data to a set of statistically independent univariate time series, using a technique known as independent component analysis (ICA). The mutually independent signals can then be bias corrected as univariate time series and back-transformed to improve the representation of spatial dependence in the data. The spatially corrected data are then bias corrected at the grid scale in the second step. The method has been applied to two CMIP5 General Circulation Model simulations for six different climate regions of Australia for two climate variables—temperature and precipitation. The results demonstrate that the ICA-based technique leads to considerable improvements in temperature simulations with more modest improvements in precipitation. Overall, the method results in current climate simulations that have greater equivalency in space and time with observational data.
Zhang, Yong; Zhong, Miner; Geng, Nana; Jiang, Yunjian
2017-01-01
The market demand for electric vehicles (EVs) has increased in recent years. Suitable models are necessary to understand and forecast EV sales. This study presents a singular spectrum analysis (SSA) as a univariate time-series model and vector autoregressive model (VAR) as a multivariate model. Empirical results suggest that SSA satisfactorily indicates the evolving trend and provides reasonable results. The VAR model, which comprised exogenous parameters related to the market on a monthly basis, can significantly improve the prediction accuracy. The EV sales in China, which are categorized into battery and plug-in EVs, are predicted in both short term (up to December 2017) and long term (up to 2020), as statistical proofs of the growth of the Chinese EV industry.
Zhang, Yong; Zhong, Miner; Geng, Nana; Jiang, Yunjian
2017-01-01
The market demand for electric vehicles (EVs) has increased in recent years. Suitable models are necessary to understand and forecast EV sales. This study presents a singular spectrum analysis (SSA) as a univariate time-series model and vector autoregressive model (VAR) as a multivariate model. Empirical results suggest that SSA satisfactorily indicates the evolving trend and provides reasonable results. The VAR model, which comprised exogenous parameters related to the market on a monthly basis, can significantly improve the prediction accuracy. The EV sales in China, which are categorized into battery and plug-in EVs, are predicted in both short term (up to December 2017) and long term (up to 2020), as statistical proofs of the growth of the Chinese EV industry. PMID:28459872
Regenerating time series from ordinal networks.
McCullough, Michael; Sakellariou, Konstantinos; Stemler, Thomas; Small, Michael
2017-03-01
Recently proposed ordinal networks not only afford novel methods of nonlinear time series analysis but also constitute stochastic approximations of the deterministic flow time series from which the network models are constructed. In this paper, we construct ordinal networks from discrete sampled continuous chaotic time series and then regenerate new time series by taking random walks on the ordinal network. We then investigate the extent to which the dynamics of the original time series are encoded in the ordinal networks and retained through the process of regenerating new time series by using several distinct quantitative approaches. First, we use recurrence quantification analysis on traditional recurrence plots and order recurrence plots to compare the temporal structure of the original time series with random walk surrogate time series. Second, we estimate the largest Lyapunov exponent from the original time series and investigate the extent to which this invariant measure can be estimated from the surrogate time series. Finally, estimates of correlation dimension are computed to compare the topological properties of the original and surrogate time series dynamics. Our findings show that ordinal networks constructed from univariate time series data constitute stochastic models which approximate important dynamical properties of the original systems.
Regenerating time series from ordinal networks
NASA Astrophysics Data System (ADS)
McCullough, Michael; Sakellariou, Konstantinos; Stemler, Thomas; Small, Michael
2017-03-01
Recently proposed ordinal networks not only afford novel methods of nonlinear time series analysis but also constitute stochastic approximations of the deterministic flow time series from which the network models are constructed. In this paper, we construct ordinal networks from discrete sampled continuous chaotic time series and then regenerate new time series by taking random walks on the ordinal network. We then investigate the extent to which the dynamics of the original time series are encoded in the ordinal networks and retained through the process of regenerating new time series by using several distinct quantitative approaches. First, we use recurrence quantification analysis on traditional recurrence plots and order recurrence plots to compare the temporal structure of the original time series with random walk surrogate time series. Second, we estimate the largest Lyapunov exponent from the original time series and investigate the extent to which this invariant measure can be estimated from the surrogate time series. Finally, estimates of correlation dimension are computed to compare the topological properties of the original and surrogate time series dynamics. Our findings show that ordinal networks constructed from univariate time series data constitute stochastic models which approximate important dynamical properties of the original systems.
Forecasting electricity usage using univariate time series models
NASA Astrophysics Data System (ADS)
Hock-Eam, Lim; Chee-Yin, Yip
2014-12-01
Electricity is one of the important energy sources. A sufficient supply of electricity is vital to support a country's development and growth. Due to the changing of socio-economic characteristics, increasing competition and deregulation of electricity supply industry, the electricity demand forecasting is even more important than before. It is imperative to evaluate and compare the predictive performance of various forecasting methods. This will provide further insights on the weakness and strengths of each method. In literature, there are mixed evidences on the best forecasting methods of electricity demand. This paper aims to compare the predictive performance of univariate time series models for forecasting the electricity demand using a monthly data of maximum electricity load in Malaysia from January 2003 to December 2013. Results reveal that the Box-Jenkins method produces the best out-of-sample predictive performance. On the other hand, Holt-Winters exponential smoothing method is a good forecasting method for in-sample predictive performance.
Multivariate Time Series Decomposition into Oscillation Components.
Matsuda, Takeru; Komaki, Fumiyasu
2017-08-01
Many time series are considered to be a superposition of several oscillation components. We have proposed a method for decomposing univariate time series into oscillation components and estimating their phases (Matsuda & Komaki, 2017 ). In this study, we extend that method to multivariate time series. We assume that several oscillators underlie the given multivariate time series and that each variable corresponds to a superposition of the projections of the oscillators. Thus, the oscillators superpose on each variable with amplitude and phase modulation. Based on this idea, we develop gaussian linear state-space models and use them to decompose the given multivariate time series. The model parameters are estimated from data using the empirical Bayes method, and the number of oscillators is determined using the Akaike information criterion. Therefore, the proposed method extracts underlying oscillators in a data-driven manner and enables investigation of phase dynamics in a given multivariate time series. Numerical results show the effectiveness of the proposed method. From monthly mean north-south sunspot number data, the proposed method reveals an interesting phase relationship.
Compounding approach for univariate time series with nonstationary variances
NASA Astrophysics Data System (ADS)
Schäfer, Rudi; Barkhofen, Sonja; Guhr, Thomas; Stöckmann, Hans-Jürgen; Kuhl, Ulrich
2015-12-01
A defining feature of nonstationary systems is the time dependence of their statistical parameters. Measured time series may exhibit Gaussian statistics on short time horizons, due to the central limit theorem. The sample statistics for long time horizons, however, averages over the time-dependent variances. To model the long-term statistical behavior, we compound the local distribution with the distribution of its parameters. Here, we consider two concrete, but diverse, examples of such nonstationary systems: the turbulent air flow of a fan and a time series of foreign exchange rates. Our main focus is to empirically determine the appropriate parameter distribution for the compounding approach. To this end, we extract the relevant time scales by decomposing the time signals into windows and determine the distribution function of the thus obtained local variances.
Compounding approach for univariate time series with nonstationary variances.
Schäfer, Rudi; Barkhofen, Sonja; Guhr, Thomas; Stöckmann, Hans-Jürgen; Kuhl, Ulrich
2015-12-01
A defining feature of nonstationary systems is the time dependence of their statistical parameters. Measured time series may exhibit Gaussian statistics on short time horizons, due to the central limit theorem. The sample statistics for long time horizons, however, averages over the time-dependent variances. To model the long-term statistical behavior, we compound the local distribution with the distribution of its parameters. Here, we consider two concrete, but diverse, examples of such nonstationary systems: the turbulent air flow of a fan and a time series of foreign exchange rates. Our main focus is to empirically determine the appropriate parameter distribution for the compounding approach. To this end, we extract the relevant time scales by decomposing the time signals into windows and determine the distribution function of the thus obtained local variances.
NASA Astrophysics Data System (ADS)
Chek, Mohd Zaki Awang; Ahmad, Abu Bakar; Ridzwan, Ahmad Nur Azam Ahmad; Jelas, Imran Md.; Jamal, Nur Faezah; Ismail, Isma Liana; Zulkifli, Faiz; Noor, Syamsul Ikram Mohd
2012-09-01
The main objective of this study is to forecast the future claims amount of Invalidity Pension Scheme (IPS). All data were derived from SOCSO annual reports from year 1972 - 2010. These claims consist of all claims amount from 7 benefits offered by SOCSO such as Invalidity Pension, Invalidity Grant, Survivors Pension, Constant Attendance Allowance, Rehabilitation, Funeral and Education. Prediction of future claims of Invalidity Pension Scheme will be made using Univariate Forecasting Models to predict the future claims among workforce in Malaysia.
Multidimensional stock network analysis: An Escoufier's RV coefficient approach
NASA Astrophysics Data System (ADS)
Lee, Gan Siew; Djauhari, Maman A.
2013-09-01
The current practice of stocks network analysis is based on the assumption that the time series of closed stock price could represent the behaviour of the each stock. This assumption leads to consider minimal spanning tree (MST) and sub-dominant ultrametric (SDU) as an indispensible tool to filter the economic information contained in the network. Recently, there is an attempt where researchers represent stock not only as a univariate time series of closed price but as a bivariate time series of closed price and volume. In this case, they developed the so-called multidimensional MST to filter the important economic information. However, in this paper, we show that their approach is only applicable for that bivariate time series only. This leads us to introduce a new methodology to construct MST where each stock is represented by a multivariate time series. An example of Malaysian stock exchange will be presented and discussed to illustrate the advantages of the method.
Advanced spectral methods for climatic time series
Ghil, M.; Allen, M.R.; Dettinger, M.D.; Ide, K.; Kondrashov, D.; Mann, M.E.; Robertson, A.W.; Saunders, A.; Tian, Y.; Varadi, F.; Yiou, P.
2002-01-01
The analysis of univariate or multivariate time series provides crucial information to describe, understand, and predict climatic variability. The discovery and implementation of a number of novel methods for extracting useful information from time series has recently revitalized this classical field of study. Considerable progress has also been made in interpreting the information so obtained in terms of dynamical systems theory. In this review we describe the connections between time series analysis and nonlinear dynamics, discuss signal- to-noise enhancement, and present some of the novel methods for spectral analysis. The various steps, as well as the advantages and disadvantages of these methods, are illustrated by their application to an important climatic time series, the Southern Oscillation Index. This index captures major features of interannual climate variability and is used extensively in its prediction. Regional and global sea surface temperature data sets are used to illustrate multivariate spectral methods. Open questions and further prospects conclude the review.
Multivariate time series clustering on geophysical data recorded at Mt. Etna from 1996 to 2003
NASA Astrophysics Data System (ADS)
Di Salvo, Roberto; Montalto, Placido; Nunnari, Giuseppe; Neri, Marco; Puglisi, Giuseppe
2013-02-01
Time series clustering is an important task in data analysis issues in order to extract implicit, previously unknown, and potentially useful information from a large collection of data. Finding useful similar trends in multivariate time series represents a challenge in several areas including geophysics environment research. While traditional time series analysis methods deal only with univariate time series, multivariate time series analysis is a more suitable approach in the field of research where different kinds of data are available. Moreover, the conventional time series clustering techniques do not provide desired results for geophysical datasets due to the huge amount of data whose sampling rate is different according to the nature of signal. In this paper, a novel approach concerning geophysical multivariate time series clustering is proposed using dynamic time series segmentation and Self Organizing Maps techniques. This method allows finding coupling among trends of different geophysical data recorded from monitoring networks at Mt. Etna spanning from 1996 to 2003, when the transition from summit eruptions to flank eruptions occurred. This information can be used to carry out a more careful evaluation of the state of volcano and to define potential hazard assessment at Mt. Etna.
NASA Astrophysics Data System (ADS)
Kurniati, Devi; Hoyyi, Abdul; Widiharih, Tatik
2018-05-01
Time series data is a series of data taken or measured based on observations at the same time interval. Time series data analysis is used to perform data analysis considering the effect of time. The purpose of time series analysis is to know the characteristics and patterns of a data and predict a data value in some future period based on data in the past. One of the forecasting methods used for time series data is the state space model. This study discusses the modeling and forecasting of electric energy consumption using the state space model for univariate data. The modeling stage is began with optimal Autoregressive (AR) order selection, determination of state vector through canonical correlation analysis, estimation of parameter, and forecasting. The result of this research shows that modeling of electric energy consumption using state space model of order 4 with Mean Absolute Percentage Error (MAPE) value 3.655%, so the model is very good forecasting category.
Intimate partner violence in Madrid: a time series analysis (2008-2016).
Sanz-Barbero, Belén; Linares, Cristina; Vives-Cases, Carmen; González, José Luis; López-Ossorio, Juan José; Díaz, Julio
2018-06-02
This study analyzes whether there are time patterns in different intimate partner violence (IPV) indicators and aims to obtain models that can predict the behavior of these time series. Univariate autoregressive moving average models were used to analyze the time series corresponding to the number of daily calls to the 016 telephone IPV helpline and the number of daily police reports filed in the Community of Madrid during the period 2008-2015. Predictions were made for both dependent variables for 2016. The daily number of calls to the 016 telephone IPV helpline decreased during January 2008-April 2012 and increased during April 2012-December 2015. No statistically significant change was observed in the trend of the number of daily IPV police reports. The number of IPV police reports filed increased on weekends and on Christmas holidays. The number of calls to the 016 IPV help line increased on Mondays. Using data from 2008 to 2015, the univariate autoregressive moving average models predicted 64.2% of calls to the 016 telephone IPV helpline and 73.2% of police reports filed during 2016 in the Community of Madrid. Our results suggest the need for an increase in police and judicial resources on nonwork days. Also, the 016 telephone IPV helpline should be especially active on work days. Copyright © 2018 Elsevier Inc. All rights reserved.
The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...
A Method for Comparing Multivariate Time Series with Different Dimensions
Tapinos, Avraam; Mendes, Pedro
2013-01-01
In many situations it is desirable to compare dynamical systems based on their behavior. Similarity of behavior often implies similarity of internal mechanisms or dependency on common extrinsic factors. While there are widely used methods for comparing univariate time series, most dynamical systems are characterized by multivariate time series. Yet, comparison of multivariate time series has been limited to cases where they share a common dimensionality. A semi-metric is a distance function that has the properties of non-negativity, symmetry and reflexivity, but not sub-additivity. Here we develop a semi-metric – SMETS – that can be used for comparing groups of time series that may have different dimensions. To demonstrate its utility, the method is applied to dynamic models of biochemical networks and to portfolios of shares. The former is an example of a case where the dependencies between system variables are known, while in the latter the system is treated (and behaves) as a black box. PMID:23393554
Process fault detection and nonlinear time series analysis for anomaly detection in safeguards
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burr, T.L.; Mullen, M.F.; Wangen, L.E.
In this paper we discuss two advanced techniques, process fault detection and nonlinear time series analysis, and apply them to the analysis of vector-valued and single-valued time-series data. We investigate model-based process fault detection methods for analyzing simulated, multivariate, time-series data from a three-tank system. The model-predictions are compared with simulated measurements of the same variables to form residual vectors that are tested for the presence of faults (possible diversions in safeguards terminology). We evaluate two methods, testing all individual residuals with a univariate z-score and testing all variables simultaneously with the Mahalanobis distance, for their ability to detect lossmore » of material from two different leak scenarios from the three-tank system: a leak without and with replacement of the lost volume. Nonlinear time-series analysis tools were compared with the linear methods popularized by Box and Jenkins. We compare prediction results using three nonlinear and two linear modeling methods on each of six simulated time series: two nonlinear and four linear. The nonlinear methods performed better at predicting the nonlinear time series and did as well as the linear methods at predicting the linear values.« less
ERIC Educational Resources Information Center
Mistler, Stephen A.; Enders, Craig K.
2017-01-01
Multiple imputation methods can generally be divided into two broad frameworks: joint model (JM) imputation and fully conditional specification (FCS) imputation. JM draws missing values simultaneously for all incomplete variables using a multivariate distribution, whereas FCS imputes variables one at a time from a series of univariate conditional…
Recursive least squares background prediction of univariate syndromic surveillance data
2009-01-01
Background Surveillance of univariate syndromic data as a means of potential indicator of developing public health conditions has been used extensively. This paper aims to improve the performance of detecting outbreaks by using a background forecasting algorithm based on the adaptive recursive least squares method combined with a novel treatment of the Day of the Week effect. Methods Previous work by the first author has suggested that univariate recursive least squares analysis of syndromic data can be used to characterize the background upon which a prediction and detection component of a biosurvellance system may be built. An adaptive implementation is used to deal with data non-stationarity. In this paper we develop and implement the RLS method for background estimation of univariate data. The distinctly dissimilar distribution of data for different days of the week, however, can affect filter implementations adversely, and so a novel procedure based on linear transformations of the sorted values of the daily counts is introduced. Seven-days ahead daily predicted counts are used as background estimates. A signal injection procedure is used to examine the integrated algorithm's ability to detect synthetic anomalies in real syndromic time series. We compare the method to a baseline CDC forecasting algorithm known as the W2 method. Results We present detection results in the form of Receiver Operating Characteristic curve values for four different injected signal to noise ratios using 16 sets of syndromic data. We find improvements in the false alarm probabilities when compared to the baseline W2 background forecasts. Conclusion The current paper introduces a prediction approach for city-level biosurveillance data streams such as time series of outpatient clinic visits and sales of over-the-counter remedies. This approach uses RLS filters modified by a correction for the weekly patterns often seen in these data series, and a threshold detection algorithm from the residuals of the RLS forecasts. We compare the detection performance of this algorithm to the W2 method recently implemented at CDC. The modified RLS method gives consistently better sensitivity at multiple background alert rates, and we recommend that it should be considered for routine application in bio-surveillance systems. PMID:19149886
Recursive least squares background prediction of univariate syndromic surveillance data.
Najmi, Amir-Homayoon; Burkom, Howard
2009-01-16
Surveillance of univariate syndromic data as a means of potential indicator of developing public health conditions has been used extensively. This paper aims to improve the performance of detecting outbreaks by using a background forecasting algorithm based on the adaptive recursive least squares method combined with a novel treatment of the Day of the Week effect. Previous work by the first author has suggested that univariate recursive least squares analysis of syndromic data can be used to characterize the background upon which a prediction and detection component of a biosurvellance system may be built. An adaptive implementation is used to deal with data non-stationarity. In this paper we develop and implement the RLS method for background estimation of univariate data. The distinctly dissimilar distribution of data for different days of the week, however, can affect filter implementations adversely, and so a novel procedure based on linear transformations of the sorted values of the daily counts is introduced. Seven-days ahead daily predicted counts are used as background estimates. A signal injection procedure is used to examine the integrated algorithm's ability to detect synthetic anomalies in real syndromic time series. We compare the method to a baseline CDC forecasting algorithm known as the W2 method. We present detection results in the form of Receiver Operating Characteristic curve values for four different injected signal to noise ratios using 16 sets of syndromic data. We find improvements in the false alarm probabilities when compared to the baseline W2 background forecasts. The current paper introduces a prediction approach for city-level biosurveillance data streams such as time series of outpatient clinic visits and sales of over-the-counter remedies. This approach uses RLS filters modified by a correction for the weekly patterns often seen in these data series, and a threshold detection algorithm from the residuals of the RLS forecasts. We compare the detection performance of this algorithm to the W2 method recently implemented at CDC. The modified RLS method gives consistently better sensitivity at multiple background alert rates, and we recommend that it should be considered for routine application in bio-surveillance systems.
The short time Fourier transform and local signals
NASA Astrophysics Data System (ADS)
Okumura, Shuhei
In this thesis, I examine the theoretical properties of the short time discrete Fourier transform (STFT). The STFT is obtained by applying the Fourier transform by a fixed-sized, moving window to input series. We move the window by one time point at a time, so we have overlapping windows. I present several theoretical properties of the STFT, applied to various types of complex-valued, univariate time series inputs, and their outputs in closed forms. In particular, just like the discrete Fourier transform, the STFT's modulus time series takes large positive values when the input is a periodic signal. One main point is that a white noise time series input results in the STFT output being a complex-valued stationary time series and we can derive the time and time-frequency dependency structure such as the cross-covariance functions. Our primary focus is the detection of local periodic signals. I present a method to detect local signals by computing the probability that the squared modulus STFT time series has consecutive large values exceeding some threshold after one exceeding observation following one observation less than the threshold. We discuss a method to reduce the computation of such probabilities by the Box-Cox transformation and the delta method, and show that it works well in comparison to the Monte Carlo simulation method.
Jeffrey P. Prestemon
2009-01-01
Timber product markets are subject to large shocks deriving from natural disturbances and policy shifts. Statistical modeling of shocks is often done to assess their economic importance. In this article, I simulate the statistical power of univariate and bivariate methods of shock detection using time series intervention models. Simulations show that bivariate methods...
Vial, Flavie; Wei, Wei; Held, Leonhard
2016-12-20
In an era of ubiquitous electronic collection of animal health data, multivariate surveillance systems (which concurrently monitor several data streams) should have a greater probability of detecting disease events than univariate systems. However, despite their limitations, univariate aberration detection algorithms are used in most active syndromic surveillance (SyS) systems because of their ease of application and interpretation. On the other hand, a stochastic modelling-based approach to multivariate surveillance offers more flexibility, allowing for the retention of historical outbreaks, for overdispersion and for non-stationarity. While such methods are not new, they are yet to be applied to animal health surveillance data. We applied an example of such stochastic model, Held and colleagues' two-component model, to two multivariate animal health datasets from Switzerland. In our first application, multivariate time series of the number of laboratories test requests were derived from Swiss animal diagnostic laboratories. We compare the performance of the two-component model to parallel monitoring using an improved Farrington algorithm and found both methods yield a satisfactorily low false alarm rate. However, the calibration test of the two-component model on the one-step ahead predictions proved satisfactory, making such an approach suitable for outbreak prediction. In our second application, the two-component model was applied to the multivariate time series of the number of cattle abortions and the number of test requests for bovine viral diarrhea (a disease that often results in abortions). We found that there is a two days lagged effect from the number of abortions to the number of test requests. We further compared the joint modelling and univariate modelling of the number of laboratory test requests time series. The joint modelling approach showed evidence of superiority in terms of forecasting abilities. Stochastic modelling approaches offer the potential to address more realistic surveillance scenarios through, for example, the inclusion of times series specific parameters, or of covariates known to have an impact on syndrome counts. Nevertheless, many methodological challenges to multivariate surveillance of animal SyS data still remain. Deciding on the amount of corroboration among data streams that is required to escalate into an alert is not a trivial task given the sparse data on the events under consideration (e.g. disease outbreaks).
Multivariate multiscale entropy of financial markets
NASA Astrophysics Data System (ADS)
Lu, Yunfan; Wang, Jun
2017-11-01
In current process of quantifying the dynamical properties of the complex phenomena in financial market system, the multivariate financial time series are widely concerned. In this work, considering the shortcomings and limitations of univariate multiscale entropy in analyzing the multivariate time series, the multivariate multiscale sample entropy (MMSE), which can evaluate the complexity in multiple data channels over different timescales, is applied to quantify the complexity of financial markets. Its effectiveness and advantages have been detected with numerical simulations with two well-known synthetic noise signals. For the first time, the complexity of four generated trivariate return series for each stock trading hour in China stock markets is quantified thanks to the interdisciplinary application of this method. We find that the complexity of trivariate return series in each hour show a significant decreasing trend with the stock trading time progressing. Further, the shuffled multivariate return series and the absolute multivariate return series are also analyzed. As another new attempt, quantifying the complexity of global stock markets (Asia, Europe and America) is carried out by analyzing the multivariate returns from them. Finally we utilize the multivariate multiscale entropy to assess the relative complexity of normalized multivariate return volatility series with different degrees.
NASA Astrophysics Data System (ADS)
Zhu, Zhe
2017-08-01
The free and open access to all archived Landsat images in 2008 has completely changed the way of using Landsat data. Many novel change detection algorithms based on Landsat time series have been developed We present a comprehensive review of four important aspects of change detection studies based on Landsat time series, including frequencies, preprocessing, algorithms, and applications. We observed the trend that the more recent the study, the higher the frequency of Landsat time series used. We reviewed a series of image preprocessing steps, including atmospheric correction, cloud and cloud shadow detection, and composite/fusion/metrics techniques. We divided all change detection algorithms into six categories, including thresholding, differencing, segmentation, trajectory classification, statistical boundary, and regression. Within each category, six major characteristics of different algorithms, such as frequency, change index, univariate/multivariate, online/offline, abrupt/gradual change, and sub-pixel/pixel/spatial were analyzed. Moreover, some of the widely-used change detection algorithms were also discussed. Finally, we reviewed different change detection applications by dividing these applications into two categories, change target and change agent detection.
A univariate model of river water nitrate time series
NASA Astrophysics Data System (ADS)
Worrall, F.; Burt, T. P.
1999-01-01
Four time series were taken from three catchments in the North and South of England. The sites chosen included two in predominantly agricultural catchments, one at the tidal limit and one downstream of a sewage treatment works. A time series model was constructed for each of these series as a means of decomposing the elements controlling river water nitrate concentrations and to assess whether this approach could provide a simple management tool for protecting water abstractions. Autoregressive (AR) modelling of the detrended and deseasoned time series showed a "memory effect". This memory effect expressed itself as an increase in the winter-summer difference in nitrate levels that was dependent upon the nitrate concentration 12 or 6 months previously. Autoregressive moving average (ARMA) modelling showed that one of the series contained seasonal, non-stationary elements that appeared as an increasing trend in the winter-summer difference. The ARMA model was used to predict nitrate levels and predictions were tested against data held back from the model construction process - predictions gave average percentage errors of less than 10%. Empirical modelling can therefore provide a simple, efficient method for constructing management models for downstream water abstraction.
Detecting a currency’s dominance using multivariate time series analysis
NASA Astrophysics Data System (ADS)
Syahidah Yusoff, Nur; Sharif, Shamshuritawati
2017-09-01
A currency exchange rate is the price of one country’s currency in terms of another country’s currency. There are four different prices; opening, closing, highest, and lowest can be achieved from daily trading activities. In the past, a lot of studies have been carried out by using closing price only. However, those four prices are interrelated to each other. Thus, the multivariate time series can provide more information than univariate time series. Therefore, the enthusiasm of this paper is to compare the results of two different approaches, which are mean vector and Escoufier’s RV coefficient in constructing similarity matrices of 20 world currencies. Consequently, both matrices are used to substitute the correlation matrix required by network topology. With the help of degree centrality measure, we can detect the currency’s dominance for both networks. The pros and cons for both approaches will be presented at the end of this paper.
Aboagye-Sarfo, Patrick; Mai, Qun; Sanfilippo, Frank M; Preen, David B; Stewart, Louise M; Fatovich, Daniel M
2015-10-01
To develop multivariate vector-ARMA (VARMA) forecast models for predicting emergency department (ED) demand in Western Australia (WA) and compare them to the benchmark univariate autoregressive moving average (ARMA) and Winters' models. Seven-year monthly WA state-wide public hospital ED presentation data from 2006/07 to 2012/13 were modelled. Graphical and VARMA modelling methods were used for descriptive analysis and model fitting. The VARMA models were compared to the benchmark univariate ARMA and Winters' models to determine their accuracy to predict ED demand. The best models were evaluated by using error correction methods for accuracy. Descriptive analysis of all the dependent variables showed an increasing pattern of ED use with seasonal trends over time. The VARMA models provided a more precise and accurate forecast with smaller confidence intervals and better measures of accuracy in predicting ED demand in WA than the ARMA and Winters' method. VARMA models are a reliable forecasting method to predict ED demand for strategic planning and resource allocation. While the ARMA models are a closely competing alternative, they under-estimated future ED demand. Copyright © 2015 Elsevier Inc. All rights reserved.
Fitzgerald, Michael G.; Karlinger, Michael R.
1983-01-01
Time-series models were constructed for analysis of daily runoff and sediment discharge data from selected rivers of the Eastern United States. Logarithmic transformation and first-order differencing of the data sets were necessary to produce second-order, stationary time series and remove seasonal trends. Cyclic models accounted for less than 42 percent of the variance in the water series and 31 percent in the sediment series. Analysis of the apparent oscillations of given frequencies occurring in the data indicates that frequently occurring storms can account for as much as 50 percent of the variation in sediment discharge. Components of the frequency analysis indicate that a linear representation is reasonable for the water-sediment system. Models that incorporate lagged water discharge as input prove superior to univariate techniques in modeling and prediction of sediment discharges. The random component of the models includes errors in measurement and model hypothesis and indicates no serial correlation. An index of sediment production within or between drain-gage basins can be calculated from model parameters.
1991-09-01
However, there is no guarantee that this would work; for instance if the data were generated by an ARCH model (Tong, 1990 pp. 116-117) then a simple...Hill, R., Griffiths, W., Lutkepohl, H., and Lee, T., Introduction to the Theory and Practice of Econometrics , 2th ed., Wiley, 1985. Kendall, M., Stuart
Ridge Polynomial Neural Network with Error Feedback for Time Series Forecasting
Ghazali, Rozaida; Herawan, Tutut
2016-01-01
Time series forecasting has gained much attention due to its many practical applications. Higher-order neural network with recurrent feedback is a powerful technique that has been used successfully for time series forecasting. It maintains fast learning and the ability to learn the dynamics of the time series over time. Network output feedback is the most common recurrent feedback for many recurrent neural network models. However, not much attention has been paid to the use of network error feedback instead of network output feedback. In this study, we propose a novel model, called Ridge Polynomial Neural Network with Error Feedback (RPNN-EF) that incorporates higher order terms, recurrence and error feedback. To evaluate the performance of RPNN-EF, we used four univariate time series with different forecasting horizons, namely star brightness, monthly smoothed sunspot numbers, daily Euro/Dollar exchange rate, and Mackey-Glass time-delay differential equation. We compared the forecasting performance of RPNN-EF with the ordinary Ridge Polynomial Neural Network (RPNN) and the Dynamic Ridge Polynomial Neural Network (DRPNN). Simulation results showed an average 23.34% improvement in Root Mean Square Error (RMSE) with respect to RPNN and an average 10.74% improvement with respect to DRPNN. That means that using network errors during training helps enhance the overall forecasting performance for the network. PMID:27959927
Ridge Polynomial Neural Network with Error Feedback for Time Series Forecasting.
Waheeb, Waddah; Ghazali, Rozaida; Herawan, Tutut
2016-01-01
Time series forecasting has gained much attention due to its many practical applications. Higher-order neural network with recurrent feedback is a powerful technique that has been used successfully for time series forecasting. It maintains fast learning and the ability to learn the dynamics of the time series over time. Network output feedback is the most common recurrent feedback for many recurrent neural network models. However, not much attention has been paid to the use of network error feedback instead of network output feedback. In this study, we propose a novel model, called Ridge Polynomial Neural Network with Error Feedback (RPNN-EF) that incorporates higher order terms, recurrence and error feedback. To evaluate the performance of RPNN-EF, we used four univariate time series with different forecasting horizons, namely star brightness, monthly smoothed sunspot numbers, daily Euro/Dollar exchange rate, and Mackey-Glass time-delay differential equation. We compared the forecasting performance of RPNN-EF with the ordinary Ridge Polynomial Neural Network (RPNN) and the Dynamic Ridge Polynomial Neural Network (DRPNN). Simulation results showed an average 23.34% improvement in Root Mean Square Error (RMSE) with respect to RPNN and an average 10.74% improvement with respect to DRPNN. That means that using network errors during training helps enhance the overall forecasting performance for the network.
Elbert, Yevgeniy; Burkom, Howard S
2009-11-20
This paper discusses further advances in making robust predictions with the Holt-Winters forecasts for a variety of syndromic time series behaviors and introduces a control-chart detection approach based on these forecasts. Using three collections of time series data, we compare biosurveillance alerting methods with quantified measures of forecast agreement, signal sensitivity, and time-to-detect. The study presents practical rules for initialization and parameterization of biosurveillance time series. Several outbreak scenarios are used for detection comparison. We derive an alerting algorithm from forecasts using Holt-Winters-generalized smoothing for prospective application to daily syndromic time series. The derived algorithm is compared with simple control-chart adaptations and to more computationally intensive regression modeling methods. The comparisons are conducted on background data from both authentic and simulated data streams. Both types of background data include time series that vary widely by both mean value and cyclic or seasonal behavior. Plausible, simulated signals are added to the background data for detection performance testing at signal strengths calculated to be neither too easy nor too hard to separate the compared methods. Results show that both the sensitivity and the timeliness of the Holt-Winters-based algorithm proved to be comparable or superior to that of the more traditional prediction methods used for syndromic surveillance.
Nonlinear Dynamics, Poor Data, and What to Make of Them?
NASA Astrophysics Data System (ADS)
Ghil, M.; Zaliapin, I. V.
2005-12-01
The analysis of univariate or multivariate time series provides crucial information to describe, understand, and predict variability in the geosciences. The discovery and implementation of a number of novel methods for extracting useful information from time series has recently revitalized this classical field of study. Considerable progress has also been made in interpreting the information so obtained in terms of dynamical systems theory. In this talk we will describe the connections between time series analysis and nonlinear dynamics, discuss signal-to-noise enhancement, and present some of the novel methods for spectral analysis. These fall into two broad categories: (i) methods that try to ferret out regularities of the time series; and (ii) methods aimed at describing the characteristics of irregular processes. The former include singular-spectrum analysis (SSA), the multi-taper method (MTM), and the maximum-entropy method (MEM). The various steps, as well as the advantages and disadvantages of these methods, will be illustrated by their application to several important climatic time series, such as the Southern Oscillation Index (SOI), paleoclimatic time series, and instrumental temperature time series. The SOI index captures major features of interannual climate variability and is used extensively in its prediction. The other time series cover interdecadal and millennial time scales. The second category includes the calculation of fractional dimension, leading Lyapunov exponents, and Hurst exponents. More recently, multi-trend analysis (MTA), binary-decomposition analysis (BDA), and related methods have attempted to describe the structure of time series that include both regular and irregular components. Within the time available, I will try to give a feeling for how these methods work, and how well.
Evaluation of a Multivariate Syndromic Surveillance System for West Nile Virus.
Faverjon, Céline; Andersson, M Gunnar; Decors, Anouk; Tapprest, Jackie; Tritz, Pierre; Sandoz, Alain; Kutasi, Orsolya; Sala, Carole; Leblond, Agnès
2016-06-01
Various methods are currently used for the early detection of West Nile virus (WNV) but their outputs are not quantitative and/or do not take into account all available information. Our study aimed to test a multivariate syndromic surveillance system to evaluate if the sensitivity and the specificity of detection of WNV could be improved. Weekly time series data on nervous syndromes in horses and mortality in both horses and wild birds were used. Baselines were fitted to the three time series and used to simulate 100 years of surveillance data. WNV outbreaks were simulated and inserted into the baselines based on historical data and expert opinion. Univariate and multivariate syndromic surveillance systems were tested to gauge how well they detected the outbreaks; detection was based on an empirical Bayesian approach. The systems' performances were compared using measures of sensitivity, specificity, and area under receiver operating characteristic curve (AUC). When data sources were considered separately (i.e., univariate systems), the best detection performance was obtained using the data set of nervous symptoms in horses compared to those of bird and horse mortality (AUCs equal to 0.80, 0.75, and 0.50, respectively). A multivariate outbreak detection system that used nervous symptoms in horses and bird mortality generated the best performance (AUC = 0.87). The proposed approach is suitable for performing multivariate syndromic surveillance of WNV outbreaks. This is particularly relevant, given that a multivariate surveillance system performed better than a univariate approach. Such a surveillance system could be especially useful in serving as an alert for the possibility of human viral infections. This approach can be also used for other diseases for which multiple sources of evidence are available.
NASA Astrophysics Data System (ADS)
Sardinha-Lourenço, A.; Andrade-Campos, A.; Antunes, A.; Oliveira, M. S.
2018-03-01
Recent research on water demand short-term forecasting has shown that models using univariate time series based on historical data are useful and can be combined with other prediction methods to reduce errors. The behavior of water demands in drinking water distribution networks focuses on their repetitive nature and, under meteorological conditions and similar consumers, allows the development of a heuristic forecast model that, in turn, combined with other autoregressive models, can provide reliable forecasts. In this study, a parallel adaptive weighting strategy of water consumption forecast for the next 24-48 h, using univariate time series of potable water consumption, is proposed. Two Portuguese potable water distribution networks are used as case studies where the only input data are the consumption of water and the national calendar. For the development of the strategy, the Autoregressive Integrated Moving Average (ARIMA) method and a short-term forecast heuristic algorithm are used. Simulations with the model showed that, when using a parallel adaptive weighting strategy, the prediction error can be reduced by 15.96% and the average error by 9.20%. This reduction is important in the control and management of water supply systems. The proposed methodology can be extended to other forecast methods, especially when it comes to the availability of multiple forecast models.
Malik, Nishant; Marwan, Norbert; Zou, Yong; Mucha, Peter J.; Kurths, Jürgen
2016-01-01
A method to identify distinct dynamical regimes and transitions between those regimes in a short univariate time series was recently introduced [1], employing the computation of fluctuations in a measure of nonlinear similarity based on local recurrence properties. In the present work, we describe the details of the analytical relationships between this newly introduced measure and the well known concepts of attractor dimensions and Lyapunov exponents. We show that the new measure has linear dependence on the effective dimension of the attractor and it measures the variations in the sum of the Lyapunov spectrum. To illustrate the practical usefulness of the method, we identify various types of dynamical transitions in different nonlinear models. We present testbed examples for the new method’s robustness against noise and missing values in the time series. We also use this method to analyze time series of social dynamics, specifically an analysis of the U.S. crime record time series from 1975 to 1993. Using this method, we find that dynamical complexity in robberies was influenced by the unemployment rate until the late 1980’s. We have also observed a dynamical transition in homicide and robbery rates in the late 1980’s and early 1990’s, leading to increase in the dynamical complexity of these rates. PMID:25019852
What does the structure of its visibility graph tell us about the nature of the time series?
NASA Astrophysics Data System (ADS)
Franke, Jasper G.; Donner, Reik V.
2017-04-01
Visibility graphs are a recently introduced method to construct complex network representations based upon univariate time series in order to study their dynamical characteristics [1]. In the last years, this approach has been successfully applied to studying a considerable variety of geoscientific research questions and data sets, including non-trivial temporal patterns in complex earthquake catalogs [2] or time-reversibility in climate time series [3]. It has been shown that several characteristic features of the thus constructed networks differ between stochastic and deterministic (possibly chaotic) processes, which is, however, relatively hard to exploit in the case of real-world applications. In this study, we propose studying two new measures related with the network complexity of visibility graphs constructed from time series, one being a special type of network entropy [4] and the other a recently introduced measure of the heterogeneity of the network's degree distribution [5]. For paradigmatic model systems exhibiting bifurcation sequences between regular and chaotic dynamics, both properties clearly trace the transitions between both types of regimes and exhibit marked quantitative differences for regular and chaotic dynamics. Moreover, for dynamical systems with a small amount of additive noise, the considered properties demonstrate gradual changes prior to the bifurcation point. This finding appears closely related to the subsequent loss of stability of the current state known to lead to a critical slowing down as the transition point is approaches. In this spirit, both considered visibility graph characteristics provide alternative tracers of dynamical early warning signals consistent with classical indicators. Our results demonstrate that measures of visibility graph complexity (i) provide a potentially useful means to tracing changes in the dynamical patterns encoded in a univariate time series that originate from increasing autocorrelation and (ii) allow to systematically distinguish regular from deterministic-chaotic dynamics. We demonstrate the application of our method for different model systems as well as selected paleoclimate time series from the North Atlantic region. Notably, visibility graph based methods are particularly suited for studying the latter type of geoscientific data, since they do not impose intrinsic restrictions or assumptions on the nature of the time series under investigation in terms of noise process, linearity and sampling homogeneity. [1] Lacasa, Lucas, et al. "From time series to complex networks: The visibility graph." Proceedings of the National Academy of Sciences 105.13 (2008): 4972-4975. [2] Telesca, Luciano, and Michele Lovallo. "Analysis of seismic sequences by using the method of visibility graph." EPL (Europhysics Letters) 97.5 (2012): 50002. [3] Donges, Jonathan F., Reik V. Donner, and Jürgen Kurths. "Testing time series irreversibility using complex network methods." EPL (Europhysics Letters) 102.1 (2013): 10004. [4] Small, Michael. "Complex networks from time series: capturing dynamics." 2013 IEEE International Symposium on Circuits and Systems (ISCAS2013), Beijing (2013): 2509-2512. [5] Jacob, Rinku, K.P. Harikrishnan, Ranjeev Misra, and G. Ambika. "Measure for degree heterogeneity in complex networks and its application to recurrence network analysis." arXiv preprint 1605.06607 (2016).
NASA Astrophysics Data System (ADS)
Gocheva-Ilieva, S.; Stoimenova, M.; Ivanov, A.; Voynikova, D.; Iliev, I.
2016-10-01
Fine particulate matter PM2.5 and PM10 air pollutants are a serious problem in many urban areas affecting both the health of the population and the environment as a whole. The availability of large data arrays for the levels of these pollutants makes it possible to perform statistical analysis, to obtain relevant information, and to find patterns within the data. Research in this field is particularly topical for a number of Bulgarian cities, European country, where in recent years regulatory air pollution health limits are constantly being exceeded. This paper examines average daily data for air pollution with PM2.5 and PM10, collected by 3 monitoring stations in the cities of Plovdiv and Asenovgrad between 2011 and 2016. The goal is to find and analyze actual relationships in data time series, to build adequate mathematical models, and to develop short-term forecasts. Modeling is carried out by stochastic univariate and multivariate time series analysis, based on Box-Jenkins methodology. The best models are selected following initial transformation of the data and using a set of standard and robust statistical criteria. The Mathematica and SPSS software were used to perform calculations. This examination showed measured concentrations of PM2.5 and PM10 in the region of Plovdiv and Asenovgrad regularly exceed permissible European and national health and safety thresholds. We obtained adequate stochastic models with high statistical fit with the data and good quality forecasting when compared against actual measurements. The mathematical approach applied provides an independent alternative to standard official monitoring and control means for air pollution in urban areas.
A multivariate time series approach to modeling and forecasting demand in the emergency department.
Jones, Spencer S; Evans, R Scott; Allen, Todd L; Thomas, Alun; Haug, Peter J; Welch, Shari J; Snow, Gregory L
2009-02-01
The goals of this investigation were to study the temporal relationships between the demands for key resources in the emergency department (ED) and the inpatient hospital, and to develop multivariate forecasting models. Hourly data were collected from three diverse hospitals for the year 2006. Descriptive analysis and model fitting were carried out using graphical and multivariate time series methods. Multivariate models were compared to a univariate benchmark model in terms of their ability to provide out-of-sample forecasts of ED census and the demands for diagnostic resources. Descriptive analyses revealed little temporal interaction between the demand for inpatient resources and the demand for ED resources at the facilities considered. Multivariate models provided more accurate forecasts of ED census and of the demands for diagnostic resources. Our results suggest that multivariate time series models can be used to reliably forecast ED patient census; however, forecasts of the demands for diagnostic resources were not sufficiently reliable to be useful in the clinical setting.
The promise of the state space approach to time series analysis for nursing research.
Levy, Janet A; Elser, Heather E; Knobel, Robin B
2012-01-01
Nursing research, particularly related to physiological development, often depends on the collection of time series data. The state space approach to time series analysis has great potential to answer exploratory questions relevant to physiological development but has not been used extensively in nursing. The aim of the study was to introduce the state space approach to time series analysis and demonstrate potential applicability to neonatal monitoring and physiology. We present a set of univariate state space models; each one describing a process that generates a variable of interest over time. Each model is presented algebraically and a realization of the process is presented graphically from simulated data. This is followed by a discussion of how the model has been or may be used in two nursing projects on neonatal physiological development. The defining feature of the state space approach is the decomposition of the series into components that are functions of time; specifically, slowly varying level, faster varying periodic, and irregular components. State space models potentially simulate developmental processes where a phenomenon emerges and disappears before stabilizing, where the periodic component may become more regular with time, or where the developmental trajectory of a phenomenon is irregular. The ultimate contribution of this approach to nursing science will require close collaboration and cross-disciplinary education between nurses and statisticians.
Santori, G; Fontana, I; Bertocchi, M; Gasloli, G; Magoni Rossi, A; Tagliamacco, A; Barocci, S; Nocera, A; Valente, U
2010-05-01
A useful approach to reduce the number of discarded marginal kidneys and to increase the nephron mass is double kidney transplantation (DKT). In this study, we retrospectively evaluated the potential predictors for patient and graft survival in a single-center series of 59 DKT procedures performed between April 21, 1999, and September 21, 2008. The kidney recipients of mean age 63.27 +/- 5.17 years included 16 women (27%) and 43 men (73%). The donors of mean age 69.54 +/- 7.48 years included 32 women (54%) and 27 men (46%). The mean posttransplant dialysis time was 2.37 +/- 3.61 days. The mean hospitalization was 20.12 +/- 13.65 days. Average serum creatinine (SCr) at discharge was 1.5 +/- 0.59 mg/dL. In view of the limited numbers of recipient deaths (n = 4) and graft losses (n = 8) that occurred in our series, the proportional hazards assumption for each Cox regression model with P < .05 was tested by using correlation coefficients between transformed survival times and scaled Schoenfeld residuals, and checked with smoothed plots of Schoenfeld residuals. For patient survival, the variables that reached statistical significance were donor SCr (P = .007), donor creatinine cleararance (P = .023), and recipient age (P = .047). Each significant model passed the Schoenfeld test. By entering these variables into a multivariate Cox model for patient survival, no further significance was observed. In the univariate Cox models performed for graft survival, statistical significance was noted for donor SCr (P = .027), SCr 3 months post-DKT (P = .043), and SCr 6 months post-DKT (P = .017). All significant univariate models for graft survival passed the Schoenfeld test. A final multivariate model retained SCr at 6 months (beta = 1.746, P = .042) and donor SCr (beta = .767, P = .090). In our analysis, SCr at 6 months seemed to emerge from both univariate and multivariate Cox models as a potential predictor of graft survival among DKT. Multicenter studies with larger recipient populations and more graft losses should be performed to confirm our findings. Copyright (c) 2010 Elsevier Inc. All rights reserved.
Nonlinear Prediction Model for Hydrologic Time Series Based on Wavelet Decomposition
NASA Astrophysics Data System (ADS)
Kwon, H.; Khalil, A.; Brown, C.; Lall, U.; Ahn, H.; Moon, Y.
2005-12-01
Traditionally forecasting and characterizations of hydrologic systems is performed utilizing many techniques. Stochastic linear methods such as AR and ARIMA and nonlinear ones such as statistical learning theory based tools have been extensively used. The common difficulty to all methods is the determination of sufficient and necessary information and predictors for a successful prediction. Relationships between hydrologic variables are often highly nonlinear and interrelated across the temporal scale. A new hybrid approach is proposed for the simulation of hydrologic time series combining both the wavelet transform and the nonlinear model. The present model employs some merits of wavelet transform and nonlinear time series model. The Wavelet Transform is adopted to decompose a hydrologic nonlinear process into a set of mono-component signals, which are simulated by nonlinear model. The hybrid methodology is formulated in a manner to improve the accuracy of a long term forecasting. The proposed hybrid model yields much better results in terms of capturing and reproducing the time-frequency properties of the system at hand. Prediction results are promising when compared to traditional univariate time series models. An application of the plausibility of the proposed methodology is provided and the results conclude that wavelet based time series model can be utilized for simulating and forecasting of hydrologic variable reasonably well. This will ultimately serve the purpose of integrated water resources planning and management.
A Nonlinear Dynamical Systems based Model for Stochastic Simulation of Streamflow
NASA Astrophysics Data System (ADS)
Erkyihun, S. T.; Rajagopalan, B.; Zagona, E. A.
2014-12-01
Traditional time series methods model the evolution of the underlying process as a linear or nonlinear function of the autocorrelation. These methods capture the distributional statistics but are incapable of providing insights into the dynamics of the process, the potential regimes, and predictability. This work develops a nonlinear dynamical model for stochastic simulation of streamflows. In this, first a wavelet spectral analysis is employed on the flow series to isolate dominant orthogonal quasi periodic timeseries components. The periodic bands are added denoting the 'signal' component of the time series and the residual being the 'noise' component. Next, the underlying nonlinear dynamics of this combined band time series is recovered. For this the univariate time series is embedded in a d-dimensional space with an appropriate lag T to recover the state space in which the dynamics unfolds. Predictability is assessed by quantifying the divergence of trajectories in the state space with time, as Lyapunov exponents. The nonlinear dynamics in conjunction with a K-nearest neighbor time resampling is used to simulate the combined band, to which the noise component is added to simulate the timeseries. We demonstrate this method by applying it to the data at Lees Ferry that comprises of both the paleo reconstructed and naturalized historic annual flow spanning 1490-2010. We identify interesting dynamics of the signal in the flow series and epochal behavior of predictability. These will be of immense use for water resources planning and management.
Time-Series Forecast Modeling on High-Bandwidth Network Measurements
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yoo, Wucherl; Sim, Alex
With the increasing number of geographically distributed scientific collaborations and the growing sizes of scientific data, it has become challenging for users to achieve the best possible network performance on a shared network. In this paper, we have developed a model to forecast expected bandwidth utilization on high-bandwidth wide area networks. The forecast model can improve the efficiency of the resource utilization and scheduling of data movements on high-bandwidth networks to accommodate ever increasing data volume for large-scale scientific data applications. A univariate time-series forecast model is developed with the Seasonal decomposition of Time series by Loess (STL) and themore » AutoRegressive Integrated Moving Average (ARIMA) on Simple Network Management Protocol (SNMP) path utilization measurement data. Compared with the traditional approach such as Box-Jenkins methodology to train the ARIMA model, our forecast model reduces computation time up to 92.6 %. It also shows resilience against abrupt network usage changes. Finally, our forecast model conducts the large number of multi-step forecast, and the forecast errors are within the mean absolute deviation (MAD) of the monitored measurements.« less
Time-Series Forecast Modeling on High-Bandwidth Network Measurements
Yoo, Wucherl; Sim, Alex
2016-06-24
With the increasing number of geographically distributed scientific collaborations and the growing sizes of scientific data, it has become challenging for users to achieve the best possible network performance on a shared network. In this paper, we have developed a model to forecast expected bandwidth utilization on high-bandwidth wide area networks. The forecast model can improve the efficiency of the resource utilization and scheduling of data movements on high-bandwidth networks to accommodate ever increasing data volume for large-scale scientific data applications. A univariate time-series forecast model is developed with the Seasonal decomposition of Time series by Loess (STL) and themore » AutoRegressive Integrated Moving Average (ARIMA) on Simple Network Management Protocol (SNMP) path utilization measurement data. Compared with the traditional approach such as Box-Jenkins methodology to train the ARIMA model, our forecast model reduces computation time up to 92.6 %. It also shows resilience against abrupt network usage changes. Finally, our forecast model conducts the large number of multi-step forecast, and the forecast errors are within the mean absolute deviation (MAD) of the monitored measurements.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Loveday, D.L.; Craggs, C.
Univariate stochastic modeling, using Box-Jenkins methods, is carried out for three air temperatures which can influence the performance of a solar-assisted heat pump system. In this system, external ambient air (the low grade source) is pre-heated by the conventional tiled roof of an occupied domestic residence. The air then crosses the evaporator of an electrically driven split heat pump which is situated in the roof space. Autocorrelation coefficients are presented for time series of the following dry-bulb temperatures: the external air, the residence internal (lounge) air, and the air in the roofspace after pre-heating but prior to crossing the heatmore » pump evaporator. Hourly data relating to a two-week period in the heating season was utilized, providing a 336-h dataset. Univariate models fitted to the first 300 observations were validated by forecasting ahead for the remaining 36 h in steps of 1 h. Comparison of forecasted and measured values showed good agreement, except for a 4-h period in which the intensity of solar radiation exceeded that which prevailed during the modeled period. It is concluded that the Box-Jenkins approach can be used to develop univariate mathematical models which adequately describe building and climate thermal behavior, and that the importance of solar radiation in this respect should not be overlooked.« less
NASA Astrophysics Data System (ADS)
Malik, Nishant; Marwan, Norbert; Zou, Yong; Mucha, Peter J.; Kurths, Jürgen
2014-06-01
A method to identify distinct dynamical regimes and transitions between those regimes in a short univariate time series was recently introduced [N. Malik et al., Europhys. Lett. 97, 40009 (2012), 10.1209/0295-5075/97/40009], employing the computation of fluctuations in a measure of nonlinear similarity based on local recurrence properties. In this work, we describe the details of the analytical relationships between this newly introduced measure and the well-known concepts of attractor dimensions and Lyapunov exponents. We show that the new measure has linear dependence on the effective dimension of the attractor and it measures the variations in the sum of the Lyapunov spectrum. To illustrate the practical usefulness of the method, we identify various types of dynamical transitions in different nonlinear models. We present testbed examples for the new method's robustness against noise and missing values in the time series. We also use this method to analyze time series of social dynamics, specifically an analysis of the US crime record time series from 1975 to 1993. Using this method, we find that dynamical complexity in robberies was influenced by the unemployment rate until the late 1980s. We have also observed a dynamical transition in homicide and robbery rates in the late 1980s and early 1990s, leading to increase in the dynamical complexity of these rates.
Monograph on the use of the multivariate Gram Charlier series Type A
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hatayodom, T.; Heydt, G.
1978-01-01
The Gram-Charlier series in an infinite series expansion for a probability density function (pdf) in which terms of the series are Hermite polynomials. There are several Gram-Charlier series - the best known is Type A. The Gram-Charlier series, Type A (GCA) exists for both univariate and multivariate random variables. This monograph introduces the multivariate GCA and illustrates its use through several examples. A brief bibliography and discussion of Hermite polynomials is also included. 9 figures, 2 tables.
Dependency structure and scaling properties of financial time series are related
Morales, Raffaello; Di Matteo, T.; Aste, Tomaso
2014-01-01
We report evidence of a deep interplay between cross-correlations hierarchical properties and multifractality of New York Stock Exchange daily stock returns. The degree of multifractality displayed by different stocks is found to be positively correlated to their depth in the hierarchy of cross-correlations. We propose a dynamical model that reproduces this observation along with an array of other empirical properties. The structure of this model is such that the hierarchical structure of heterogeneous risks plays a crucial role in the time evolution of the correlation matrix, providing an interpretation to the mechanism behind the interplay between cross-correlation and multifractality in financial markets, where the degree of multifractality of stocks is associated to their hierarchical positioning in the cross-correlation structure. Empirical observations reported in this paper present a new perspective towards the merging of univariate multi scaling and multivariate cross-correlation properties of financial time series. PMID:24699417
Dependency structure and scaling properties of financial time series are related
NASA Astrophysics Data System (ADS)
Morales, Raffaello; Di Matteo, T.; Aste, Tomaso
2014-04-01
We report evidence of a deep interplay between cross-correlations hierarchical properties and multifractality of New York Stock Exchange daily stock returns. The degree of multifractality displayed by different stocks is found to be positively correlated to their depth in the hierarchy of cross-correlations. We propose a dynamical model that reproduces this observation along with an array of other empirical properties. The structure of this model is such that the hierarchical structure of heterogeneous risks plays a crucial role in the time evolution of the correlation matrix, providing an interpretation to the mechanism behind the interplay between cross-correlation and multifractality in financial markets, where the degree of multifractality of stocks is associated to their hierarchical positioning in the cross-correlation structure. Empirical observations reported in this paper present a new perspective towards the merging of univariate multi scaling and multivariate cross-correlation properties of financial time series.
Multivariate exploration of non-intrusive load monitoring via spatiotemporal pattern network
Liu, Chao; Akintayo, Adedotun; Jiang, Zhanhong; ...
2017-12-18
Non-intrusive load monitoring (NILM) of electrical demand for the purpose of identifying load components has thus far mostly been studied using univariate data, e.g., using only whole building electricity consumption time series to identify a certain type of end-use such as lighting load. However, using additional variables in the form of multivariate time series data may provide more information in terms of extracting distinguishable features in the context of energy disaggregation. In this work, a novel probabilistic graphical modeling approach, namely the spatiotemporal pattern network (STPN) is proposed for energy disaggregation using multivariate time-series data. The STPN framework is shownmore » to be capable of handling diverse types of multivariate time-series to improve the energy disaggregation performance. The technique outperforms the state of the art factorial hidden Markov models (FHMM) and combinatorial optimization (CO) techniques in multiple real-life test cases. Furthermore, based on two homes' aggregate electric consumption data, a similarity metric is defined for the energy disaggregation of one home using a trained model based on the other home (i.e., out-of-sample case). The proposed similarity metric allows us to enhance scalability via learning supervised models for a few homes and deploying such models to many other similar but unmodeled homes with significantly high disaggregation accuracy.« less
Multivariate exploration of non-intrusive load monitoring via spatiotemporal pattern network
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Chao; Akintayo, Adedotun; Jiang, Zhanhong
Non-intrusive load monitoring (NILM) of electrical demand for the purpose of identifying load components has thus far mostly been studied using univariate data, e.g., using only whole building electricity consumption time series to identify a certain type of end-use such as lighting load. However, using additional variables in the form of multivariate time series data may provide more information in terms of extracting distinguishable features in the context of energy disaggregation. In this work, a novel probabilistic graphical modeling approach, namely the spatiotemporal pattern network (STPN) is proposed for energy disaggregation using multivariate time-series data. The STPN framework is shownmore » to be capable of handling diverse types of multivariate time-series to improve the energy disaggregation performance. The technique outperforms the state of the art factorial hidden Markov models (FHMM) and combinatorial optimization (CO) techniques in multiple real-life test cases. Furthermore, based on two homes' aggregate electric consumption data, a similarity metric is defined for the energy disaggregation of one home using a trained model based on the other home (i.e., out-of-sample case). The proposed similarity metric allows us to enhance scalability via learning supervised models for a few homes and deploying such models to many other similar but unmodeled homes with significantly high disaggregation accuracy.« less
Univariate Time Series Prediction of Solar Power Using a Hybrid Wavelet-ARMA-NARX Prediction Method
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nazaripouya, Hamidreza; Wang, Yubo; Chu, Chi-Cheng
This paper proposes a new hybrid method for super short-term solar power prediction. Solar output power usually has a complex, nonstationary, and nonlinear characteristic due to intermittent and time varying behavior of solar radiance. In addition, solar power dynamics is fast and is inertia less. An accurate super short-time prediction is required to compensate for the fluctuations and reduce the impact of solar power penetration on the power system. The objective is to predict one step-ahead solar power generation based only on historical solar power time series data. The proposed method incorporates discrete wavelet transform (DWT), Auto-Regressive Moving Average (ARMA)more » models, and Recurrent Neural Networks (RNN), while the RNN architecture is based on Nonlinear Auto-Regressive models with eXogenous inputs (NARX). The wavelet transform is utilized to decompose the solar power time series into a set of richer-behaved forming series for prediction. ARMA model is employed as a linear predictor while NARX is used as a nonlinear pattern recognition tool to estimate and compensate the error of wavelet-ARMA prediction. The proposed method is applied to the data captured from UCLA solar PV panels and the results are compared with some of the common and most recent solar power prediction methods. The results validate the effectiveness of the proposed approach and show a considerable improvement in the prediction precision.« less
Performance of time-series methods in forecasting the demand for red blood cell transfusion.
Pereira, Arturo
2004-05-01
Planning the future blood collection efforts must be based on adequate forecasts of transfusion demand. In this study, univariate time-series methods were investigated for their performance in forecasting the monthly demand for RBCs at one tertiary-care, university hospital. Three time-series methods were investigated: autoregressive integrated moving average (ARIMA), the Holt-Winters family of exponential smoothing models, and one neural-network-based method. The time series consisted of the monthly demand for RBCs from January 1988 to December 2002 and was divided into two segments: the older one was used to fit or train the models, and the younger to test for the accuracy of predictions. Performance was compared across forecasting methods by calculating goodness-of-fit statistics, the percentage of months in which forecast-based supply would have met the RBC demand (coverage rate), and the outdate rate. The RBC transfusion series was best fitted by a seasonal ARIMA(0,1,1)(0,1,1)(12) model. Over 1-year time horizons, forecasts generated by ARIMA or exponential smoothing laid within the +/- 10 percent interval of the real RBC demand in 79 percent of months (62% in the case of neural networks). The coverage rate for the three methods was 89, 91, and 86 percent, respectively. Over 2-year time horizons, exponential smoothing largely outperformed the other methods. Predictions by exponential smoothing laid within the +/- 10 percent interval of real values in 75 percent of the 24 forecasted months, and the coverage rate was 87 percent. Over 1-year time horizons, predictions of RBC demand generated by ARIMA or exponential smoothing are accurate enough to be of help in the planning of blood collection efforts. For longer time horizons, exponential smoothing outperforms the other forecasting methods.
Effect of correlation on covariate selection in linear and nonlinear mixed effect models.
Bonate, Peter L
2017-01-01
The effect of correlation among covariates on covariate selection was examined with linear and nonlinear mixed effect models. Demographic covariates were extracted from the National Health and Nutrition Examination Survey III database. Concentration-time profiles were Monte Carlo simulated where only one covariate affected apparent oral clearance (CL/F). A series of univariate covariate population pharmacokinetic models was fit to the data and compared with the reduced model without covariate. The "best" covariate was identified using either the likelihood ratio test statistic or AIC. Weight and body surface area (calculated using Gehan and George equation, 1970) were highly correlated (r = 0.98). Body surface area was often selected as a better covariate than weight, sometimes as high as 1 in 5 times, when weight was the covariate used in the data generating mechanism. In a second simulation, parent drug concentration and three metabolites were simulated from a thorough QT study and used as covariates in a series of univariate linear mixed effects models of ddQTc interval prolongation. The covariate with the largest significant LRT statistic was deemed the "best" predictor. When the metabolite was formation-rate limited and only parent concentrations affected ddQTc intervals the metabolite was chosen as a better predictor as often as 1 in 5 times depending on the slope of the relationship between parent concentrations and ddQTc intervals. A correlated covariate can be chosen as being a better predictor than another covariate in a linear or nonlinear population analysis by sheer correlation These results explain why for the same drug different covariates may be identified in different analyses. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Levine, Matthew E; Albers, David J; Hripcsak, George
2016-01-01
Time series analysis methods have been shown to reveal clinical and biological associations in data collected in the electronic health record. We wish to develop reliable high-throughput methods for identifying adverse drug effects that are easy to implement and produce readily interpretable results. To move toward this goal, we used univariate and multivariate lagged regression models to investigate associations between twenty pairs of drug orders and laboratory measurements. Multivariate lagged regression models exhibited higher sensitivity and specificity than univariate lagged regression in the 20 examples, and incorporating autoregressive terms for labs and drugs produced more robust signals in cases of known associations among the 20 example pairings. Moreover, including inpatient admission terms in the model attenuated the signals for some cases of unlikely associations, demonstrating how multivariate lagged regression models' explicit handling of context-based variables can provide a simple way to probe for health-care processes that confound analyses of EHR data.
NASA Astrophysics Data System (ADS)
Papacharalampous, Georgia; Tyralis, Hristos; Koutsoyiannis, Demetris
2018-02-01
We investigate the predictability of monthly temperature and precipitation by applying automatic univariate time series forecasting methods to a sample of 985 40-year-long monthly temperature and 1552 40-year-long monthly precipitation time series. The methods include a naïve one based on the monthly values of the last year, as well as the random walk (with drift), AutoRegressive Fractionally Integrated Moving Average (ARFIMA), exponential smoothing state-space model with Box-Cox transformation, ARMA errors, Trend and Seasonal components (BATS), simple exponential smoothing, Theta and Prophet methods. Prophet is a recently introduced model inspired by the nature of time series forecasted at Facebook and has not been applied to hydrometeorological time series before, while the use of random walk, BATS, simple exponential smoothing and Theta is rare in hydrology. The methods are tested in performing multi-step ahead forecasts for the last 48 months of the data. We further investigate how different choices of handling the seasonality and non-normality affect the performance of the models. The results indicate that: (a) all the examined methods apart from the naïve and random walk ones are accurate enough to be used in long-term applications; (b) monthly temperature and precipitation can be forecasted to a level of accuracy which can barely be improved using other methods; (c) the externally applied classical seasonal decomposition results mostly in better forecasts compared to the automatic seasonal decomposition used by the BATS and Prophet methods; and (d) Prophet is competitive, especially when it is combined with externally applied classical seasonal decomposition.
Morrison, Kathryn T; Shaddick, Gavin; Henderson, Sarah B; Buckeridge, David L
2016-08-15
This paper outlines a latent process model for forecasting multiple health outcomes arising from a common environmental exposure. Traditionally, surveillance models in environmental health do not link health outcome measures, such as morbidity or mortality counts, to measures of exposure, such as air pollution. Moreover, different measures of health outcomes are treated as independent, while it is known that they are correlated with one another over time as they arise in part from a common underlying exposure. We propose modelling an environmental exposure as a latent process, and we describe the implementation of such a model within a hierarchical Bayesian framework and its efficient computation using integrated nested Laplace approximations. Through a simulation study, we compare distinct univariate models for each health outcome with a bivariate approach. The bivariate model outperforms the univariate models in bias and coverage of parameter estimation, in forecast accuracy and in computational efficiency. The methods are illustrated with a case study using healthcare utilization and air pollution data from British Columbia, Canada, 2003-2011, where seasonal wildfires produce high levels of air pollution, significantly impacting population health. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Adam Smith in the Mathematics Classroom
ERIC Educational Resources Information Center
Lipsey, Sally I.
1975-01-01
The author describes a series of current economic ideas and situations which can be used in the mathematics classroom to illustrate the use of signed numbers, the coordinate system, univariate and multivariate functions, linear programing, and variation. (SD)
Time series modelling of global mean temperature for managerial decision-making.
Romilly, Peter
2005-07-01
Climate change has important implications for business and economic activity. Effective management of climate change impacts will depend on the availability of accurate and cost-effective forecasts. This paper uses univariate time series techniques to model the properties of a global mean temperature dataset in order to develop a parsimonious forecasting model for managerial decision-making over the short-term horizon. Although the model is estimated on global temperature data, the methodology could also be applied to temperature data at more localised levels. The statistical techniques include seasonal and non-seasonal unit root testing with and without structural breaks, as well as ARIMA and GARCH modelling. A forecasting evaluation shows that the chosen model performs well against rival models. The estimation results confirm the findings of a number of previous studies, namely that global mean temperatures increased significantly throughout the 20th century. The use of GARCH modelling also shows the presence of volatility clustering in the temperature data, and a positive association between volatility and global mean temperature.
Prognostic importance of DNA ploidy in non-endometrioid, high-risk endometrial carcinomas.
Sorbe, Bengt
2016-03-01
The present study investigated the predictive and prognostic impact of DNA ploidy together with other well-known prognostic factors in a series of non-endometrioid, high-risk endometrial carcinomas. From a complete consecutive series of 4,543 endometrial carcinomas of International Federation of Gynecology and Obstetrics (FIGO) stages I-IV, 94 serous carcinomas, 48 clear cell carcinomas and 231 carcinosarcomas were selected as a non-endometrioid, high-risk group for further studies regarding prognosis. The impact of DNA ploidy, as assessed by flow cytometry, was of particular focus. The age of the patients, FIGO stage, depth of myometrial infiltration and tumor expression of p53 were also included in the analyses (univariate and multivariate). In the complete series of cases, the recurrence rate was 37%, and the 5-year overall survival rate was 39% with no difference between the three histological subtypes. The primary cure rate (78%) was also similar for all tumor types studied. DNA ploidy was a significant predictive factor (on univariate analysis) for primary tumor cure rate, and a prognostic factor for survival rate (on univariate and multivariate analyses). The predictive and prognostic impact of DNA ploidy was higher in carcinosarcomas than in serous and clear cell carcinomas. In the majority of multivariate analyses, FIGO stage and depth of myometrial infiltration were the most important predictive (tumor recurrence) and prognostic (survival rate) factors. DNA ploidy status is a less important predictive and prognostic factor in non-endometrioid, high-risk endometrial carcinomas than in the common endometrioid carcinomas, in which FIGO and nuclear grade also are highly significant and important factors.
Hot spots of multivariate extreme anomalies in Earth observations
NASA Astrophysics Data System (ADS)
Flach, M.; Sippel, S.; Bodesheim, P.; Brenning, A.; Denzler, J.; Gans, F.; Guanche, Y.; Reichstein, M.; Rodner, E.; Mahecha, M. D.
2016-12-01
Anomalies in Earth observations might indicate data quality issues, extremes or the change of underlying processes within a highly multivariate system. Thus, considering the multivariate constellation of variables for extreme detection yields crucial additional information over conventional univariate approaches. We highlight areas in which multivariate extreme anomalies are more likely to occur, i.e. hot spots of extremes in global atmospheric Earth observations that impact the Biosphere. In addition, we present the year of the most unusual multivariate extreme between 2001 and 2013 and show that these coincide with well known high impact extremes. Technically speaking, we account for multivariate extremes by using three sophisticated algorithms adapted from computer science applications. Namely an ensemble of the k-nearest neighbours mean distance, a kernel density estimation and an approach based on recurrences is used. However, the impact of atmosphere extremes on the Biosphere might largely depend on what is considered to be normal, i.e. the shape of the mean seasonal cycle and its inter-annual variability. We identify regions with similar mean seasonality by means of dimensionality reduction in order to estimate in each region both the `normal' variance and robust thresholds for detecting the extremes. In addition, we account for challenges like heteroscedasticity in Northern latitudes. Apart from hot spot areas, those anomalies in the atmosphere time series are of particular interest, which can only be detected by a multivariate approach but not by a simple univariate approach. Such an anomalous constellation of atmosphere variables is of interest if it impacts the Biosphere. The multivariate constellation of such an anomalous part of a time series is shown in one case study indicating that multivariate anomaly detection can provide novel insights into Earth observations.
Hanna, Andrew N; Datta, Jashodeep; Ginzberg, Sara; Dasher, Kevin; Ginsberg, Gregory G; Dempsey, Daniel T
2018-04-01
Although laparoscopic Heller myotomy (LHM) has been the standard of care for achalasia, per oral endoscopic myotomy (POEM) has gained popularity as a viable alternative. This retrospective study aimed to compare patient-reported outcomes between LHM and POEM in a consecutive series of achalasia patients with more than 1 year of follow-up. We reviewed demographic and procedure-related data for patients who underwent either LHM or POEM for achalasia between January 2011 and May 2016. Phone interviews were conducted assessing post-procedure achalasia symptoms via the Eckardt score and achalasia severity questionnaire (ASQ). Demographics, disease factors, and survey results were compared between LHM and POEM patients using univariate analysis. Significant predictors of procedure failure were analyzed using univariate and multivariate analysis. There were no serious complications in 110 consecutive patients who underwent LHM or POEM during the study period, and 96 (87%) patients completed phone surveys. There was a nonsignificant trend toward better patient-reported outcomes with POEM. There were significant differences in patient characteristics including sex, achalasia type, mean residual lower esophageal pressure (rLESP), and follow-up time. The only univariate predictors of an unsatisfactory Eckardt score or ASQ were longer follow-up and lower rLESP, with follow-up length being the only predictor on multivariate analysis. There were significant demographic and clinical differences in patient selection for POEM vs LHM in our group. Although the 2 procedures have similar patient-reported effectiveness, subjective outcomes seem to decline as a result of time rather than procedure type. Copyright © 2018 American College of Surgeons. Published by Elsevier Inc. All rights reserved.
Night of the sirens: analysis of carbon monoxide-detector experience in suburban Chicago.
Bizovi, K E; Leikin, J B; Hryhorczuk, D O; Frateschi, L J
1998-06-01
To determine the pattern and environmental causes of carbon monoxide (CO)-detector alarms. Data including time, location, detector manufacturer, CO measurements in the home, reported illness, cause, and actions taken were collected between July 15, 1994, and January 26, 1995, on all calls to 17 suburban Chicago fire departments for CO-detector alarms. We used univariate time-series analysis involving joint estimation of model parameters and outlier effects to analyze data and compared data on ambient CO levels from the Illinois Environmental Protection Agency to the number of calls per day. During the study period, 777 calls for sounding CO detectors were made to the fire departments in question. The median number of calls per day was three. Our univariate time series identified 3 days with a significant excess of calls (December 12, 29 calls; December 21, 69; December 22, 128; P < .001). The average ambient CO readings on these days were 0.99, 3.25, and 3.89 ppm, respectively, compared with an overall mean of 8.8 ppm. In-home CO levels among all 828 measurements taken from the 777 domestic calls ranged from 0 to 425 ppm, 0 in 249 (30%), 1 to 10 in 340 (41%), 11 to 50 in 149 (18%), 51 to 100 in 22 (9%), and more than 100 in 11 (1.3%). No measurement was taken in six cases. Cause of alarm was listed as furnace in 25 cases, auto exhaust in 24, stove/oven in 22, poor location of detector in 14, water heater in 11, outside sources in 7, and multiple sources in 7. Other sources accounted for fewer than 1% each. The participating fire departments considered 242 cases (31%) to be false alarms. Cause was not determined in 400 calls (51%). In 37 calls (4.8%), people reported illness. Above-average ambient CO levels coincided with a significant increase in the number of calls and may have contributed to the triggering of CO alarms.
Validation of a contemporary prostate cancer grading system using prostate cancer death as outcome.
Berney, Daniel M; Beltran, Luis; Fisher, Gabrielle; North, Bernard V; Greenberg, David; Møller, Henrik; Soosay, Geraldine; Scardino, Peter; Cuzick, Jack
2016-05-10
Gleason scoring (GS) has major deficiencies and a novel system of five grade groups (GS⩽6; 3+4; 4+3; 8; ⩾9) has been recently agreed and included in the WHO 2016 classification. Although verified in radical prostatectomies using PSA relapse for outcome, it has not been validated using prostate cancer death as an outcome in biopsy series. There is debate whether an 'overall' or 'worst' GS in biopsies series should be used. Nine hundred and eighty-eight prostate cancer biopsy cases were identified between 1990 and 2003, and treated conservatively. Diagnosis and grade was assigned to each core as well as an overall grade. Follow-up for prostate cancer death was until 31 December 2012. A log-rank test assessed univariable differences between the five grade groups based on overall and worst grade seen, and using univariable and multivariable Cox proportional hazards. Regression was used to quantify differences in outcome. Using both 'worst' and 'overall' GS yielded highly significant results on univariate and multivariate analysis with overall GS slightly but insignificantly outperforming worst GS. There was a strong correlation with the five grade groups and prostate cancer death. This is the largest conservatively treated prostate cancer cohort with long-term follow-up and contemporary assessment of grade. It validates the formation of five grade groups and suggests that the 'worst' grade is a valid prognostic measure.
Structural connectome topology relates to regional BOLD signal dynamics in the mouse brain
NASA Astrophysics Data System (ADS)
Sethi, Sarab S.; Zerbi, Valerio; Wenderoth, Nicole; Fornito, Alex; Fulcher, Ben D.
2017-04-01
Brain dynamics are thought to unfold on a network determined by the pattern of axonal connections linking pairs of neuronal elements; the so-called connectome. Prior work has indicated that structural brain connectivity constrains pairwise correlations of brain dynamics ("functional connectivity"), but it is not known whether inter-regional axonal connectivity is related to the intrinsic dynamics of individual brain areas. Here we investigate this relationship using a weighted, directed mesoscale mouse connectome from the Allen Mouse Brain Connectivity Atlas and resting state functional MRI (rs-fMRI) time-series data measured in 184 brain regions in eighteen anesthetized mice. For each brain region, we measured degree, betweenness, and clustering coefficient from weighted and unweighted, and directed and undirected versions of the connectome. We then characterized the univariate rs-fMRI dynamics in each brain region by computing 6930 time-series properties using the time-series analysis toolbox, hctsa. After correcting for regional volume variations, strong and robust correlations between structural connectivity properties and rs-fMRI dynamics were found only when edge weights were accounted for, and were associated with variations in the autocorrelation properties of the rs-fMRI signal. The strongest relationships were found for weighted in-degree, which was positively correlated to the autocorrelation of fMRI time series at time lag τ = 34 s (partial Spearman correlation ρ = 0.58 ), as well as a range of related measures such as relative high frequency power (f > 0.4 Hz: ρ = - 0.43 ). Our results indicate that the topology of inter-regional axonal connections of the mouse brain is closely related to intrinsic, spontaneous dynamics such that regions with a greater aggregate strength of incoming projections display longer timescales of activity fluctuations.
Method and data evaluation at NASA endocrine laboratory. [Skylab 3 experiments
NASA Technical Reports Server (NTRS)
Johnston, D. A.
1974-01-01
The biomedical data of the astronauts on Skylab 3 were analyzed to evaluate the univariate statistical methods for comparing endocrine series experiments in relation to other medical experiments. It was found that an information storage and retrieval system was needed to facilitate statistical analyses.
Drought Analysis for Kuwait Using Standardized Precipitation Index
2014-01-01
Implementation of adequate measures to assess and monitor droughts is recognized as a major matter challenging researchers involved in water resources management. The objective of this study is to assess the hydrologic drought characteristics from the historical rainfall records of Kuwait with arid environment by employing the criterion of Standardized Precipitation Index (SPI). A wide range of monthly total precipitation data from January 1967 to December 2009 is used for the assessment. The computation of the SPI series is performed for intermediate- and long-time scales of 3, 6, 12, and 24 months. The drought severity and duration are also estimated. The bivariate probability distribution for these two drought characteristics is constructed by using Clayton copula. It has been shown that the drought SPI series for the time scales examined have no systematic trend component but a seasonal pattern related to rainfall data. The results are used to perform univariate and bivariate frequency analyses for the drought events. The study will help evaluating the risk of future droughts in the region, assessing their consequences on economy, environment, and society, and adopting measures for mitigating the effect of droughts. PMID:25386598
Whomersley, P; Schratzberger, M; Huxham, M; Bates, H; Rees, H
2007-01-01
Sewage sludge was disposed of in Liverpool Bay for over 100 years. Annual amounts increased from 0.5 million tonnes per annum in 1900 to approximately 2 million tonnes per annum by 1995. Macrofauna and a suite of environmental variables were collected at a station adjacent to, and a reference station distant from, the disposal site over 13 years, spanning a pre- (1990-1998) and post- (1999-2003) cessation period. Univariate and multivariate analyses of the time-series data showed significant community differences between reference and disposal site stations and multivariate analyses revealed station-specific community development post-disposal. Temporal variability of communities collected at the disposal station post-cessation was higher than during years of disposal, when temporally stable dominance patterns of disturbance-tolerant species had established. Alterations of community structure post-disturbance reflected successional changes possibly driven by facilitation. Subtle faunistic changes at the Liverpool Bay disposal site indicate that the near-field effects of the disposal of sewage sludge were small and therefore could be considered environmentally acceptable.
Tobacco, Alcohol and Marijuana Use among First Year U.S. College Students: A time series analysis
Dierker, Lisa; Stolar, Marilyn; Richardson, Elizabeth; Tiffany, Stephen; Flay, Brian; Collins, Linda; Nichter, Mimi; Nichter, Mark; Bailey, Steffani; Clayton, Richard
2009-01-01
The present study sought to evaluate the day-to-day patterns of tobacco, alcohol and marijuana use among first year college students in the U.S. Using 210 days of weekly time-line follow-back diary data collected in 2002-2003, the authors examined within-person patterns of use. The sample was 48% female and 90% Caucasian. Sixty eight percent of the participants were permanent residents of Indiana. Univariate time series analysis was employed to evaluate behavioral trends for each substance across the academic year and to determine the predictive value of day-to-day substance use. Some of the most common trends included higher levels of substance use at the beginning and/or end of the academic year. Use on any given day could be predicted best from the amount of corresponding substance use one day prior. Conclusions While universal intervention might best be focused in the earliest weeks on campus and at the end of the year when substance use is at its highest, the diversity of substance use trajectories suggests the need for more targeted approaches to intervention. Study limitations are noted. PMID:18393083
Comparison of time series models for predicting campylobacteriosis risk in New Zealand.
Al-Sakkaf, A; Jones, G
2014-05-01
Predicting campylobacteriosis cases is a matter of considerable concern in New Zealand, after the number of the notified cases was the highest among the developed countries in 2006. Thus, there is a need to develop a model or a tool to predict accurately the number of campylobacteriosis cases as the Microbial Risk Assessment Model used to predict the number of campylobacteriosis cases failed to predict accurately the number of actual cases. We explore the appropriateness of classical time series modelling approaches for predicting campylobacteriosis. Finding the most appropriate time series model for New Zealand data has additional practical considerations given a possible structural change, that is, a specific and sudden change in response to the implemented interventions. A univariate methodological approach was used to predict monthly disease cases using New Zealand surveillance data of campylobacteriosis incidence from 1998 to 2009. The data from the years 1998 to 2008 were used to model the time series with the year 2009 held out of the data set for model validation. The best two models were then fitted to the full 1998-2009 data and used to predict for each month of 2010. The Holt-Winters (multiplicative) and ARIMA (additive) intervention models were considered the best models for predicting campylobacteriosis in New Zealand. It was noticed that the prediction by an additive ARIMA with intervention was slightly better than the prediction by a Holt-Winter multiplicative method for the annual total in year 2010, the former predicting only 23 cases less than the actual reported cases. It is confirmed that classical time series techniques such as ARIMA with intervention and Holt-Winters can provide a good prediction performance for campylobacteriosis risk in New Zealand. The results reported by this study are useful to the New Zealand Health and Safety Authority's efforts in addressing the problem of the campylobacteriosis epidemic. © 2013 Blackwell Verlag GmbH.
Reprint of: Relationship between cataract severity and socioeconomic status.
Wesolosky, Jason D; Rudnisky, Christopher J
2015-06-01
To determine the relationship between cataract severity and socioeconomic status (SES). Retrospective, observational case series. A total of 1350 eyes underwent phacoemulsification cataract extraction by a single surgeon using an Alcon Infiniti system. Cataract severity was measured using phaco time in seconds. SES was measured using area-level aggregate census data: median income, education, proportion of common-law couples, and employment rate. Preoperative best corrected visual acuity was obtained and converted to logarithm of the minimum angle of resolution values. For patients undergoing bilateral surgery, the generalized estimating equation was used to account for the correlation between eyes. Univariate analyses were performed using simple regression, and multivariate analyses were performed to account for variables with significant relationships (p < 0.05) on univariate testing. Sensitivity analyses were performed to assess the effect of including patient age in the controlled analyses. Multivariate analyses demonstrated that cataracts were more severe when the median income was lower (p = 0.001) and the proportion of common-law couples living in a patient's community (p = 0.012) and the unemployment rate (p = 0.002) were higher. These associations persisted even when controlling for patient age. Patients of lower SES have more severe cataracts. Copyright © 2015. Published by Elsevier Inc.
Validation of a contemporary prostate cancer grading system using prostate cancer death as outcome
Berney, Daniel M; Beltran, Luis; Fisher, Gabrielle; North, Bernard V; Greenberg, David; Møller, Henrik; Soosay, Geraldine; Scardino, Peter; Cuzick, Jack
2016-01-01
Background: Gleason scoring (GS) has major deficiencies and a novel system of five grade groups (GS⩽6; 3+4; 4+3; 8; ⩾9) has been recently agreed and included in the WHO 2016 classification. Although verified in radical prostatectomies using PSA relapse for outcome, it has not been validated using prostate cancer death as an outcome in biopsy series. There is debate whether an ‘overall' or ‘worst' GS in biopsies series should be used. Methods: Nine hundred and eighty-eight prostate cancer biopsy cases were identified between 1990 and 2003, and treated conservatively. Diagnosis and grade was assigned to each core as well as an overall grade. Follow-up for prostate cancer death was until 31 December 2012. A log-rank test assessed univariable differences between the five grade groups based on overall and worst grade seen, and using univariable and multivariable Cox proportional hazards. Regression was used to quantify differences in outcome. Results: Using both ‘worst' and ‘overall' GS yielded highly significant results on univariate and multivariate analysis with overall GS slightly but insignificantly outperforming worst GS. There was a strong correlation with the five grade groups and prostate cancer death. Conclusions: This is the largest conservatively treated prostate cancer cohort with long-term follow-up and contemporary assessment of grade. It validates the formation of five grade groups and suggests that the ‘worst' grade is a valid prognostic measure. PMID:27100731
Ramseyer, Fabian; Kupper, Zeno; Caspar, Franz; Znoj, Hansjörg; Tschacher, Wolfgang
2014-10-01
Processes occurring in the course of psychotherapy are characterized by the simple fact that they unfold in time and that the multiple factors engaged in change processes vary highly between individuals (idiographic phenomena). Previous research, however, has neglected the temporal perspective by its traditional focus on static phenomena, which were mainly assessed at the group level (nomothetic phenomena). To support a temporal approach, the authors introduce time-series panel analysis (TSPA), a statistical methodology explicitly focusing on the quantification of temporal, session-to-session aspects of change in psychotherapy. TSPA-models are initially built at the level of individuals and are subsequently aggregated at the group level, thus allowing the exploration of prototypical models. TSPA is based on vector auto-regression (VAR), an extension of univariate auto-regression models to multivariate time-series data. The application of TSPA is demonstrated in a sample of 87 outpatient psychotherapy patients who were monitored by postsession questionnaires. Prototypical mechanisms of change were derived from the aggregation of individual multivariate models of psychotherapy process. In a 2nd step, the associations between mechanisms of change (TSPA) and pre- to postsymptom change were explored. TSPA allowed a prototypical process pattern to be identified, where patient's alliance and self-efficacy were linked by a temporal feedback-loop. Furthermore, therapist's stability over time in both mastery and clarification interventions was positively associated with better outcomes. TSPA is a statistical tool that sheds new light on temporal mechanisms of change. Through this approach, clinicians may gain insight into prototypical patterns of change in psychotherapy. PsycINFO Database Record (c) 2014 APA, all rights reserved.
A new algorithm for automatic Outlier Detection in GPS Time Series
NASA Astrophysics Data System (ADS)
Cannavo', Flavio; Mattia, Mario; Rossi, Massimo; Palano, Mimmo; Bruno, Valentina
2010-05-01
Nowadays continuous GPS time series are considered a crucial product of GPS permanent networks, useful in many geo-science fields, such as active tectonics, seismology, crustal deformation and volcano monitoring (Altamimi et al. 2002, Elósegui et al. 2006, Aloisi et al. 2009). Although the GPS data elaboration software has increased in reliability, the time series are still affected by different kind of noise, from the intrinsic noise (e.g. thropospheric delay) to the un-modeled noise (e.g. cycle slips, satellite faults, parameters changing). Typically GPS Time Series present characteristic noise that is a linear combination of white noise and correlated colored noise, and this characteristic is fractal in the sense that is evident for every considered time scale or sampling rate. The un-modeled noise sources result in spikes, outliers and steps. These kind of errors can appreciably influence the estimation of velocities of the monitored sites. The outlier detection in generic time series is a widely treated problem in literature (Wei, 2005), while is not fully developed for the specific kind of GPS series. We propose a robust automatic procedure for cleaning the GPS time series from the outliers and, especially for long daily series, steps due to strong seismic or volcanic events or merely instrumentation changing such as antenna and receiver upgrades. The procedure is basically divided in two steps: a first step for the colored noise reduction and a second step for outlier detection through adaptive series segmentation. Both algorithms present novel ideas and are nearly unsupervised. In particular, we propose an algorithm to estimate an autoregressive model for colored noise in GPS time series in order to subtract the effect of non Gaussian noise on the series. This step is useful for the subsequent step (i.e. adaptive segmentation) which requires the hypothesis of Gaussian noise. The proposed algorithms are tested in a benchmark case study and the results confirm that the algorithms are effective and reasonable. Bibliography - Aloisi M., A. Bonaccorso, F. Cannavò, S. Gambino, M. Mattia, G. Puglisi, E. Boschi, A new dyke intrusion style for the Mount Etna May 2008 eruption modelled through continuous tilt and GPS data, Terra Nova, Volume 21 Issue 4 , Pages 316 - 321, doi: 10.1111/j.1365-3121.2009.00889.x (August 2009) - Altamimi Z., Sillard P., Boucher C., ITRF2000: A new release of the International Terrestrial Reference frame for earth science applications, J Geophys Res-Solid Earth, 107 (B10): art. no.-2214, (Oct 2002) - Elósegui, P., J. L. Davis, D. Oberlander, R. Baena, and G. Ekström , Accuracy of high-rate GPS for seismology, Geophys. Res. Lett., 33, L11308, doi:10.1029/2006GL026065 (2006) - Wei W. S., Time Series Analysis: Univariate and Multivariate Methods, Addison Wesley (2 edition), ISBN-10: 0321322169 (July, 2005)
On the Choice of Variable for Atmospheric Moisture Analysis
NASA Technical Reports Server (NTRS)
Dee, Dick P.; DaSilva, Arlindo M.; Atlas, Robert (Technical Monitor)
2002-01-01
The implications of using different control variables for the analysis of moisture observations in a global atmospheric data assimilation system are investigated. A moisture analysis based on either mixing ratio or specific humidity is prone to large extrapolation errors, due to the high variability in space and time of these parameters and to the difficulties in modeling their error covariances. Using the logarithm of specific humidity does not alleviate these problems, and has the further disadvantage that very dry background estimates cannot be effectively corrected by observations. Relative humidity is a better choice from a statistical point of view, because this field is spatially and temporally more coherent and error statistics are therefore easier to obtain. If, however, the analysis is designed to preserve relative humidity in the absence of moisture observations, then the analyzed specific humidity field depends entirely on analyzed temperature changes. If the model has a cool bias in the stratosphere this will lead to an unstable accumulation of excess moisture there. A pseudo-relative humidity can be defined by scaling the mixing ratio by the background saturation mixing ratio. A univariate pseudo-relative humidity analysis will preserve the specific humidity field in the absence of moisture observations. A pseudorelative humidity analysis is shown to be equivalent to a mixing ratio analysis with flow-dependent covariances. In the presence of multivariate (temperature-moisture) observations it produces analyzed relative humidity values that are nearly identical to those produced by a relative humidity analysis. Based on a time series analysis of radiosonde observed-minus-background differences it appears to be more justifiable to neglect specific humidity-temperature correlations (in a univariate pseudo-relative humidity analysis) than to neglect relative humidity-temperature correlations (in a univariate relative humidity analysis). A pseudo-relative humidity analysis is easily implemented in an existing moisture analysis system, by simply scaling observed-minus background moisture residuals prior to solving the analysis equation, and rescaling the analyzed increments afterward.
Category-Specific Comparison of Univariate Alerting Methods for Biosurveillance Decision Support
Elbert, Yevgeniy; Hung, Vivian; Burkom, Howard
2013-01-01
Objective For a multi-source decision support application, we sought to match univariate alerting algorithms to surveillance data types to optimize detection performance. Introduction Temporal alerting algorithms commonly used in syndromic surveillance systems are often adjusted for data features such as cyclic behavior but are subject to overfitting or misspecification errors when applied indiscriminately. In a project for the Armed Forces Health Surveillance Center to enable multivariate decision support, we obtained 4.5 years of out-patient, prescription and laboratory test records from all US military treatment facilities. A proof-of-concept project phase produced 16 events with multiple evidence corroboration for comparison of alerting algorithms for detection performance. We used the representative streams from each data source to compare sensitivity of 6 algorithms to injected spikes, and we used all data streams from 16 known events to compare them for detection timeliness. Methods The six methods compared were: Holt-Winters generalized exponential smoothing method (1)automated choice between daily methods, regression and an exponential weighted moving average (2)adaptive daily Shewhart-type chartadaptive one-sided daily CUSUMEWMA applied to 7-day means with a trend correction; and7-day temporal scan statistic Sensitivity testing: We conducted comparative sensitivity testing for categories of time series with similar scales and seasonal behavior. We added multiples of the standard deviation of each time series as single-day injects in separate algorithm runs. For each candidate method, we then used as a sensitivity measure the proportion of these runs for which the output of each algorithm was below alerting thresholds estimated empirically for each algorithm using simulated data streams. We identified the algorithm(s) whose sensitivity was most consistently high for each data category. For each syndromic query applied to each data source (outpatient, lab test orders, and prescriptions), 502 authentic time series were derived, one for each reporting treatment facility. Data categories were selected in order to group time series with similar expected algorithm performance: Median > 100 < Median ≤ 10Median = 0Lag 7 Autocorrelation Coefficient ≥ 0.2Lag 7 Autocorrelation Coefficient < 0.2 Timeliness testing: For the timeliness testing, we avoided artificiality of simulated signals by measuring alerting detection delays in the 16 corroborated outbreaks. The multiple time series from these events gave a total of 141 time series with outbreak intervals for timeliness testing. The following measures were computed to quantify timeliness of detection: Median Detection Delay – median number of days to detect the outbreak.Penalized Mean Detection Delay –mean number of days to detect the outbreak with outbreak misses penalized as 1 day plus the maximum detection time. Results Based on the injection results, the Holt-Winters algorithm was most sensitive among time series with positive medians. The adaptive CUSUM and the Shewhart methods were most sensitive for data streams with median zero. Table 1 provides timeliness results using the 141 outbreak-associated streams on sparse (Median=0) and non-sparse data categories. [Insert table #1 here] Data median Detection Delay, days Holt-winters Regression EWMA Adaptive Shewhart Adaptive CUSUM 7-day Trend-adj. EWMA 7-day Temporal Scan Median 0 Median 3 2 4 2 4.5 2 Penalized Mean 7.2 7 6.6 6.2 7.3 7.6 Median >0 Median 2 2 2.5 2 6 4 Penalized Mean 6.1 7 7.2 7.1 7.7 6.6 The gray shading in the table 1 indicates methods with shortest detection delays for sparse and non-sparse data streams. The Holt-Winters method was again superior for non-sparse data. For data with median=0, the adaptive CUSUM was superior for a daily false alarm probability of 0.01, but the Shewhart method was timelier for more liberal thresholds. Conclusions Both kinds of detection performance analysis showed the method based on Holt-Winters exponential smoothing superior on non-sparse time series with day-of-week effects. The adaptive CUSUM and She-whart methods proved optimal on sparse data and data without weekly patterns.
Further summation formulae related to generalized harmonic numbers
NASA Astrophysics Data System (ADS)
Zheng, De-Yin
2007-11-01
By employing the univariate series expansion of classical hypergeometric series formulae, Shen [L.-C. Shen, Remarks on some integrals and series involving the Stirling numbers and [zeta](n), Trans. Amer. Math. Soc. 347 (1995) 1391-1399] and Choi and Srivastava [J. Choi, H.M. Srivastava, Certain classes of infinite series, Monatsh. Math. 127 (1999) 15-25; J. Choi, H.M. Srivastava, Explicit evaluation of Euler and related sums, Ramanujan J. 10 (2005) 51-70] investigated the evaluation of infinite series related to generalized harmonic numbers. More summation formulae have systematically been derived by Chu [W. Chu, Hypergeometric series and the Riemann Zeta function, Acta Arith. 82 (1997) 103-118], who developed fully this approach to the multivariate case. The present paper will explore the hypergeometric series method further and establish numerous summation formulae expressing infinite series related to generalized harmonic numbers in terms of the Riemann Zeta function [zeta](m) with m=5,6,7, including several known ones as examples.
Tošić, Tamara; Sellers, Kristin K; Fröhlich, Flavio; Fedotenkova, Mariia; Beim Graben, Peter; Hutt, Axel
2015-01-01
For decades, research in neuroscience has supported the hypothesis that brain dynamics exhibits recurrent metastable states connected by transients, which together encode fundamental neural information processing. To understand the system's dynamics it is important to detect such recurrence domains, but it is challenging to extract them from experimental neuroscience datasets due to the large trial-to-trial variability. The proposed methodology extracts recurrent metastable states in univariate time series by transforming datasets into their time-frequency representations and computing recurrence plots based on instantaneous spectral power values in various frequency bands. Additionally, a new statistical inference analysis compares different trial recurrence plots with corresponding surrogates to obtain statistically significant recurrent structures. This combination of methods is validated by applying it to two artificial datasets. In a final study of visually-evoked Local Field Potentials in partially anesthetized ferrets, the methodology is able to reveal recurrence structures of neural responses with trial-to-trial variability. Focusing on different frequency bands, the δ-band activity is much less recurrent than α-band activity. Moreover, α-activity is susceptible to pre-stimuli, while δ-activity is much less sensitive to pre-stimuli. This difference in recurrence structures in different frequency bands indicates diverse underlying information processing steps in the brain.
Tošić, Tamara; Sellers, Kristin K.; Fröhlich, Flavio; Fedotenkova, Mariia; beim Graben, Peter; Hutt, Axel
2016-01-01
For decades, research in neuroscience has supported the hypothesis that brain dynamics exhibits recurrent metastable states connected by transients, which together encode fundamental neural information processing. To understand the system's dynamics it is important to detect such recurrence domains, but it is challenging to extract them from experimental neuroscience datasets due to the large trial-to-trial variability. The proposed methodology extracts recurrent metastable states in univariate time series by transforming datasets into their time-frequency representations and computing recurrence plots based on instantaneous spectral power values in various frequency bands. Additionally, a new statistical inference analysis compares different trial recurrence plots with corresponding surrogates to obtain statistically significant recurrent structures. This combination of methods is validated by applying it to two artificial datasets. In a final study of visually-evoked Local Field Potentials in partially anesthetized ferrets, the methodology is able to reveal recurrence structures of neural responses with trial-to-trial variability. Focusing on different frequency bands, the δ-band activity is much less recurrent than α-band activity. Moreover, α-activity is susceptible to pre-stimuli, while δ-activity is much less sensitive to pre-stimuli. This difference in recurrence structures in different frequency bands indicates diverse underlying information processing steps in the brain. PMID:26834580
Modelling world gold prices and USD foreign exchange relationship using multivariate GARCH model
NASA Astrophysics Data System (ADS)
Ping, Pung Yean; Ahmad, Maizah Hura Binti
2014-12-01
World gold price is a popular investment commodity. The series have often been modeled using univariate models. The objective of this paper is to show that there is a co-movement between gold price and USD foreign exchange rate. Using the effect of the USD foreign exchange rate on the gold price, a model that can be used to forecast future gold prices is developed. For this purpose, the current paper proposes a multivariate GARCH (Bivariate GARCH) model. Using daily prices of both series from 01.01.2000 to 05.05.2014, a causal relation between the two series understudied are found and a bivariate GARCH model is produced.
Forecasting Container Throughput at the Doraleh Port in Djibouti through Time Series Analysis
NASA Astrophysics Data System (ADS)
Mohamed Ismael, Hawa; Vandyck, George Kobina
The Doraleh Container Terminal (DCT) located in Djibouti has been noted as the most technologically advanced container terminal on the African continent. DCT's strategic location at the crossroads of the main shipping lanes connecting Asia, Africa and Europe put it in a unique position to provide important shipping services to vessels plying that route. This paper aims to forecast container throughput through the Doraleh Container Port in Djibouti by Time Series Analysis. A selection of univariate forecasting models has been used, namely Triple Exponential Smoothing Model, Grey Model and Linear Regression Model. By utilizing the above three models and their combination, the forecast of container throughput through the Doraleh port was realized. A comparison of the different forecasting results of the three models, in addition to the combination forecast is then undertaken, based on commonly used evaluation criteria Mean Absolute Deviation (MAD) and Mean Absolute Percentage Error (MAPE). The study found that the Linear Regression forecasting Model was the best prediction method for forecasting the container throughput, since its forecast error was the least. Based on the regression model, a ten (10) year forecast for container throughput at DCT has been made.
Network Reconstruction Using Nonparametric Additive ODE Models
Henderson, James; Michailidis, George
2014-01-01
Network representations of biological systems are widespread and reconstructing unknown networks from data is a focal problem for computational biologists. For example, the series of biochemical reactions in a metabolic pathway can be represented as a network, with nodes corresponding to metabolites and edges linking reactants to products. In a different context, regulatory relationships among genes are commonly represented as directed networks with edges pointing from influential genes to their targets. Reconstructing such networks from data is a challenging problem receiving much attention in the literature. There is a particular need for approaches tailored to time-series data and not reliant on direct intervention experiments, as the former are often more readily available. In this paper, we introduce an approach to reconstructing directed networks based on dynamic systems models. Our approach generalizes commonly used ODE models based on linear or nonlinear dynamics by extending the functional class for the functions involved from parametric to nonparametric models. Concomitantly we limit the complexity by imposing an additive structure on the estimated slope functions. Thus the submodel associated with each node is a sum of univariate functions. These univariate component functions form the basis for a novel coupling metric that we define in order to quantify the strength of proposed relationships and hence rank potential edges. We show the utility of the method by reconstructing networks using simulated data from computational models for the glycolytic pathway of Lactocaccus Lactis and a gene network regulating the pluripotency of mouse embryonic stem cells. For purposes of comparison, we also assess reconstruction performance using gene networks from the DREAM challenges. We compare our method to those that similarly rely on dynamic systems models and use the results to attempt to disentangle the distinct roles of linearity, sparsity, and derivative estimation. PMID:24732037
Performance of univariate forecasting on seasonal diseases: the case of tuberculosis.
Permanasari, Adhistya Erna; Rambli, Dayang Rohaya Awang; Dominic, P Dhanapal Durai
2011-01-01
The annual disease incident worldwide is desirable to be predicted for taking appropriate policy to prevent disease outbreak. This chapter considers the performance of different forecasting method to predict the future number of disease incidence, especially for seasonal disease. Six forecasting methods, namely linear regression, moving average, decomposition, Holt-Winter's, ARIMA, and artificial neural network (ANN), were used for disease forecasting on tuberculosis monthly data. The model derived met the requirement of time series with seasonality pattern and downward trend. The forecasting performance was compared using similar error measure in the base of the last 5 years forecast result. The findings indicate that ARIMA model was the most appropriate model since it obtained the less relatively error than the other model.
Fast and Flexible Multivariate Time Series Subsequence Search
NASA Technical Reports Server (NTRS)
Bhaduri, Kanishka; Oza, Nikunj C.; Zhu, Qiang; Srivastava, Ashok N.
2010-01-01
Multivariate Time-Series (MTS) are ubiquitous, and are generated in areas as disparate as sensor recordings in aerospace systems, music and video streams, medical monitoring, and financial systems. Domain experts are often interested in searching for interesting multivariate patterns from these MTS databases which often contain several gigabytes of data. Surprisingly, research on MTS search is very limited. Most of the existing work only supports queries with the same length of data, or queries on a fixed set of variables. In this paper, we propose an efficient and flexible subsequence search framework for massive MTS databases, that, for the first time, enables querying on any subset of variables with arbitrary time delays between them. We propose two algorithms to solve this problem (1) a List Based Search (LBS) algorithm which uses sorted lists for indexing, and (2) a R*-tree Based Search (RBS) which uses Minimum Bounding Rectangles (MBR) to organize the subsequences. Both algorithms guarantee that all matching patterns within the specified thresholds will be returned (no false dismissals). The very few false alarms can be removed by a post-processing step. Since our framework is also capable of Univariate Time-Series (UTS) subsequence search, we first demonstrate the efficiency of our algorithms on several UTS datasets previously used in the literature. We follow this up with experiments using two large MTS databases from the aviation domain, each containing several millions of observations. Both these tests show that our algorithms have very high prune rates (>99%) thus needing actual disk access for only less than 1% of the observations. To the best of our knowledge, MTS subsequence search has never been attempted on datasets of the size we have used in this paper.
Nonlinear analysis and dynamic structure in the energy market
NASA Astrophysics Data System (ADS)
Aghababa, Hajar
This research assesses the dynamic structure of the energy sector of the aggregate economy in the context of nonlinear mechanisms. Earlier studies have focused mainly on the price of the energy products when detecting nonlinearities in time series data of the energy market, and there is little mention of the production side of the market. Moreover, there is a lack of exploration about the implication of high dimensionality and time aggregation when analyzing the market's fundamentals. This research will address these gaps by including the quantity side of the market in addition to the price and by systematically incorporating various frequencies for sample sizes in three essays. The goal of this research is to provide an inclusive and exhaustive examination of the dynamics in the energy markets. The first essay begins with the application of statistical techniques, and it incorporates the most well-known univariate tests for nonlinearity with distinct power functions over alternatives and tests different null hypotheses. It utilizes the daily spot price observations on five major products in the energy market. The results suggest that the time series daily spot prices of the energy products are highly nonlinear in their nature. They demonstrate apparent evidence of general nonlinear serial dependence in each individual series, as well as nonlinearity in the first, second, and third moments of the series. The second essay examines the underlying mechanism of crude oil production and identifies the nonlinear structure of the production market by utilizing various monthly time series observations of crude oil production: the U.S. field, Organization of the Petroleum Exporting Countries (OPEC), non-OPEC, and the world production of crude oil. The finding implies that the time series data of the U.S. field, OPEC, and the world production of crude oil exhibit deep nonlinearity in their structure and are generated by nonlinear mechanisms. However, the dynamics of the non-OPEC production time series data does not reveal signs of nonlinearity. The third essay explores nonlinear structure in the case of high dimensionality of the observations, different frequencies of sample sizes, and division of the samples into sub-samples. It systematically examines the robustness of the inference methods at various levels of time aggregation by employing daily spot prices on crude oil for 26 years as well as monthly spot price index on crude oil for 41 years. The daily and monthly samples are divided into sub-samples as well. All the tests detect strong evidence of nonlinear structure in the daily spot price of crude oil; whereas in monthly observations the evidence of nonlinear dependence is less dramatic, indicating that the nonlinear serial dependence will not be as intense when the time aggregation increase in time series observations.
Relationship between cataract severity and socioeconomic status.
Wesolosky, Jason D; Rudnisky, Christopher J
2013-12-01
To determine the relationship between cataract severity and socioeconomic status (SES). Retrospective, observational case series. A total of 1350 eyes underwent phacoemulsification cataract extraction by a single surgeon using an Alcon Infiniti system. Cataract severity was measured using phaco time in seconds. SES was measured using area-level aggregate census data: median income, education, proportion of common-law couples, and employment rate. Preoperative best corrected visual acuity was obtained and converted to logarithm of the minimum angle of resolution values. For patients undergoing bilateral surgery, the generalized estimating equation was used to account for the correlation between eyes. Univariate analyses were performed using simple regression, and multivariate analyses were performed to account for variables with significant relationships (p < 0.05) on univariate testing. Sensitivity analyses were performed to assess the effect of including patient age in the controlled analyses. Multivariate analyses demonstrated that cataracts were more severe when the median income was lower (p = 0.001) and the proportion of common-law couples living in a patient's community (p = 0.012) and the unemployment rate (p = 0.002) were higher. These associations persisted even when controlling for patient age. Patients of lower SES have more severe cataracts. Copyright © 2013 Canadian Ophthalmological Society. Published by Elsevier Inc. All rights reserved.
Measures of dependence for multivariate Lévy distributions
NASA Astrophysics Data System (ADS)
Boland, J.; Hurd, T. R.; Pivato, M.; Seco, L.
2001-02-01
Recent statistical analysis of a number of financial databases is summarized. Increasing agreement is found that logarithmic equity returns show a certain type of asymptotic behavior of the largest events, namely that the probability density functions have power law tails with an exponent α≈3.0. This behavior does not vary much over different stock exchanges or over time, despite large variations in trading environments. The present paper proposes a class of multivariate distributions which generalizes the observed qualities of univariate time series. A new consequence of the proposed class is the "spectral measure" which completely characterizes the multivariate dependences of the extreme tails of the distribution. This measure on the unit sphere in M-dimensions, in principle completely general, can be determined empirically by looking at extreme events. If it can be observed and determined, it will prove to be of importance for scenario generation in portfolio risk management.
Localizing epileptic seizure onsets with Granger causality
NASA Astrophysics Data System (ADS)
Adhikari, Bhim M.; Epstein, Charles M.; Dhamala, Mukesh
2013-09-01
Accurate localization of the epileptic seizure onset zones (SOZs) is crucial for successful surgery, which usually depends on the information obtained from intracranial electroencephalography (IEEG) recordings. The visual criteria and univariate methods of analyzing IEEG recordings have not always produced clarity on the SOZs for resection and ultimate seizure freedom for patients. Here, to contribute to improving the localization of the SOZs and to understanding the mechanism of seizure propagation over the brain, we applied spectral interdependency methods to IEEG time series recorded from patients during seizures. We found that the high-frequency (>80 Hz) Granger causality (GC) occurs before the onset of any visible ictal activity and causal relationships involve the recording electrodes where clinically identifiable seizures later develop. These results suggest that high-frequency oscillatory network activities precede and underlie epileptic seizures, and that GC spectral measures derived from IEEG can assist in precise delineation of seizure onset times and SOZs.
Time-series analysis of delta13C from tree rings. I. Time trends and autocorrelation.
Monserud, R A; Marshall, J D
2001-09-01
Univariate time-series analyses were conducted on stable carbon isotope ratios obtained from tree-ring cellulose. We looked for the presence and structure of autocorrelation. Significant autocorrelation violates the statistical independence assumption and biases hypothesis tests. Its presence would indicate the existence of lagged physiological effects that persist for longer than the current year. We analyzed data from 28 trees (60-85 years old; mean = 73 years) of western white pine (Pinus monticola Dougl.), ponderosa pine (Pinus ponderosa Laws.), and Douglas-fir (Pseudotsuga menziesii (Mirb.) Franco var. glauca) growing in northern Idaho. Material was obtained by the stem analysis method from rings laid down in the upper portion of the crown throughout each tree's life. The sampling protocol minimized variation caused by changing light regimes within each tree. Autoregressive moving average (ARMA) models were used to describe the autocorrelation structure over time. Three time series were analyzed for each tree: the stable carbon isotope ratio (delta(13)C); discrimination (delta); and the difference between ambient and internal CO(2) concentrations (c(a) - c(i)). The effect of converting from ring cellulose to whole-leaf tissue did not affect the analysis because it was almost completely removed by the detrending that precedes time-series analysis. A simple linear or quadratic model adequately described the time trend. The residuals from the trend had a constant mean and variance, thus ensuring stationarity, a requirement for autocorrelation analysis. The trend over time for c(a) - c(i) was particularly strong (R(2) = 0.29-0.84). Autoregressive moving average analyses of the residuals from these trends indicated that two-thirds of the individual tree series contained significant autocorrelation, whereas the remaining third were random (white noise) over time. We were unable to distinguish between individuals with and without significant autocorrelation beforehand. Significant ARMA models were all of low order, with either first- or second-order (i.e., lagged 1 or 2 years, respectively) models performing well. A simple autoregressive (AR(1)), model was the most common. The most useful generalization was that the same ARMA model holds for each of the three series (delta(13)C, delta, c(a) - c(i)) for an individual tree, if the time trend has been properly removed for each series. The mean series for the two pine species were described by first-order ARMA models (1-year lags), whereas the Douglas-fir mean series were described by second-order models (2-year lags) with negligible first-order effects. Apparently, the process of constructing a mean time series for a species preserves an underlying signal related to delta(13)C while canceling some of the random individual tree variation. Furthermore, the best model for the overall mean series (e.g., for a species) cannot be inferred from a consensus of the individual tree model forms, nor can its parameters be estimated reliably from the mean of the individual tree parameters. Because two-thirds of the individual tree time series contained significant autocorrelation, the normal assumption of a random structure over time is unwarranted, even after accounting for the time trend. The residuals of an appropriate ARMA model satisfy the independence assumption, and can be used to make hypothesis tests.
Forecasting daily attendances at an emergency department to aid resource planning
Sun, Yan; Heng, Bee Hoon; Seow, Yian Tay; Seow, Eillyne
2009-01-01
Background Accurate forecasting of emergency department (ED) attendances can be a valuable tool for micro and macro level planning. Methods Data for analysis was the counts of daily patient attendances at the ED of an acute care regional general hospital from July 2005 to Mar 2008. Patients were stratified into three acuity categories; i.e. P1, P2 and P3, with P1 being the most acute and P3 being the least acute. The autoregressive integrated moving average (ARIMA) method was separately applied to each of the three acuity categories and total patient attendances. Independent variables included in the model were public holiday (yes or no), ambient air quality measured by pollution standard index (PSI), daily ambient average temperature and daily relative humidity. The seasonal components of weekly and yearly periodicities in the time series of daily attendances were also studied. Univariate analysis by t-tests and multivariate time series analysis were carried out in SPSS version 15. Results By time series analyses, P1 attendances did not show any weekly or yearly periodicity and was only predicted by ambient air quality of PSI > 50. P2 and total attendances showed weekly periodicities, and were also significantly predicted by public holiday. P3 attendances were significantly correlated with day of the week, month of the year, public holiday, and ambient air quality of PSI > 50. After applying the developed models to validate the forecast, the MAPE of prediction by the models were 16.8%, 6.7%, 8.6% and 4.8% for P1, P2, P3 and total attendances, respectively. The models were able to account for most of the significant autocorrelations present in the data. Conclusion Time series analysis has been shown to provide a useful, readily available tool for predicting emergency department workload that can be used to plan staff roster and resource planning. PMID:19178716
Testing for significance of phase synchronisation dynamics in the EEG.
Daly, Ian; Sweeney-Reed, Catherine M; Nasuto, Slawomir J
2013-06-01
A number of tests exist to check for statistical significance of phase synchronisation within the Electroencephalogram (EEG); however, the majority suffer from a lack of generality and applicability. They may also fail to account for temporal dynamics in the phase synchronisation, regarding synchronisation as a constant state instead of a dynamical process. Therefore, a novel test is developed for identifying the statistical significance of phase synchronisation based upon a combination of work characterising temporal dynamics of multivariate time-series and Markov modelling. We show how this method is better able to assess the significance of phase synchronisation than a range of commonly used significance tests. We also show how the method may be applied to identify and classify significantly different phase synchronisation dynamics in both univariate and multivariate datasets.
Jerlström, Tomas; Gårdmark, Truls; Carringer, Malcolm; Holmäng, Sten; Liedberg, Fredrik; Hosseini, Abolfazl; Malmström, Per-Uno; Ljungberg, Börje; Hagberg, Oskar; Jahnson, Staffan
2014-08-01
Cystectomy combined with pelvic lymph-node dissection and urinary diversion entails high morbidity and mortality. Improvements are needed, and a first step is to collect information on the current situation. In 2011, this group took the initiative to start a population-based database in Sweden (population 9.5 million in 2011) with prospective registration of patients and complications until 90 days after cystectomy. This article reports findings from the first year of registration. Participation was voluntary, and data were reported by local urologists or research nurses. Perioperative parameters and early complications classified according to the modified Clavien system were registered, and selected variables of possible importance for complications were analysed by univariate and multivariate logistic regression. During 2011, 285 (65%) of 435 cystectomies performed in Sweden were registered in the database, the majority reported by the seven academic centres. Median blood loss was 1000 ml, operating time 318 min, and length of hospital stay 15 days. Any complications were registered for 103 patients (36%). Clavien grades 1-2 and 3-5 were noted in 19% and 15%, respectively. Thirty-seven patients (13%) were reoperated on at least once. In logistic regression analysis elevated risk of complications was significantly associated with operating time exceeding 318 min in both univariate and multivariate analysis, and with age 76-89 years only in multivariate analysis. It was feasible to start a national population-based registry of radical cystectomies for bladder cancer. The evaluation of the first year shows an increased risk of complications in patients with longer operating time and higher age. The results agree with some previously published series but should be interpreted with caution considering the relatively low coverage, which is expected to be higher in the future.
Economic Expansion Is a Major Determinant of Physician Supply and Utilization
Cooper, Richard A; Getzen, Thomas E; Laud, Prakash
2003-01-01
Objective To assess the relationship between levels of economic development and the supply and utilization of physicians. Data Sources Data were obtained from the American Medical Association, American Osteopathic Association, Organization for Economic Cooperation and Development (OECD), Bureau of Health Professions, Bureau of Labor Statistics, Bureau of Economic Analysis, Census Bureau, Health Care Financing Administration, and historical sources. Study Design Economic development, expressed as real per capita gross domestic product (GDP) or personal income, was correlated with per capita health care labor and physician supply within countries and states over periods of time spanning 25–70 years and across countries, states, and metropolitan statistical areas (MSAs) at multiple points in time over periods of up to 30 years. Longitudinal data were analyzed in four complementary ways: (1) simple univariate regressions; (2) regressions in which temporal trends were partialled out; (3) time series comparing percentage differences across segments of time; and (4) a bivariate Granger causality test. Cross-sectional data were assessed at multiple time points by means of univariate regression analyses. Principal Findings Under each analytic scenario, physician supply correlated with differences in GDP or personal income. Longitudinal correlations were associated with temporal lags of approximately 5 years for health employment and 10 years for changes in physician supply. The magnitude of changes in per capita physician supply in the United States was equivalent to differences of approximately 0.75 percent for each 1.0 percent difference in GDP. The greatest effects of economic expansion were on the medical specialties, whereas the surgical and hospital-based specialties were affected to a lesser degree, and levels of economic expansion had little influence on family/general practice. Conclusions Economic expansion has a strong, lagged relationship with changes in physician supply. This suggests that economic projections could serve as a gauge for projecting the future utilization of physician services. PMID:12785567
Topological Isomorphisms of Human Brain and Financial Market Networks
Vértes, Petra E.; Nicol, Ruth M.; Chapman, Sandra C.; Watkins, Nicholas W.; Robertson, Duncan A.; Bullmore, Edward T.
2011-01-01
Although metaphorical and conceptual connections between the human brain and the financial markets have often been drawn, rigorous physical or mathematical underpinnings of this analogy remain largely unexplored. Here, we apply a statistical and graph theoretic approach to the study of two datasets – the time series of 90 stocks from the New York stock exchange over a 3-year period, and the fMRI-derived time series acquired from 90 brain regions over the course of a 10-min-long functional MRI scan of resting brain function in healthy volunteers. Despite the many obvious substantive differences between these two datasets, graphical analysis demonstrated striking commonalities in terms of global network topological properties. Both the human brain and the market networks were non-random, small-world, modular, hierarchical systems with fat-tailed degree distributions indicating the presence of highly connected hubs. These properties could not be trivially explained by the univariate time series statistics of stock price returns. This degree of topological isomorphism suggests that brains and markets can be regarded broadly as members of the same family of networks. The two systems, however, were not topologically identical. The financial market was more efficient and more modular – more highly optimized for information processing – than the brain networks; but also less robust to systemic disintegration as a result of hub deletion. We conclude that the conceptual connections between brains and markets are not merely metaphorical; rather these two information processing systems can be rigorously compared in the same mathematical language and turn out often to share important topological properties in common to some degree. There will be interesting scientific arbitrage opportunities in further work at the graph-theoretically mediated interface between systems neuroscience and the statistical physics of financial markets. PMID:22007161
Generalized multiplicative error models: Asymptotic inference and empirical analysis
NASA Astrophysics Data System (ADS)
Li, Qian
This dissertation consists of two parts. The first part focuses on extended Multiplicative Error Models (MEM) that include two extreme cases for nonnegative series. These extreme cases are common phenomena in high-frequency financial time series. The Location MEM(p,q) model incorporates a location parameter so that the series are required to have positive lower bounds. The estimator for the location parameter turns out to be the minimum of all the observations and is shown to be consistent. The second case captures the nontrivial fraction of zero outcomes feature in a series and combines a so-called Zero-Augmented general F distribution with linear MEM(p,q). Under certain strict stationary and moment conditions, we establish a consistency and asymptotic normality of the semiparametric estimation for these two new models. The second part of this dissertation examines the differences and similarities between trades in the home market and trades in the foreign market of cross-listed stocks. We exploit the multiplicative framework to model trading duration, volume per trade and price volatility for Canadian shares that are cross-listed in the New York Stock Exchange (NYSE) and the Toronto Stock Exchange (TSX). We explore the clustering effect, interaction between trading variables, and the time needed for price equilibrium after a perturbation for each market. The clustering effect is studied through the use of univariate MEM(1,1) on each variable, while the interactions among duration, volume and price volatility are captured by a multivariate system of MEM(p,q). After estimating these models by a standard QMLE procedure, we exploit the Impulse Response function to compute the calendar time for a perturbation in these variables to be absorbed into price variance, and use common statistical tests to identify the difference between the two markets in each aspect. These differences are of considerable interest to traders, stock exchanges and policy makers.
Ouma, Paul O; Agutu, Nathan O; Snow, Robert W; Noor, Abdisalan M
2017-09-18
Precise quantification of health service utilisation is important for the estimation of disease burden and allocation of health resources. Current approaches to mapping health facility utilisation rely on spatial accessibility alone as the predictor. However, other spatially varying social, demographic and economic factors may affect the use of health services. The exclusion of these factors can lead to the inaccurate estimation of health facility utilisation. Here, we compare the accuracy of a univariate spatial model, developed only from estimated travel time, to a multivariate model that also includes relevant social, demographic and economic factors. A theoretical surface of travel time to the nearest public health facility was developed. These were assigned to each child reported to have had fever in the Kenya demographic and health survey of 2014 (KDHS 2014). The relationship of child treatment seeking for fever with travel time, household and individual factors from the KDHS2014 were determined using multilevel mixed modelling. Bayesian information criterion (BIC) and likelihood ratio test (LRT) tests were carried out to measure how selected factors improve parsimony and goodness of fit of the time model. Using the mixed model, a univariate spatial model of health facility utilisation was fitted using travel time as the predictor. The mixed model was also used to compute a multivariate spatial model of utilisation, using travel time and modelled surfaces of selected household and individual factors as predictors. The univariate and multivariate spatial models were then compared using the receiver operating area under the curve (AUC) and a percent correct prediction (PCP) test. The best fitting multivariate model had travel time, household wealth index and number of children in household as the predictors. These factors reduced BIC of the time model from 4008 to 2959, a change which was confirmed by the LRT test. Although there was a high correlation of the two modelled probability surfaces (Adj R 2 = 88%), the multivariate model had better AUC compared to the univariate model; 0.83 versus 0.73 and PCP 0.61 versus 0.45 values. Our study shows that a model that uses travel time, as well as household and individual-level socio-demographic factors, results in a more accurate estimation of use of health facilities for the treatment of childhood fever, compared to one that relies on only travel time.
Cain, Meghan K; Zhang, Zhiyong; Yuan, Ke-Hai
2017-10-01
Nonnormality of univariate data has been extensively examined previously (Blanca et al., Methodology: European Journal of Research Methods for the Behavioral and Social Sciences, 9(2), 78-84, 2013; Miceeri, Psychological Bulletin, 105(1), 156, 1989). However, less is known of the potential nonnormality of multivariate data although multivariate analysis is commonly used in psychological and educational research. Using univariate and multivariate skewness and kurtosis as measures of nonnormality, this study examined 1,567 univariate distriubtions and 254 multivariate distributions collected from authors of articles published in Psychological Science and the American Education Research Journal. We found that 74 % of univariate distributions and 68 % multivariate distributions deviated from normal distributions. In a simulation study using typical values of skewness and kurtosis that we collected, we found that the resulting type I error rates were 17 % in a t-test and 30 % in a factor analysis under some conditions. Hence, we argue that it is time to routinely report skewness and kurtosis along with other summary statistics such as means and variances. To facilitate future report of skewness and kurtosis, we provide a tutorial on how to compute univariate and multivariate skewness and kurtosis by SAS, SPSS, R and a newly developed Web application.
Variation in ancient Egyptian stature and body proportions.
Zakrzewski, Sonia R
2003-07-01
Stature and the pattern of body proportions were investigated in a series of six time-successive Egyptian populations in order to investigate the biological effects on human growth of the development and intensification of agriculture, and the formation of state-level social organization. Univariate analyses of variance were performed to assess differences between the sexes and among various time periods. Significant differences were found both in stature and in raw long bone length measurements between the early semipastoral population and the later intensive agricultural population. The size differences were greater in males than in females. This disparity is suggested to be due to greater male response to poor nutrition in the earlier populations, and with the increasing development of social hierarchy, males were being provisioned preferentially over females. Little change in body shape was found through time, suggesting that all body segments were varying in size in response to environmental and social conditions. The change found in body plan is suggested to be the result of the later groups having a more tropical (Nilotic) form than the preceding populations. Copyright 2003 Wiley-Liss, Inc.
He, Jie; Zhao, Yunfeng; Zhao, Jingli; Gao, Jin; Han, Dandan; Xu, Pao; Yang, Runqing
2017-11-02
Because of their high economic importance, growth traits in fish are under continuous improvement. For growth traits that are recorded at multiple time-points in life, the use of univariate and multivariate animal models is limited because of the variable and irregular timing of these measures. Thus, the univariate random regression model (RRM) was introduced for the genetic analysis of dynamic growth traits in fish breeding. We used a multivariate random regression model (MRRM) to analyze genetic changes in growth traits recorded at multiple time-point of genetically-improved farmed tilapia. Legendre polynomials of different orders were applied to characterize the influences of fixed and random effects on growth trajectories. The final MRRM was determined by optimizing the univariate RRM for the analyzed traits separately via penalizing adaptively the likelihood statistical criterion, which is superior to both the Akaike information criterion and the Bayesian information criterion. In the selected MRRM, the additive genetic effects were modeled by Legendre polynomials of three orders for body weight (BWE) and body length (BL) and of two orders for body depth (BD). By using the covariance functions of the MRRM, estimated heritabilities were between 0.086 and 0.628 for BWE, 0.155 and 0.556 for BL, and 0.056 and 0.607 for BD. Only heritabilities for BD measured from 60 to 140 days of age were consistently higher than those estimated by the univariate RRM. All genetic correlations between growth time-points exceeded 0.5 for either single or pairwise time-points. Moreover, correlations between early and late growth time-points were lower. Thus, for phenotypes that are measured repeatedly in aquaculture, an MRRM can enhance the efficiency of the comprehensive selection for BWE and the main morphological traits.
Björklund, M; Gustafsson, L
2017-07-01
Understanding the magnitude and long-term patterns of selection in natural populations is of importance, for example, when analysing the evolutionary impact of climate change. We estimated univariate and multivariate directional, quadratic and correlational selection on four morphological traits (adult wing, tarsus and tail length, body mass) over a time period of 33 years (≈ 19 000 observations) in a nest-box breeding population of collared flycatchers (Ficedula albicollis). In general, selection was weak in both males and females over the years regardless of fitness measure (fledged young, recruits and survival) with only few cases with statistically significant selection. When data were analysed in a multivariate context and as time series, a number of patterns emerged; there was a consistent, but weak, selection for longer wings in both sexes, selection was stronger on females when the number of fledged young was used as a fitness measure, there were no indications of sexually antagonistic selection, and we found a negative correlation between selection on tarsus and wing length in both sexes but using different fitness measures. Uni- and multivariate selection gradients were correlated only for wing length and mass. Multivariate selection gradient vectors were longer than corresponding vector of univariate gradients and had more constrained direction. Correlational selection had little importance. Overall, the fitness surface was more or less flat with few cases of significant curvature, indicating that the adaptive peak with regard to body size in this species is broader than the phenotypic distribution, which has resulted in weak estimates of selection. © 2017 European Society For Evolutionary Biology. Journal of Evolutionary Biology © 2017 European Society For Evolutionary Biology.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Baker, Ryan; Han Gang; Sarangkasiri, Siriporn
2013-01-01
Purpose: To report clinical and dosimetric factors predictive of radiation pneumonitis (RP) in patients receiving lung stereotactic body radiation therapy (SBRT) from a series of 240 patients. Methods and Materials: Of the 297 isocenters treating 263 patients, 240 patients (n=263 isocenters) had evaluable information regarding RP. Age, gender, current smoking status and pack-years, O{sub 2} use, Charlson Comorbidity Index, prior lung radiation therapy (yes/no), dose/fractionation, V{sub 5}, V{sub 13}, V{sub 20}, V{sub prescription}, mean lung dose, planning target volume (PTV), total lung volume, and PTV/lung volume ratio were recorded. Results: Twenty-nine patients (11.0%) developed symptomatic pneumonitis (26 grade 2, 3more » grade 3). The mean V{sub 20} was 6.5% (range, 0.4%-20.2%), and the average mean lung dose was 5.03 Gy (0.547-12.2 Gy). In univariable analysis female gender (P=.0257) and Charlson Comorbidity index (P=.0366) were significantly predictive of RP. Among dosimetric parameters, V{sub 5} (P=.0186), V{sub 13} (P=.0438), and V{sub prescription} (where dose = 60 Gy) (P=.0128) were significant. There was only a trend toward significance for V{sub 20} (P=.0610). Planning target volume/normal lung volume ratio was highly significant (P=.0024). In multivariable analysis the clinical factors of female gender, pack-years smoking, and larger gross internal tumor volume and PTV were predictive (P=.0094, .0312, .0364, and .052, respectively), but no dosimetric factors were significant. Conclusions: Rate of symptomatic RP was 11%. Our mean lung dose was <600 cGy in most cases and V20 <10%. In univariable analysis, dosimetric factors were predictive, while tumor size (or tumor/lung volume ratio) played a role in multivariable and univariable and analysis, respectively.« less
Reported gum disease as a cardiovascular risk factor in adults with intellectual disabilities.
Hsieh, K; Murthy, S; Heller, T; Rimmer, J H; Yen, G
2018-03-01
Several risk factors for cardiovascular disease (CVD) have been identified among adults with intellectual disabilities (ID). Periodontitis has been reported to increase the risk of developing a CVD in the general population. Given that individuals with ID have been reported to have a higher prevalence of poor oral health than the general population, the purpose of this study was to determine whether adults with ID with informant reported gum disease present greater reported CVD than those who do not have reported gum disease and whether gum disease can be considered a risk factor for CVD. Using baseline data from the Longitudinal Health and Intellectual Disability Study from which informant survey data were collected, 128 participants with reported gum disease and 1252 subjects without reported gum disease were identified. A series of univariate logistic regressions was conducted to identify potential confounding factors for a multiple logistic regression. The series of univariate logistic regressions identified age, Down syndrome, hypercholesterolemia, hypertension, reported gum disease, daily consumption of fruits and vegetables and the addition of table salt as significant risk factors for reported CVD. When the significant factors from the univariate logistic regression were included in the multiple logistic analysis, reported gum disease remained as an independent risk factor for reported CVD after adjusting for the remaining risk factors. Compared with the adults with ID without reported gum disease, adults in the gum disease group demonstrated a significantly higher prevalence of reported CVD (19.5% vs. 9.7%; P = .001). After controlling for other risk factors, reported gum disease among adults with ID may be associated with a higher risk of CVD. However, further research that also includes clinical indices of periodontal disease and CVD for this population is needed to determine if there is a causal relationship between gum disease and CVD. © 2017 MENCAP and International Association of the Scientific Study of Intellectual and Developmental Disabilities and John Wiley & Sons Ltd.
NASA Astrophysics Data System (ADS)
Braga, Jez Willian Batista; Trevizan, Lilian Cristina; Nunes, Lidiane Cristina; Rufini, Iolanda Aparecida; Santos, Dário, Jr.; Krug, Francisco José
2010-01-01
The application of laser induced breakdown spectrometry (LIBS) aiming the direct analysis of plant materials is a great challenge that still needs efforts for its development and validation. In this way, a series of experimental approaches has been carried out in order to show that LIBS can be used as an alternative method to wet acid digestions based methods for analysis of agricultural and environmental samples. The large amount of information provided by LIBS spectra for these complex samples increases the difficulties for selecting the most appropriated wavelengths for each analyte. Some applications have suggested that improvements in both accuracy and precision can be achieved by the application of multivariate calibration in LIBS data when compared to the univariate regression developed with line emission intensities. In the present work, the performance of univariate and multivariate calibration, based on partial least squares regression (PLSR), was compared for analysis of pellets of plant materials made from an appropriate mixture of cryogenically ground samples with cellulose as the binding agent. The development of a specific PLSR model for each analyte and the selection of spectral regions containing only lines of the analyte of interest were the best conditions for the analysis. In this particular application, these models showed a similar performance, but PLSR seemed to be more robust due to a lower occurrence of outliers in comparison to the univariate method. Data suggests that efforts dealing with sample presentation and fitness of standards for LIBS analysis must be done in order to fulfill the boundary conditions for matrix independent development and validation.
Kroeker, Kristine; Widdifield, Jessica; Muthukumarana, Saman; Jiang, Depeng; Lix, Lisa M
2017-01-01
Objective This research proposes a model-based method to facilitate the selection of disease case definitions from validation studies for administrative health data. The method is demonstrated for a rheumatoid arthritis (RA) validation study. Study design and setting Data were from 148 definitions to ascertain cases of RA in hospital, physician and prescription medication administrative data. We considered: (A) separate univariate models for sensitivity and specificity, (B) univariate model for Youden’s summary index and (C) bivariate (ie, joint) mixed-effects model for sensitivity and specificity. Model covariates included the number of diagnoses in physician, hospital and emergency department records, physician diagnosis observation time, duration of time between physician diagnoses and number of RA-related prescription medication records. Results The most common case definition attributes were: 1+ hospital diagnosis (65%), 2+ physician diagnoses (43%), 1+ specialist physician diagnosis (51%) and 2+ years of physician diagnosis observation time (27%). Statistically significant improvements in sensitivity and/or specificity for separate univariate models were associated with (all p values <0.01): 2+ and 3+ physician diagnoses, unlimited physician diagnosis observation time, 1+ specialist physician diagnosis and 1+ RA-related prescription medication records (65+ years only). The bivariate model produced similar results. Youden’s index was associated with these same case definition criteria, except for the length of the physician diagnosis observation time. Conclusion A model-based method provides valuable empirical evidence to aid in selecting a definition(s) for ascertaining diagnosed disease cases from administrative health data. The choice between univariate and bivariate models depends on the goals of the validation study and number of case definitions. PMID:28645978
Postoperative urinary retention after inguinal hernia repair: a single institution experience.
Blair, A B; Dwarakanath, A; Mehta, A; Liang, H; Hui, X; Wyman, C; Ouanes, J P P; Nguyen, H T
2017-12-01
Inguinal hernia repair is a common general surgery procedure with low morbidity. However, postoperative urinary retention (PUR) occurs in up to 22% of patients, resulting in further extraneous treatments.This single institution series investigates whether patient comorbidities, surgical approaches, and anesthesia methods are associated with developing PUR after inguinal hernia repairs. This is a single institution retrospective review of inguinal hernia from 2012 to 2015. PUR was defined as patients without a postoperative urinary catheter who subsequently required bladder decompression due to an inability to void. Univariate and multivariate logistic regressions were performed to quantify the associations between patient, surgical, and anesthetic factors with PUR. Stratification analysis was conducted at age of 50 years. 445 patients were included (42.9% laparoscopic and 57.1% open). Overall rate of PUR was 11.2% (12% laparoscopic, 10.6% open, and p = 0.64). In univariate analysis, PUR was significantly associated with patient age >50 and history of benign prostatic hyperplasia (BPH). Risk stratification for age >50 revealed in this cohort a 2.49 times increased PUR risk with lack of intraoperative bladder decompression (p = 0.013). At our institution, we found that patient age, history of BPH, and bilateral repair were associated with PUR after inguinal hernia repair. No association was found with PUR and laparoscopic vs open approach. Older males may be at higher risk without intraoperative bladder decompression, and therefore, catheter placement should be considered in this population, regardless of surgical approach.
Prolonged instability prior to a regime shift
Spanbauer, Trisha; Allen, Craig R.; Angeler, David G.; Eason, Tarsha; Fritz, Sherilyn C.; Garmestani, Ahjond S.; Nash, Kirsty L.; Stone, Jeffery R.
2014-01-01
Regime shifts are generally defined as the point of ‘abrupt’ change in the state of a system. However, a seemingly abrupt transition can be the product of a system reorganization that has been ongoing much longer than is evident in statistical analysis of a single component of the system. Using both univariate and multivariate statistical methods, we tested a long-term high-resolution paleoecological dataset with a known change in species assemblage for a regime shift. Analysis of this dataset with Fisher Information and multivariate time series modeling showed that there was a∼2000 year period of instability prior to the regime shift. This period of instability and the subsequent regime shift coincide with regional climate change, indicating that the system is undergoing extrinsic forcing. Paleoecological records offer a unique opportunity to test tools for the detection of thresholds and stable-states, and thus to examine the long-term stability of ecosystems over periods of multiple millennia.
Forecasting irregular variations of UT1-UTC and LOD data caused by ENSO
NASA Astrophysics Data System (ADS)
Niedzielski, T.; Kosek, W.
2008-04-01
The research focuses on prediction of LOD and UT1-UTC time series up to one-year in the future with the particular emphasis on the prediction improvement during El Nĩ o or La Nĩ a n n events. The polynomial-harmonic least-squares model is applied to fit the deterministic function to LOD data. The stochastic residuals computed as the difference between LOD data and the polynomial- harmonic model reveal the extreme values driven by El Nĩ o or La Nĩ a. These peaks are modeled by the n n stochastic bivariate autoregressive prediction. This approach focuses on the auto- and cross-correlations between LOD and the axial component of the atmospheric angular momentum. This technique allows one to derive more accurate predictions than purely univariate forecasts, particularly during El Nĩ o/La n Nĩ a events. n
Forecasting models for flow and total dissolved solids in Karoun river-Iran
NASA Astrophysics Data System (ADS)
Salmani, Mohammad Hassan; Salmani Jajaei, Efat
2016-04-01
Water quality is one of the most important factors contributing to a healthy life. From the water quality management point of view, TDS (total dissolved solids) is the most important factor and many water developing plans have been implemented in recognition of this factor. However, these plans have not been perfect and very successful in overcoming the poor water quality problem, so there are a good volume of related studies in the literature. We study TDS and the water flow of the Karoun river in southwest Iran. We collected the necessary time series data from the Harmaleh station located in the river. We present two Univariate Seasonal Autoregressive Integrated Movement Average (ARIMA) models to forecast TDS and water flow in this river. Then, we build up a Transfer Function (TF) model to formulate the TDS as a function of water flow volume. A performance comparison between the Seasonal ARIMA and the TF models are presented.
Correlation dimension and phase space contraction via extreme value theory
NASA Astrophysics Data System (ADS)
Faranda, Davide; Vaienti, Sandro
2018-04-01
We show how to obtain theoretical and numerical estimates of correlation dimension and phase space contraction by using the extreme value theory. The maxima of suitable observables sampled along the trajectory of a chaotic dynamical system converge asymptotically to classical extreme value laws where: (i) the inverse of the scale parameter gives the correlation dimension and (ii) the extremal index is associated with the rate of phase space contraction for backward iteration, which in dimension 1 and 2, is closely related to the positive Lyapunov exponent and in higher dimensions is related to the metric entropy. We call it the Dynamical Extremal Index. Numerical estimates are straightforward to obtain as they imply just a simple fit to a univariate distribution. Numerical tests range from low dimensional maps, to generalized Henon maps and climate data. The estimates of the indicators are particularly robust even with relatively short time series.
A regressive methodology for estimating missing data in rainfall daily time series
NASA Astrophysics Data System (ADS)
Barca, E.; Passarella, G.
2009-04-01
The "presence" of gaps in environmental data time series represents a very common, but extremely critical problem, since it can produce biased results (Rubin, 1976). Missing data plagues almost all surveys. The problem is how to deal with missing data once it has been deemed impossible to recover the actual missing values. Apart from the amount of missing data, another issue which plays an important role in the choice of any recovery approach is the evaluation of "missingness" mechanisms. When data missing is conditioned by some other variable observed in the data set (Schafer, 1997) the mechanism is called MAR (Missing at Random). Otherwise, when the missingness mechanism depends on the actual value of the missing data, it is called NCAR (Not Missing at Random). This last is the most difficult condition to model. In the last decade interest arose in the estimation of missing data by using regression (single imputation). More recently multiple imputation has become also available, which returns a distribution of estimated values (Scheffer, 2002). In this paper an automatic methodology for estimating missing data is presented. In practice, given a gauging station affected by missing data (target station), the methodology checks the randomness of the missing data and classifies the "similarity" between the target station and the other gauging stations spread over the study area. Among different methods useful for defining the similarity degree, whose effectiveness strongly depends on the data distribution, the Spearman correlation coefficient was chosen. Once defined the similarity matrix, a suitable, nonparametric, univariate, and regressive method was applied in order to estimate missing data in the target station: the Theil method (Theil, 1950). Even though the methodology revealed to be rather reliable an improvement of the missing data estimation can be achieved by a generalization. A first possible improvement consists in extending the univariate technique to the multivariate approach. Another approach follows the paradigm of the "multiple imputation" (Rubin, 1987; Rubin, 1988), which consists in using a set of "similar stations" instead than the most similar. This way, a sort of estimation range can be determined allowing the introduction of uncertainty. Finally, time series can be grouped on the basis of monthly rainfall rates defining classes of wetness (i.e.: dry, moderately rainy and rainy), in order to achieve the estimation using homogeneous data subsets. We expect that integrating the methodology with these enhancements will certainly improve its reliability. The methodology was applied to the daily rainfall time series data registered in the Candelaro River Basin (Apulia - South Italy) from 1970 to 2001. REFERENCES D.B., Rubin, 1976. Inference and Missing Data. Biometrika 63 581-592 D.B. Rubin, 1987. Multiple Imputation for Nonresponce in Surveys, New York: John Wiley & Sons, Inc. D.B. Rubin, 1988. An overview of multiple imputation. In Survey Research Section, pp. 79-84, American Statistical Association, 1988. J.L., Schafer, 1997. Analysis of Incomplete Multivariate Data, Chapman & Hall. J., Scheffer, 2002. Dealing with Missing Data. Res. Lett. Inf. Math. Sci. 3, 153-160. Available online at http://www.massey.ac.nz/~wwiims/research/letters/ H. Theil, 1950. A rank-invariant method of linear and polynomial regression analysis. Indicationes Mathematicae, 12, pp.85-91.
Event time analysis of longitudinal neuroimage data.
Sabuncu, Mert R; Bernal-Rusiel, Jorge L; Reuter, Martin; Greve, Douglas N; Fischl, Bruce
2014-08-15
This paper presents a method for the statistical analysis of the associations between longitudinal neuroimaging measurements, e.g., of cortical thickness, and the timing of a clinical event of interest, e.g., disease onset. The proposed approach consists of two steps, the first of which employs a linear mixed effects (LME) model to capture temporal variation in serial imaging data. The second step utilizes the extended Cox regression model to examine the relationship between time-dependent imaging measurements and the timing of the event of interest. We demonstrate the proposed method both for the univariate analysis of image-derived biomarkers, e.g., the volume of a structure of interest, and the exploratory mass-univariate analysis of measurements contained in maps, such as cortical thickness and gray matter density. The mass-univariate method employs a recently developed spatial extension of the LME model. We applied our method to analyze structural measurements computed using FreeSurfer, a widely used brain Magnetic Resonance Image (MRI) analysis software package. We provide a quantitative and objective empirical evaluation of the statistical performance of the proposed method on longitudinal data from subjects suffering from Mild Cognitive Impairment (MCI) at baseline. Copyright © 2014 Elsevier Inc. All rights reserved.
Inouye, David I.; Ravikumar, Pradeep; Dhillon, Inderjit S.
2016-01-01
We develop Square Root Graphical Models (SQR), a novel class of parametric graphical models that provides multivariate generalizations of univariate exponential family distributions. Previous multivariate graphical models (Yang et al., 2015) did not allow positive dependencies for the exponential and Poisson generalizations. However, in many real-world datasets, variables clearly have positive dependencies. For example, the airport delay time in New York—modeled as an exponential distribution—is positively related to the delay time in Boston. With this motivation, we give an example of our model class derived from the univariate exponential distribution that allows for almost arbitrary positive and negative dependencies with only a mild condition on the parameter matrix—a condition akin to the positive definiteness of the Gaussian covariance matrix. Our Poisson generalization allows for both positive and negative dependencies without any constraints on the parameter values. We also develop parameter estimation methods using node-wise regressions with ℓ1 regularization and likelihood approximation methods using sampling. Finally, we demonstrate our exponential generalization on a synthetic dataset and a real-world dataset of airport delay times. PMID:27563373
Chan, Emily H; Sahai, Vikram; Conrad, Corrie; Brownstein, John S
2011-05-01
A variety of obstacles including bureaucracy and lack of resources have interfered with timely detection and reporting of dengue cases in many endemic countries. Surveillance efforts have turned to modern data sources, such as Internet search queries, which have been shown to be effective for monitoring influenza-like illnesses. However, few have evaluated the utility of web search query data for other diseases, especially those of high morbidity and mortality or where a vaccine may not exist. In this study, we aimed to assess whether web search queries are a viable data source for the early detection and monitoring of dengue epidemics. Bolivia, Brazil, India, Indonesia and Singapore were chosen for analysis based on available data and adequate search volume. For each country, a univariate linear model was then built by fitting a time series of the fraction of Google search query volume for specific dengue-related queries from that country against a time series of official dengue case counts for a time-frame within 2003-2010. The specific combination of queries used was chosen to maximize model fit. Spurious spikes in the data were also removed prior to model fitting. The final models, fit using a training subset of the data, were cross-validated against both the overall dataset and a holdout subset of the data. All models were found to fit the data quite well, with validation correlations ranging from 0.82 to 0.99. Web search query data were found to be capable of tracking dengue activity in Bolivia, Brazil, India, Indonesia and Singapore. Whereas traditional dengue data from official sources are often not available until after some substantial delay, web search query data are available in near real-time. These data represent valuable complement to assist with traditional dengue surveillance.
Influence of stromal refractive index and hydration on corneal laser refractive surgery.
de Ortueta, Diego; von Rüden, Dennis; Magnago, Thomas; Arba Mosquera, Samuel
2014-06-01
To evaluate the influence of the stromal refractive index and hydration on postoperative outcomes in eyes that had corneal laser refractive surgery using the Amaris laser system. Augenzentrum Recklinghausen, Recklinghausen, Germany. Comparative case series. At the 6-month follow-up, right eyes were retrospectively analyzed. The effect of the stromal refractive index and hydration on refractive outcomes was assessed using univariate linear and multilinear correlations. Sixty eyes were analyzed. Univariate linear analyses showed that the stromal refractive index and hydration were correlated with the thickness of the preoperative exposed stroma and was statistically different for laser in situ keratomileusis and laser-assisted subepithelial keratectomy treatments. Univariate multilinear analyses showed that the spherical equivalent (SE) was correlated with the attempted SE and stromal refractive index (or hydration). Analyses suggest overcorrections for higher stromal refractive index values and for lower hydration values. The stromal refractive index and hydration affected postoperative outcomes in a subtle, yet significant manner. An adjustment toward greater attempted correction in highly hydrated corneas and less intended correction in low hydrated corneas might help optimize refractive outcomes. Mr. Magnago and Dr. Arba-Mosquera are employees of and Dr. Diego de Ortueta is a consultant to Schwind eye-tech-solutions GmbH & Co. KG. Mr. Rüden has no financial or proprietary interest in any material or method mentioned. Copyright © 2014 ASCRS and ESCRS. Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Vrac, Mathieu
2018-06-01
Climate simulations often suffer from statistical biases with respect to observations or reanalyses. It is therefore common to correct (or adjust) those simulations before using them as inputs into impact models. However, most bias correction (BC) methods are univariate and so do not account for the statistical dependences linking the different locations and/or physical variables of interest. In addition, they are often deterministic, and stochasticity is frequently needed to investigate climate uncertainty and to add constrained randomness to climate simulations that do not possess a realistic variability. This study presents a multivariate method of rank resampling for distributions and dependences (R2D2) bias correction allowing one to adjust not only the univariate distributions but also their inter-variable and inter-site dependence structures. Moreover, the proposed R2D2 method provides some stochasticity since it can generate as many multivariate corrected outputs as the number of statistical dimensions (i.e., number of grid cell × number of climate variables) of the simulations to be corrected. It is based on an assumption of stability in time of the dependence structure - making it possible to deal with a high number of statistical dimensions - that lets the climate model drive the temporal properties and their changes in time. R2D2 is applied on temperature and precipitation reanalysis time series with respect to high-resolution reference data over the southeast of France (1506 grid cell). Bivariate, 1506-dimensional and 3012-dimensional versions of R2D2 are tested over a historical period and compared to a univariate BC. How the different BC methods behave in a climate change context is also illustrated with an application to regional climate simulations over the 2071-2100 period. The results indicate that the 1d-BC basically reproduces the climate model multivariate properties, 2d-R2D2 is only satisfying in the inter-variable context, 1506d-R2D2 strongly improves inter-site properties and 3012d-R2D2 is able to account for both. Applications of the proposed R2D2 method to various climate datasets are relevant for many impact studies. The perspectives of improvements are numerous, such as introducing stochasticity in the dependence itself, questioning its stability assumption, and accounting for temporal properties adjustment while including more physics in the adjustment procedures.
Entropy Information of Cardiorespiratory Dynamics in Neonates during Sleep.
Lucchini, Maristella; Pini, Nicolò; Fifer, William P; Burtchen, Nina; Signorini, Maria G
2017-05-01
Sleep is a central activity in human adults and characterizes most of the newborn infant life. During sleep, autonomic control acts to modulate heart rate variability (HRV) and respiration. Mechanisms underlying cardiorespiratory interactions in different sleep states have been studied but are not yet fully understood. Signal processing approaches have focused on cardiorespiratory analysis to elucidate this co-regulation. This manuscript proposes to analyze heart rate (HR), respiratory variability and their interrelationship in newborn infants to characterize cardiorespiratory interactions in different sleep states (active vs. quiet). We are searching for indices that could detect regulation alteration or malfunction, potentially leading to infant distress. We have analyzed inter-beat (RR) interval series and respiration in a population of 151 newborns, and followed up with 33 at 1 month of age. RR interval series were obtained by recognizing peaks of the QRS complex in the electrocardiogram (ECG), corresponding to the ventricles depolarization. Univariate time domain, frequency domain and entropy measures were applied. In addition, Transfer Entropy was considered as a bivariate approach able to quantify the bidirectional information flow from one signal (respiration) to another (RR series). Results confirm the validity of the proposed approach. Overall, HRV is higher in active sleep, while high frequency (HF) power characterizes more quiet sleep. Entropy analysis provides higher indices for SampEn and Quadratic Sample entropy (QSE) in quiet sleep. Transfer Entropy values were higher in quiet sleep and point to a major influence of respiration on the RR series. At 1 month of age, time domain parameters show an increase in HR and a decrease in variability. No entropy differences were found across ages. The parameters employed in this study help to quantify the potential for infants to adapt their cardiorespiratory responses as they mature. Thus, they could be useful as early markers of risk for infant cardiorespiratory vulnerabilities.
Single-Fraction Proton Beam Stereotactic Radiosurgery for Cerebral Arteriovenous Malformations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hattangadi-Gluth, Jona A.; Chapman, Paul H.; Kim, Daniel
2014-06-01
Purpose/Objective(s): To evaluate the obliteration rate and potential adverse effects of single-fraction proton beam stereotactic radiosurgery (PSRS) in patients with cerebral arteriovenous malformations (AVMs). Methods and Materials: From 1991 to 2010, 248 consecutive patients with 254 cerebral AVMs received single-fraction PSRS at our institution. The median AVM nidus volume was 3.5 cc (range, 0.1-28.1 cc), 23% of AVMs were in critical/deep locations (basal ganglia, thalamus, or brainstem), and the most common prescription dose was 15 Gy(relative biological effectiveness [RBE]). Univariable and multivariable analyses were performed to assess factors associated with obliteration and hemorrhage. Results: At a median follow-up time of 35 months (range, 6-198 months),more » 64.6% of AVMs were obliterated. The median time to total obliteration was 31 months (range, 6-127 months), and the 5-year and 10-year cumulative incidence of total obliteration was 70% and 91%, respectively. On univariable analysis, smaller target volume (hazard ratio [HR] 0.78, 95% confidence interval [CI] 0.86-0.93, P<.0001), smaller treatment volume (HR 0.93, 95% CI 0.90-0.96, P<.0001), higher prescription dose (HR 1.16, 95% CI 1.07-1.26, P=.001), and higher maximum dose (HR 1.14, 95% CI 1.05-1.23, P=.002) were associated with total obliteration. Deep/critical location was also associated with decreased likelihood of obliteration (HR 0.68, 95% CI 0.47-0.98, P=.04). On multivariable analysis, critical location (adjusted HR [AHR] 0.42, 95% CI 0.27-0.65, P<.001) and smaller target volume (AHR 0.81, 95% CI 0.68-0.97, P=.02) remained associated with total obliteration. Posttreatment hemorrhage occurred in 13 cases (5-year cumulative incidence of 7%), all among patients with less than total obliteration, and 3 of these events were fatal. The most common complication was seizure, controlled with medications, both acutely (8%) and in the long term (9.1%). Conclusions: The current series is the largest modern series of PSRS for cerebral AVMs. PSRS can achieve a high obliteration rate with minimal morbidity. Post-treatment hemorrhage remains a potentially fatal risk among patients who have not yet responded to treatment.« less
Dynamic compositional modeling of pedestrian crash counts on urban roads in Connecticut.
Serhiyenko, Volodymyr; Ivan, John N; Ravishanker, Nalini; Islam, Md Saidul
2014-03-01
Uncovering the temporal trend in crash counts provides a good understanding of the context for pedestrian safety. With a rareness of pedestrian crashes it is impossible to investigate monthly temporal effects with an individual segment/intersection level data, thus the time dependence should be derived from the aggregated level data. Most previous studies have used annual data to investigate the differences in pedestrian crashes between different regions or countries in a given year, and/or to look at time trends of fatal pedestrian injuries annually. Use of annual data unfortunately does not provide sufficient information on patterns in time trends or seasonal effects. This paper describes statistical methods uncovering patterns in monthly pedestrian crashes aggregated on urban roads in Connecticut from January 1995 to December 2009. We investigate the temporal behavior of injury severity levels, including fatal (K), severe injury (A), evident minor injury (B), and non-evident possible injury and property damage only (C and O), as proportions of all pedestrian crashes in each month, taking into consideration effects of time trend, seasonal variations and VMT (vehicle miles traveled). This type of dependent multivariate data is characterized by positive components which sum to one, and occurs in several applications in science and engineering. We describe a dynamic framework with vector autoregressions (VAR) for modeling and predicting compositional time series. Combining these predictions with predictions from a univariate statistical model for total crash counts will then enable us to predict pedestrian crash counts with the different injury severity levels. We compare these predictions with those obtained from fitting separate univariate models to time series of crash counts at each injury severity level. We also show that the dynamic models perform better than the corresponding static models. We implement the Integrated Nested Laplace Approximation (INLA) approach to enable fast Bayesian posterior computation. Taking CO injury severity level as a baseline for the compositional analysis, we conclude that there was a noticeable shift in the proportion of pedestrian crashes from injury severity A to B, while the increase for injury severity K was extremely small over time. This shift to the less severe injury level (from A to B) suggests that the overall safety on urban roads in Connecticut is improving. In January and February, there was some increase in the proportions for levels A and B over the baseline, indicating a seasonal effect. We found evidence that an increase in VMT would result in a decrease of proportions over the baseline for all injury severity levels. Our dynamic model uncovered a decreasing trend in all pedestrian crash counts before April 2005, followed by a noticeable increase and a flattening out until the end of the fitting period. This appears to be largely due to the behavior of injury severity level A pedestrian crashes. Copyright © 2013 Elsevier Ltd. All rights reserved.
Social Consequences of Alcohol Use among Urban Poor: A Cross-sectional Study in Kathmandu Valley.
Thapa, P; Mishra, S R; Bista, B; Dhungana, R R; Adhikari, N; Soti, L; Puri, S; Aryal, K K
2015-01-01
Nepal is not an exception to alcohol use; urban poor are more prone than the general population. The question of social consequences of alcohol use among urban poor remains largely unanswered in Nepal. Study explored the alcohol linked social consequences among the urban poor of Kathmandu Valley. Taking 422 urban poor from four squatter settlements of Kathmandu Valley, a cross-sectional study was carried out. A series of univariate and bivariate analysis were performed in R version 3.1.2. Four out of 10 current drinkers (42.86%, 95% CI: 31.4-54.3) encountered various social consequences. The number one consequence hitting 23.19% drinkers was money loss. Male drinkers were 4.43 times (95% CI: 1.810.8) more likely to face social consequences than their female counterparts. Being male frequent drinker increased the odds of social consequence 3.80 times (95% CI:1.3-11.0) than that of female frequent drinker. A behaviour change communication campaign needs initiation; male populace and frequent drinkers being the target.
Ulnar osteosarcoma in dogs: 30 cases (1992-2008).
Sivacolundhu, Ramesh K; Runge, Jeffrey J; Donovan, Taryn A; Barber, Lisa G; Saba, Corey F; Clifford, Craig A; de Lorimier, Louis-Philippe; Atwater, Stephen W; DiBernardi, Lisa; Freeman, Kim P; Bergman, Philip J
2013-07-01
To examine the biological behavior of ulnar osteosarcoma and evaluate predictors of survival time in dogs. Retrospective case series. 30 dogs with primary ulnar osteosarcoma. Medical records were reviewed. Variables recorded and examined to identify predictors of survival time were signalment, tumor location in the ulna, tumor length, serum alkaline phosphatase activity, surgery type, completeness of excision, tumor stage, tumor grade, histologic subtype, development of metastases, and use of chemotherapy. 30 cases were identified from 9 institutions. Eleven dogs were treated with partial ulnar ostectomy and 14 with amputation; in 5 dogs, a resection was not performed. Twenty-two dogs received chemotherapy. Median disease-free interval and survival time were 437 and 463 days, respectively. Negative prognostic factors for survival time determined via univariate analyses were histologic subtype and development of lung metastases. Telangiectatic or telangiectatic-mixed subtype (n = 5) was the only negative prognostic factor identified via multivariate analysis (median survival time, 208 days). Dogs with telangiectatic subtype were 6.99 times as likely to die of the disease. The prognosis for ulnar osteosarcoma in this population was no worse and may have been better than the prognosis for dogs with osteosarcoma involving other appendicular sites. Partial ulnar ostectomy was associated with a low complication rate and good to excellent function and did not compromise survival time. Telangiectatic or telangiectatic-mixed histologic subtype was a negative prognostic factor for survival time. The efficacy of chemotherapy requires further evaluation.
Mai, Qun; Aboagye-Sarfo, Patrick; Sanfilippo, Frank M; Preen, David B; Fatovich, Daniel M
2015-02-01
To predict the number of ED presentations in Western Australia (WA) in the next 5 years, stratified by place of treatment, age, triage and disposition. We conducted a population-based time series analysis of 7 year monthly WA statewide ED presentation data from the financial years 2006/07 to 2012/13 using univariate autoregressive integrated moving average (ARIMA) and multivariate vector-ARIMA techniques. ED presentations in WA were predicted to increase from 990,342 in 2012/13 to 1,250,991 (95% CI: 982,265-1,519,718) in 2017/18, an increase of 260,649 (or 26.3%). The majority of this increase would occur in metropolitan WA (84.2%). The compound annual growth rate (CAGR) in metropolitan WA in the next 5 years was predicted to be 6.5% compared with 2.0% in the non-metropolitan area. The greatest growth in metropolitan WA would be in ages 65 and over (CAGR, 6.9%), triage categories 2 and 3 (8.3% and 7.7%, respectively) and admitted (9.8%) cohorts. The only predicted decrease was triage category 5 (-5.3%). ED demand in WA will exceed population growth. The highest growth will be in patients with complex care needs. An integrated system-wide strategy is urgently required to ensure access, quality and sustainability of the health system. © 2015 Australasian College for Emergency Medicine and Australasian Society for Emergency Medicine.
Synchronous Motions Across the Instrumental Climate Record
NASA Astrophysics Data System (ADS)
Carl, Peter
The Earth's climate system bears a rich variety of feedback mechanisms that may give rise to complex, evolving modal structures under internal and external control. Various types of synchronization may be identified in the system's motion when looking at representative time series of the instrumental period through the glasses of an advanced technique of sparse data approximation, the Matching Pursuit (MP) approach. To disentangle the emerging network of oscillatory modes to the degree that climate dynamics turns out to be separable, a large dictionary of "Gaussian logons," i.e. frequency modulated (FM) Gabor atoms, is applied. Though the extracted modes make up linear decompositions, this flexible analyzing signal matches highly nonlinear waveforms. Univariate analyses over the period 1870-1997 are presented of a set of customary time series in annual resolution, comprising global and regional climate, central European synoptic systems, German precipitation, and runoff of the Elbe river near Dresden. All the evidence from this first-generation MP-FM study, obtained in subsequent multivariate syntheses, points to dynamically excited regimes of an organized yet complex climate system under permanent change—perhaps a (pre)chaotic one at centennial timescales, suggesting a "chaos control" perspective on global climate dynamics and change. Findings and conclusions include, among others, internal structure of reconstructed insolation, the episodic nature of global warming as reflected in multidecadal temperature modes, their swarm of "interdomain" companions across the whole system that unveils an unknown regime character of interannual climate dynamics, and the apparent onset early in the 1990s of the present thermal stagnation.
Changes in Landscape Greenness and Climatic Factors over ...
Monitoring and quantifying changes in vegetation cover over large areas using remote sensing can be achieved using the Normalized Difference Vegetation Index (NDVI), an indicator of greenness. However, distinguishing gradual shifts in NDVI (e.g. climate change) versus direct and rapid changes (e.g., fire, land development) is challenging as changes can be confounded by time-dependent patterns, and variation associated with climatic factors. In the present study we leveraged a method, that we previously developed for a pilot study, to address these confounding factors by evaluating NDVI change using autoregression techniques that compare results from univariate (NDVI vs. time) and multivariate analyses (NDVI vs. time and climatic factors) for ~7,660,636 1-km2 pixels comprising the 48 contiguous states of the USA, over a 25-year period (1989−2013). NDVI changed significantly for 48% of the nation over the 25-year in the univariate analyses where most significant trends (85%) indicated an increase in greenness over time. By including climatic factors in the multivariate analyses of NDVI over time, the detection of significant NDVI trends increased to 53% (an increase of 5%). Comparisons of univariate and multivariate analyses for each pixel showed that less than 4% of the pixels had a significant NDVI trend attributable to gradual climatic changes while the remainder of pixels with a significant NDVI trend indicated that changes were due to direct factors. Whi
Adolescent Immunization Coverage and Implementation of New School Requirements in Michigan, 2010
DeVita, Stefanie F.; Vranesich, Patricia A.; Boulton, Matthew L.
2014-01-01
Objectives. We examined the effect of Michigan’s new school rules and vaccine coadministration on time to completion of all the school-required vaccine series, the individual adolescent vaccines newly required for sixth grade in 2010, and initiation of the human papillomavirus (HPV) vaccine series, which was recommended but not required for girls. Methods. Data were derived from the Michigan Care Improvement Registry, a statewide Immunization Information System. We assessed the immunization status of Michigan children enrolled in sixth grade in 2009 or 2010. We used univariable and multivariable Cox regression models to identify significant associations between each factor and school completeness. Results. Enrollment in sixth grade in 2010 and coadministration of adolescent vaccines at the first adolescent visit were significantly associated with completion of the vaccines required for Michigan’s sixth graders. Children enrolled in sixth grade in 2010 had higher coverage with the newly required adolescent vaccines by age 13 years than did sixth graders in 2009, but there was little difference in the rate of HPV vaccine initiation among girls. Conclusions. Education and outreach efforts, particularly regarding the importance and benefits of coadministration of all recommended vaccines in adolescents, should be directed toward health care providers, parents, and adolescents. PMID:24922144
Papadia, Andrea; Bellati, Filippo; Bogani, Giorgio; Ditto, Antonino; Martinelli, Fabio; Lorusso, Domenica; Donfrancesco, Cristina; Gasparri, Maria Luisa; Raspagliesi, Francesco
2015-12-01
The aim of this study was to identify clinical variables that may predict the need for adjuvant radiotherapy after neoadjuvant chemotherapy (NACT) and radical surgery in locally advanced cervical cancer patients. A retrospective series of cervical cancer patients with International Federation of Gynecology and Obstetrics (FIGO) stages IB2-IIB treated with NACT followed by radical surgery was analyzed. Clinical predictors of persistence of intermediate- and/or high-risk factors at final pathological analysis were investigated. Statistical analysis was performed using univariate and multivariate analysis and using a model based on artificial intelligence known as artificial neuronal network (ANN) analysis. Overall, 101 patients were available for the analyses. Fifty-two (51 %) patients were considered at high risk secondary to parametrial, resection margin and/or lymph node involvement. When disease was confined to the cervix, four (4 %) patients were considered at intermediate risk. At univariate analysis, FIGO grade 3, stage IIB disease at diagnosis and the presence of enlarged nodes before NACT predicted the presence of intermediate- and/or high-risk factors at final pathological analysis. At multivariate analysis, only FIGO grade 3 and tumor diameter maintained statistical significance. The specificity of ANN models in evaluating predictive variables was slightly superior to conventional multivariable models. FIGO grade, stage, tumor diameter, and histology are associated with persistence of pathological intermediate- and/or high-risk factors after NACT and radical surgery. This information is useful in counseling patients at the time of treatment planning with regard to the probability of being subjected to pelvic radiotherapy after completion of the initially planned treatment.
An algebraic method for constructing stable and consistent autoregressive filters
DOE Office of Scientific and Technical Information (OSTI.GOV)
Harlim, John, E-mail: jharlim@psu.edu; Department of Meteorology, the Pennsylvania State University, University Park, PA 16802; Hong, Hoon, E-mail: hong@ncsu.edu
2015-02-15
In this paper, we introduce an algebraic method to construct stable and consistent univariate autoregressive (AR) models of low order for filtering and predicting nonlinear turbulent signals with memory depth. By stable, we refer to the classical stability condition for the AR model. By consistent, we refer to the classical consistency constraints of Adams–Bashforth methods of order-two. One attractive feature of this algebraic method is that the model parameters can be obtained without directly knowing any training data set as opposed to many standard, regression-based parameterization methods. It takes only long-time average statistics as inputs. The proposed method provides amore » discretization time step interval which guarantees the existence of stable and consistent AR model and simultaneously produces the parameters for the AR models. In our numerical examples with two chaotic time series with different characteristics of decaying time scales, we find that the proposed AR models produce significantly more accurate short-term predictive skill and comparable filtering skill relative to the linear regression-based AR models. These encouraging results are robust across wide ranges of discretization times, observation times, and observation noise variances. Finally, we also find that the proposed model produces an improved short-time prediction relative to the linear regression-based AR-models in forecasting a data set that characterizes the variability of the Madden–Julian Oscillation, a dominant tropical atmospheric wave pattern.« less
Prognostic significance of blood coagulation tests in carcinoma of the lung and colon.
Wojtukiewicz, M Z; Zacharski, L R; Moritz, T E; Hur, K; Edwards, R L; Rickles, F R
1992-08-01
Blood coagulation test results were collected prospectively in patients with previously untreated, advanced lung or colon cancer who entered into a clinical trial. In patients with colon cancer, reduced survival was associated (in univariate analysis) with higher values obtained at entry to the study for fibrinogen, fibrin(ogen) split products, antiplasmin, and fibrinopeptide A and accelerated euglobulin lysis times. In patients with non-small cell lung cancer, reduced survival was associated (in univariate analysis) with higher fibrinogen and fibrin(ogen) split products, platelet counts and activated partial thromboplastin times. In patients with small cell carcinoma of the lung, only higher activated partial thromboplastin times were associated (in univariate analysis) with reduced survival in patients with disseminated disease. In multivariate analysis, higher activated partial thromboplastin times were a significant independent predictor of survival for patients with non-small cell lung cancer limited to one hemithorax and with disseminated small cell carcinoma of the lung. Fibrin(ogen) split product levels were an independent predictor of survival for patients with disseminated non-small cell lung cancer as were both the fibrinogen and fibrinopeptide A levels for patients with disseminated colon cancer. These results suggest that certain tests of blood coagulation may be indicative of prognosis in lung and colon cancer. The heterogeneity of these results suggests that the mechanism(s), intensity, and pathophysiological significance of coagulation activation in cancer may differ between tumour types.
Kaye, Stephen B
2009-04-01
To provide a scalar measure of refractive error, based on geometric lens power through principal, orthogonal and oblique meridians, that is not limited to the paraxial and sag height approximations. A function is derived to model sections through the principal meridian of a lens, followed by rotation of the section through orthogonal and oblique meridians. Average focal length is determined using the definition for the average of a function. Average univariate power in the principal meridian (including spherical aberration), can be computed from the average of a function over the angle of incidence as determined by the parameters of the given lens, or adequately computed from an integrated series function. Average power through orthogonal and oblique meridians, can be similarly determined using the derived formulae. The widely used computation for measuring refractive error, the spherical equivalent, introduces non-constant approximations, leading to a systematic bias. The equations proposed provide a good univariate representation of average lens power and are not subject to a systematic bias. They are particularly useful for the analysis of aggregate data, correlating with biological treatment variables and for developing analyses, which require a scalar equivalent representation of refractive power.
Laryngospasm during emergency department ketamine sedation: a case-control study.
Green, Steven M; Roback, Mark G; Krauss, Baruch
2010-11-01
The objective of this study was to assess predictors of emergency department (ED) ketamine-associated laryngospasm using case-control techniques. We performed a matched case-control analysis of a sample of 8282 ED ketamine sedations (including 22 occurrences of laryngospasm) assembled from 32 prior published series. We sequentially studied the association of each of 7 clinical variables with laryngospasm by assigning 4 controls to each case while matching for the remaining 6 variables. We then used univariate statistics and conditional logistic regression to analyze the matched sets. We found no statistical association of age, dose, oropharyngeal procedure, underlying physical illness, route, or coadministered anticholinergics with laryngospasm. Coadministered benzodiazepines showed a borderline association in the multivariate but not univariate analysis that was considered anomalous. This case-control analysis of the largest available sample of ED ketamine-associated laryngospasm did not demonstrate evidence of association with age, dose, or other clinical factors. Such laryngospasm seems to be idiosyncratic, and accordingly, clinicians administering ketamine must be prepared for its rapid identification and management. Given no evidence that they decrease the risk of laryngospasm, coadministered anticholinergics seem unnecessary.
NASA Astrophysics Data System (ADS)
Amora Jofipasi, Chesilia; Miftahuddin; Hizir
2018-05-01
Weather is a phenomenon that occurs in certain areas that indicate a change in natural activity. Weather can be predicted using data in previous periods over a period. The purpose of this study is to get the best ETS model to predict the weather in Aceh Besar. The ETS model is a time series univariate forecasting method; its use focuses on trend and seasonal components. The data used are air temperature, dew point, sea level pressure, station pressure, visibility, wind speed, and sea surface temperature from January 2006 to December 2016. Based on AIC, AICc and BIC the smallest values obtained the conclusion that the ETS (M, N, A) is used to predict air temperature, and sea surface temperature, ETS (A, N, A) is used to predict dew point, sea level pressure and station pressure, ETS (A, A, N) is used to predict visibility, and ETS (A, N, N) is used to predict wind speed.
Prediction of the space adaptation syndrome
NASA Technical Reports Server (NTRS)
Reschke, M. F.; Homick, J. L.; Ryan, P.; Moseley, E. C.
1984-01-01
The univariate and multivariate relationships of provocative measures used to produce motion sickness symptoms were described. Normative subjects were used to develop and cross-validate sets of linear equations that optimally predict motion sickness in parabolic flights. The possibility of reducing the number of measurements required for prediction was assessed. After describing the variables verbally and statistically for 159 subjects, a factor analysis of 27 variables was completed to improve understanding of the relationships between variables and to reduce the number of measures for prediction purposes. The results of this analysis show that none of variables are significantly related to the responses to parabolic flights. A set of variables was selected to predict responses to KC-135 flights. A series of discriminant analyses were completed. Results indicate that low, moderate, or severe susceptibility could be correctly predicted 64 percent and 53 percent of the time on original and cross-validation samples, respectively. Both the factor analysis and the discriminant analysis provided no basis for reducing the number of tests.
Simultaneous calibration of ensemble river flow predictions over an entire range of lead times
NASA Astrophysics Data System (ADS)
Hemri, S.; Fundel, F.; Zappa, M.
2013-10-01
Probabilistic estimates of future water levels and river discharge are usually simulated with hydrologic models using ensemble weather forecasts as main inputs. As hydrologic models are imperfect and the meteorological ensembles tend to be biased and underdispersed, the ensemble forecasts for river runoff typically are biased and underdispersed, too. Thus, in order to achieve both reliable and sharp predictions statistical postprocessing is required. In this work Bayesian model averaging (BMA) is applied to statistically postprocess ensemble runoff raw forecasts for a catchment in Switzerland, at lead times ranging from 1 to 240 h. The raw forecasts have been obtained using deterministic and ensemble forcing meteorological models with different forecast lead time ranges. First, BMA is applied based on mixtures of univariate normal distributions, subject to the assumption of independence between distinct lead times. Then, the independence assumption is relaxed in order to estimate multivariate runoff forecasts over the entire range of lead times simultaneously, based on a BMA version that uses multivariate normal distributions. Since river runoff is a highly skewed variable, Box-Cox transformations are applied in order to achieve approximate normality. Both univariate and multivariate BMA approaches are able to generate well calibrated probabilistic forecasts that are considerably sharper than climatological forecasts. Additionally, multivariate BMA provides a promising approach for incorporating temporal dependencies into the postprocessed forecasts. Its major advantage against univariate BMA is an increase in reliability when the forecast system is changing due to model availability.
Jia, Zhongwei; Wan, Fangning; Zhu, Yao; Shi, Guohai; Zhang, Hailiang; Dai, Bo; Ye, Dingwei
2018-06-01
Previous studies have demonstrated that several members of the Forkhead-box (FOX) family of genes are associated with tumor progression and metastasis. The objective of the current study was to screen candidate FOX family genes identified from analysis of molecular networks in clear cell renal cell carcinoma (ccRCC). The expression of FOX family genes as well as FOX family-associated genes was examined, and Kaplan-Meier survival analysis was performed in The Cancer Genome Atlas (TCGA) cohort (n=525). Patient characteristics, including sex, age, tumor diameter, laterality, tumor-node-metastasis, tumor grade, stage, white blood cell count, platelet count, the levels of hemoglobin, overall survival (OS) and disease-free survival (DFS), were collected for univariate and multivariate Cox proportional hazards ratio analyses. A total of seven candidate FOX family genes were selected from the TCGA database subsequent to univariate and multivariate Cox proportional hazards ratio analyses. FOXA1, FOXA2, FOXD1, FOXD4L2, FOXK2 and FOXL1 were associated with poor OS time, while FOXA1, FOXA2, FOXD1 and FOXK2 were associated with poor DFS time (P<0.05). FOXN2 was associated with favorable outcomes for overall and disease-free survival (P<0.05). In the gene cluster network analysis, the expression of FOX family-associated genes, including nuclear receptor coactivator ( NCOA ) 1 , NADH-ubiquinone oxidoreductase flavoprotein 3 ( NDUFV3 ), phosphatidylserine decarboxylase ( PISD ) and pyruvate kinase liver and red blood cell ( PKLR ), were independent prognostic factors for OS in patients with ccRCC. Results of the present study revealed that the expression of FOX family genes, including FOXA1, FOXA2, FOXD1, FOXD4L2, FOXK2 and FOXL1 , and FOX family-associated genes, including NCOA1, NDUFV3, PISD and PKLR , are independent prognostic factors for patients with ccRCC.
Lopes Antunes, Ana Carolina; Dórea, Fernanda; Halasa, Tariq; Toft, Nils
2016-05-01
Surveillance systems are critical for accurate, timely monitoring and effective disease control. In this study, we investigated the performance of univariate process monitoring control algorithms in detecting changes in seroprevalence for endemic diseases. We also assessed the effect of sample size (number of sentinel herds tested in the surveillance system) on the performance of the algorithms. Three univariate process monitoring control algorithms were compared: Shewart p Chart(1) (PSHEW), Cumulative Sum(2) (CUSUM) and Exponentially Weighted Moving Average(3) (EWMA). Increases in seroprevalence were simulated from 0.10 to 0.15 and 0.20 over 4, 8, 24, 52 and 104 weeks. Each epidemic scenario was run with 2000 iterations. The cumulative sensitivity(4) (CumSe) and timeliness were used to evaluate the algorithms' performance with a 1% false alarm rate. Using these performance evaluation criteria, it was possible to assess the accuracy and timeliness of the surveillance system working in real-time. The results showed that EWMA and PSHEW had higher CumSe (when compared with the CUSUM) from week 1 until the end of the period for all simulated scenarios. Changes in seroprevalence from 0.10 to 0.20 were more easily detected (higher CumSe) than changes from 0.10 to 0.15 for all three algorithms. Similar results were found with EWMA and PSHEW, based on the median time to detection. Changes in the seroprevalence were detected later with CUSUM, compared to EWMA and PSHEW for the different scenarios. Increasing the sample size 10 fold halved the time to detection (CumSe=1), whereas increasing the sample size 100 fold reduced the time to detection by a factor of 6. This study investigated the performance of three univariate process monitoring control algorithms in monitoring endemic diseases. It was shown that automated systems based on these detection methods identified changes in seroprevalence at different times. Increasing the number of tested herds would lead to faster detection. However, the practical implications of increasing the sample size (such as the costs associated with the disease) should also be taken into account. Copyright © 2016 Elsevier B.V. All rights reserved.
Joint modelling of annual maximum drought severity and corresponding duration
NASA Astrophysics Data System (ADS)
Tosunoglu, Fatih; Kisi, Ozgur
2016-12-01
In recent years, the joint distribution properties of drought characteristics (e.g. severity, duration and intensity) have been widely evaluated using copulas. However, history of copulas in modelling drought characteristics obtained from streamflow data is still short, especially in semi-arid regions, such as Turkey. In this study, unlike previous studies, drought events are characterized by annual maximum severity (AMS) and corresponding duration (CD) which are extracted from daily streamflow of the seven gauge stations located in Çoruh Basin, Turkey. On evaluation of the various univariate distributions, the Exponential, Weibull and Logistic distributions are identified as marginal distributions for the AMS and CD series. Archimedean copulas, namely Ali-Mikhail-Haq, Clayton, Frank and Gumbel-Hougaard, are then employed to model joint distribution of the AMS and CD series. With respect to the Anderson Darling and Cramér-von Mises statistical tests and the tail dependence assessment, Gumbel-Hougaard copula is identified as the most suitable model for joint modelling of the AMS and CD series at each station. Furthermore, the developed Gumbel-Hougaard copulas are used to derive the conditional and joint return periods of the AMS and CD series which can be useful for designing and management of reservoirs in the basin.
Mass univariate analysis of event-related brain potentials/fields I: a critical tutorial review.
Groppe, David M; Urbach, Thomas P; Kutas, Marta
2011-12-01
Event-related potentials (ERPs) and magnetic fields (ERFs) are typically analyzed via ANOVAs on mean activity in a priori windows. Advances in computing power and statistics have produced an alternative, mass univariate analyses consisting of thousands of statistical tests and powerful corrections for multiple comparisons. Such analyses are most useful when one has little a priori knowledge of effect locations or latencies, and for delineating effect boundaries. Mass univariate analyses complement and, at times, obviate traditional analyses. Here we review this approach as applied to ERP/ERF data and four methods for multiple comparison correction: strong control of the familywise error rate (FWER) via permutation tests, weak control of FWER via cluster-based permutation tests, false discovery rate control, and control of the generalized FWER. We end with recommendations for their use and introduce free MATLAB software for their implementation. Copyright © 2011 Society for Psychophysiological Research.
Chan, Emily H.; Sahai, Vikram; Conrad, Corrie; Brownstein, John S.
2011-01-01
Background A variety of obstacles including bureaucracy and lack of resources have interfered with timely detection and reporting of dengue cases in many endemic countries. Surveillance efforts have turned to modern data sources, such as Internet search queries, which have been shown to be effective for monitoring influenza-like illnesses. However, few have evaluated the utility of web search query data for other diseases, especially those of high morbidity and mortality or where a vaccine may not exist. In this study, we aimed to assess whether web search queries are a viable data source for the early detection and monitoring of dengue epidemics. Methodology/Principal Findings Bolivia, Brazil, India, Indonesia and Singapore were chosen for analysis based on available data and adequate search volume. For each country, a univariate linear model was then built by fitting a time series of the fraction of Google search query volume for specific dengue-related queries from that country against a time series of official dengue case counts for a time-frame within 2003–2010. The specific combination of queries used was chosen to maximize model fit. Spurious spikes in the data were also removed prior to model fitting. The final models, fit using a training subset of the data, were cross-validated against both the overall dataset and a holdout subset of the data. All models were found to fit the data quite well, with validation correlations ranging from 0.82 to 0.99. Conclusions/Significance Web search query data were found to be capable of tracking dengue activity in Bolivia, Brazil, India, Indonesia and Singapore. Whereas traditional dengue data from official sources are often not available until after some substantial delay, web search query data are available in near real-time. These data represent valuable complement to assist with traditional dengue surveillance. PMID:21647308
Giebel, Ole; Wirtz, Anna; Nachreiner, Friedhelm
2008-04-01
In order to analyze whether impairments to health and well-being under flexible working hours can be predicted from specific characteristics of the work schedules, periodic components in flexible working hours and their interference with the circadian temperature rhythm were analyzed applying univariate and bivariate spectrum analyses to both time series. The resulting indicators of spectral power and phase shift of these components were then related to reported health impairments using regression analysis. The results show that a suppression of both the 24 and the 168 h components in the work schedules (i.e., a lack of periodicity) can be used to predict reported health impairments, and that if there are relatively strong 24 and 168 h components left in the work schedules, their phase difference with the temperature rhythm (as an indicator of the interference between working time and the circadian rhythm) further predicts impairment. The results indicate that the periodicity of working hours and the amount of (circadian) desynchronization induced by flexible work schedules can be used for predicting the impairing effects of flexible work schedules on health and well-being. The results can thus be used for evaluating and designing flexible shift rosters.
Ictal fear: Associations with age, gender, and other experiential phenomena.
Chong, Derek J; Dugan, Patricia
2016-09-01
The aim of this study was to determine the relationship of fear to other auras and to gender and age using a large database. The Epilepsy Phenome/Genome Project (EPGP) is a multicenter, multicontinental cross-sectional study in which ictal symptomatology and other data were ascertained in a standardized series of questionnaires then corroborated by epilepsy specialists. Auras were classified into subgroups of symptoms, with ictal fear, panic, or anxiety as a single category. Of 536 participants with focal epilepsy, 72 were coded as having ictal fear/panic/anxiety. Reviewing raw patient responses, 12 participants were deemed not to have fear, and 24 had inadequate data, leaving 36 (7%) of 512 with definite ictal fear. In univariate analyses, fear was significantly associated with auras historically considered temporal lobe in origin, including cephalic, olfactory, and visceral complaints; déjà vu; and derealization. On both univariate and multivariate stepwise analyses, fear was associated with jamais vu and auras with cardiac symptoms, dyspnea, and chest tightening. Expressive aphasia was associated with fear on univariate analysis only, but the general category of aphasias was associated with fear only in the multivariate model. There was no age or gender relationship with fear when compared to the overall population with focal epilepsy that was studied under the EPGP. Patients with ictal fear were more likely to have a right hemisphere seizure focus. Ictal fear was strongly associated with other auras considered to originate from the limbic system. No relationship of fear with age or gender was observed. Copyright © 2016 Elsevier Inc. All rights reserved.
Bernhardt, Denise; Adeberg, Sebastian; Bozorgmehr, Farastuk; Opfermann, Nils; Hoerner-Rieber, Juliane; König, Laila; Kappes, Jutta; Thomas, Michael; Herth, Felix; Heußel, Claus Peter; Warth, Arne; Debus, Jürgen; Steins, Martin; Rieken, Stefan
2017-08-01
The purpose of this study was to evaluate prognostic factors associated with overall survival (OS) and neurological progression free survival (nPFS) in small-cell lung cancer (SCLC) patients with brain metastases who received whole-brain radiotherapy (WBRT). From 2003 to 2015, 229 SCLC patients diagnosed with brain metastases who received WBRT were analyzed retrospectively. In this cohort 219 patients (95%) received a total photon dose of 30 Gy in 10 fractions. The prognostic factors evaluated for OS and nPFS were: age, Karnofsky Performance Status (KPS), number of brain metastases, synchronous versus metachronous disease, initial response to chemotherapy, the Radiation Therapy Oncology Group recursive partitioning analysis (RPA) class and thoracic radiation. Median OS after WBRT was 6 months and the median nPFS after WBRT was 11 months. Patients with synchronous cerebral metastases had a significantly better median OS with 8 months compared to patients with metachronous metastases with a median survival of 3 months (p < 0.0001; HR 0.46; 95% CI 0.31-0.67). Based on RPA classification median survival after WBRT was 17 months in RPA class I, 7 months in class II and 3 months in class III (p < 0.0001). Karnofsky performance status scale (KPS < 70%) was significantly associated with OS in both univariate (HR 2.84; p < 0.001) and multivariate analyses (HR 2.56; p = 0.011). Further, metachronous brain metastases (HR 1.8; p < 0.001), initial response to first-line chemotherapy (HR 0.51, p < 0.001) and RPA class III (HR 2.74; p < 0.001) were significantly associated with OS in univariate analysis. In multivariate analysis metachronous disease (HR 1.89; p < 0.001) and initial response to chemotherapy (HR 0.61; p < 0.001) were further identified as significant prognostic factors. NPFS was negatively significantly influenced by poor KPS (HR 2.56; p = 0.011), higher number of brain metastases (HR 1.97; p = 0.02), and higher RPA class (HR 2.26; p = 0.03) in univariate analysis. In this series, the main prognostic factors associated with OS were performance status, time of appearance of intracranial disease (synchronous vs. metachronous), initial response to chemotherapy and higher RPA class. NPFS was negatively influenced by poor KPS, multiplicity of brain metastases, and higher RPA class in univariate analysis. For patients with low performance status, metachronous disease or RPA class III, WBRT should be weighed against supportive therapy with steroids alone or palliative chemotherapy.
de Brito, Aila Riany; Santos Reis, Nadabe Dos; Silva, Tatielle Pereira; Ferreira Bonomo, Renata Cristina; Trovatti Uetanabaro, Ana Paula; de Assis, Sandra Aparecida; da Silva, Erik Galvão Paranhos; Aguiar-Oliveira, Elizama; Oliveira, Julieta Rangel; Franco, Marcelo
2017-11-26
Endoglucanase production by Aspergillus oryzae ATCC 10124 cultivated in rice husks or peanut shells was optimized by experimental design as a function of humidity, time, and temperature. The optimum temperature for the endoglucanase activity was estimated by a univariate analysis (one factor at the time) as 50°C (rice husks) and 60°C (peanut shells), however, by a multivariate analysis (synergism of factors), it was determined a different temperature (56°C) for endoglucanase from peanut shells. For the optimum pH, values determined by univariate and multivariate analysis were 5 and 5.2 (rice husk) and 5 and 7.6 (peanut shells). In addition, the best half-lives were observed at 50°C as 22.8 hr (rice husks) and 7.3 hr (peanut shells), also, 80% of residual activities was obtained between 30 and 50°C for both substrates, and the pH stability was improved at 5-7 (rice hulls) and 6-9 (peanut shells). Both endoglucanases obtained presented different characteristics as a result of the versatility of fungi in different substrates.
A Guideline to Univariate Statistical Analysis for LC/MS-Based Untargeted Metabolomics-Derived Data
Vinaixa, Maria; Samino, Sara; Saez, Isabel; Duran, Jordi; Guinovart, Joan J.; Yanes, Oscar
2012-01-01
Several metabolomic software programs provide methods for peak picking, retention time alignment and quantification of metabolite features in LC/MS-based metabolomics. Statistical analysis, however, is needed in order to discover those features significantly altered between samples. By comparing the retention time and MS/MS data of a model compound to that from the altered feature of interest in the research sample, metabolites can be then unequivocally identified. This paper reports on a comprehensive overview of a workflow for statistical analysis to rank relevant metabolite features that will be selected for further MS/MS experiments. We focus on univariate data analysis applied in parallel on all detected features. Characteristics and challenges of this analysis are discussed and illustrated using four different real LC/MS untargeted metabolomic datasets. We demonstrate the influence of considering or violating mathematical assumptions on which univariate statistical test rely, using high-dimensional LC/MS datasets. Issues in data analysis such as determination of sample size, analytical variation, assumption of normality and homocedasticity, or correction for multiple testing are discussed and illustrated in the context of our four untargeted LC/MS working examples. PMID:24957762
A Guideline to Univariate Statistical Analysis for LC/MS-Based Untargeted Metabolomics-Derived Data.
Vinaixa, Maria; Samino, Sara; Saez, Isabel; Duran, Jordi; Guinovart, Joan J; Yanes, Oscar
2012-10-18
Several metabolomic software programs provide methods for peak picking, retention time alignment and quantification of metabolite features in LC/MS-based metabolomics. Statistical analysis, however, is needed in order to discover those features significantly altered between samples. By comparing the retention time and MS/MS data of a model compound to that from the altered feature of interest in the research sample, metabolites can be then unequivocally identified. This paper reports on a comprehensive overview of a workflow for statistical analysis to rank relevant metabolite features that will be selected for further MS/MS experiments. We focus on univariate data analysis applied in parallel on all detected features. Characteristics and challenges of this analysis are discussed and illustrated using four different real LC/MS untargeted metabolomic datasets. We demonstrate the influence of considering or violating mathematical assumptions on which univariate statistical test rely, using high-dimensional LC/MS datasets. Issues in data analysis such as determination of sample size, analytical variation, assumption of normality and homocedasticity, or correction for multiple testing are discussed and illustrated in the context of our four untargeted LC/MS working examples.
NASA Astrophysics Data System (ADS)
Keener, V. W.; Feyereisen, G. W.; Lall, U.; Jones, J. W.; Bosch, D. D.; Lowrance, R.
2010-02-01
SummaryAs climate variability increases, it is becoming increasingly critical to find predictable patterns that can still be identified despite overall uncertainty. The El-Niño/Southern Oscillation is the best known pattern. Its global effects on weather, hydrology, ecology and human health have been well documented. Climate variability manifested through ENSO has strong effects in the southeast United States, seen in precipitation and stream flow data. However, climate variability may also affect water quality in nutrient concentrations and loads, and have impacts on ecosystems, health, and food availability in the southeast. In this research, we establish a teleconnection between ENSO and the Little River Watershed (LRW), GA., as seen in a shared 3-7 year mode of variability for precipitation, stream flow, and nutrient load time series. Univariate wavelet analysis of the NINO 3.4 index of sea surface temperature (SST) and of precipitation, stream flow, NO 3 concentration and load time series from the watershed was used to identify common signals. Shared 3-7 year modes of variability were seen in all variables, most strongly in precipitation, stream flow and nutrient load in strong El Niño years. The significance of shared 3-7 year periodicity over red noise with 95% confidence in SST and precipitation, stream flow, and NO 3 load time series was confirmed through cross-wavelet and wavelet-coherence transforms, in which common high power and co-variance were computed for each set of data. The strongest 3-7 year shared power was seen in SST and stream flow data, while the strongest co-variance was seen in SST and NO 3 load data. The strongest cross-correlation was seen as a positive value between the NINO 3.4 and NO 3 load with a three-month lag. The teleconnection seen in the LRW between the NINO 3.4 index and precipitation, stream flow, and NO 3 load can be utilized in a model to predict monthly nutrient loads based on short-term climate variability, facilitating management in high risk seasons.
McGrath, Trevor A; McInnes, Matthew D F; Korevaar, Daniël A; Bossuyt, Patrick M M
2016-10-01
Purpose To determine whether authors of systematic reviews of diagnostic accuracy studies published in imaging journals used recommended methods for meta-analysis, and to evaluate the effect of traditional methods on summary estimates of sensitivity and specificity. Materials and Methods Medline was searched for published systematic reviews that included meta-analysis of test accuracy data limited to imaging journals published from January 2005 to May 2015. Two reviewers independently extracted study data and classified methods for meta-analysis as traditional (univariate fixed- or random-effects pooling or summary receiver operating characteristic curve) or recommended (bivariate model or hierarchic summary receiver operating characteristic curve). Use of methods was analyzed for variation with time, geographical location, subspecialty, and journal. Results from reviews in which study authors used traditional univariate pooling methods were recalculated with a bivariate model. Results Three hundred reviews met the inclusion criteria, and in 118 (39%) of those, authors used recommended meta-analysis methods. No change in the method used was observed with time (r = 0.54, P = .09); however, there was geographic (χ(2) = 15.7, P = .001), subspecialty (χ(2) = 46.7, P < .001), and journal (χ(2) = 27.6, P < .001) heterogeneity. Fifty-one univariate random-effects meta-analyses were reanalyzed with the bivariate model; the average change in the summary estimate was -1.4% (P < .001) for sensitivity and -2.5% (P < .001) for specificity. The average change in width of the confidence interval was 7.7% (P < .001) for sensitivity and 9.9% (P ≤ .001) for specificity. Conclusion Recommended methods for meta-analysis of diagnostic accuracy in imaging journals are used in a minority of reviews; this has not changed significantly with time. Traditional (univariate) methods allow overestimation of diagnostic accuracy and provide narrower confidence intervals than do recommended (bivariate) methods. (©) RSNA, 2016 Online supplemental material is available for this article.
Boyer, Alexandre; Couallier, Vincent; Clouzeau, Benjamin; Lasheras, Agnes; M'zali, Fatima; Kann, Michael; Rogues, Anne-Marie; Gruson, Didier
2015-12-01
This study was undertaken to determine the temporal relationship between implementation of different interventions in an intensive care unit (ICU) and control of endemic nosocomial acquisition of extended-spectrum β-lactamase Enterobacteriaceae (ESBLE). This was a prospective observational study with time-series analysis of the monthly incidence of ESBLE and its predictors. In November 2007, after a 14-month baseline period, an intervention consisting of restriction of third-generation cephalosporins (3 GC) and increased use of alcohol-based hand rubs was implemented. In January 2008, an increased health care worker (HCW):patient ratio was also implemented. In March 2010, the ICU was closed, and patients were moved to a clean ICU. The first intervention resulted in global reduction in 3 GC and increased use of alcohol-based hand rub. A significant change in ESBLE incidence was observed in a full segmented univariate regression analysis (mean change in level, -0.91 ± 0.19; P < .0001). After ICU closure, there was a dramatic reduction in ESBLE acquisition. According to the multivariate model, the ICU closure was the main protective factor. Before ICU closure, an increase in the HCW:patient ratio of 0.1 point tended to be associated with a decreased risk of ESBLE acquisition (relative risk, 0.28; 95% confidence interval, 0.06-1.25; P = .09). This study shows that ICU closure was associated with, but not necessarily the reason for, control of ESBLE cross-transmission in a nonoutbreak setting. Environmental ESBE sources may play a role in cross-transmission. Copyright © 2015 Association for Professionals in Infection Control and Epidemiology, Inc. Published by Elsevier Inc. All rights reserved.
Malm, Keziah; Peprah, Nana Yaw; Silal, Sheetal P.
2018-01-01
Background Malaria incidence is largely influenced by vector abundance. Among the many interconnected factors relating to malaria transmission, weather conditions such as rainfall and temperature are known to create suitable environmental conditions that sustain reproduction and propagation of anopheles mosquitoes and malaria parasites. In Ghana, climatic conditions vary across the country. Understanding the heterogeneity of malaria morbidity using data sourced from a recently setup data repository for routine health facility data could support planning. Methods Monthly aggregated confirmed uncomplicated malaria cases from the District Health Information Management System and average monthly rainfall and temperature records obtained from the Ghana Meteorological Agency from 2008 to 2016 were analysed. Univariate time series models were fitted to the malaria, rainfall and temperature data series. After pre-whitening the morbidity data, cross correlation analyses were performed. Subsequently, transfer function models were developed for the relationship between malaria morbidity and rainfall and temperature. Results Malaria morbidity patterns vary across zones. In the Guinea savannah, morbidity peaks once in the year and twice in both the Transitional forest and Coastal savannah, following similar patterns of rainfall at the zonal level. While the effects of rainfall on malaria morbidity are delayed by a month in the Guinea savannah and Transitional Forest zones those of temperature are delayed by two months in the Transitional forest zone. In the Coastal savannah however, incidence of malaria is significantly associated with two months lead in rainfall and temperature. Conclusion Data captured on the District Health Information Management System has been used to demonstrate heterogeneity in the dynamics of malaria morbidity across the country. Timing of these variations could guide the deployment of interventions such as indoor residual spraying, Seasonal Malaria Chemoprevention or vaccines to optimise effectiveness on zonal basis. PMID:29377908
Awine, Timothy; Malm, Keziah; Peprah, Nana Yaw; Silal, Sheetal P
2018-01-01
Malaria incidence is largely influenced by vector abundance. Among the many interconnected factors relating to malaria transmission, weather conditions such as rainfall and temperature are known to create suitable environmental conditions that sustain reproduction and propagation of anopheles mosquitoes and malaria parasites. In Ghana, climatic conditions vary across the country. Understanding the heterogeneity of malaria morbidity using data sourced from a recently setup data repository for routine health facility data could support planning. Monthly aggregated confirmed uncomplicated malaria cases from the District Health Information Management System and average monthly rainfall and temperature records obtained from the Ghana Meteorological Agency from 2008 to 2016 were analysed. Univariate time series models were fitted to the malaria, rainfall and temperature data series. After pre-whitening the morbidity data, cross correlation analyses were performed. Subsequently, transfer function models were developed for the relationship between malaria morbidity and rainfall and temperature. Malaria morbidity patterns vary across zones. In the Guinea savannah, morbidity peaks once in the year and twice in both the Transitional forest and Coastal savannah, following similar patterns of rainfall at the zonal level. While the effects of rainfall on malaria morbidity are delayed by a month in the Guinea savannah and Transitional Forest zones those of temperature are delayed by two months in the Transitional forest zone. In the Coastal savannah however, incidence of malaria is significantly associated with two months lead in rainfall and temperature. Data captured on the District Health Information Management System has been used to demonstrate heterogeneity in the dynamics of malaria morbidity across the country. Timing of these variations could guide the deployment of interventions such as indoor residual spraying, Seasonal Malaria Chemoprevention or vaccines to optimise effectiveness on zonal basis.
A Balanced Approach to Adaptive Probability Density Estimation.
Kovacs, Julio A; Helmick, Cailee; Wriggers, Willy
2017-01-01
Our development of a Fast (Mutual) Information Matching (FIM) of molecular dynamics time series data led us to the general problem of how to accurately estimate the probability density function of a random variable, especially in cases of very uneven samples. Here, we propose a novel Balanced Adaptive Density Estimation (BADE) method that effectively optimizes the amount of smoothing at each point. To do this, BADE relies on an efficient nearest-neighbor search which results in good scaling for large data sizes. Our tests on simulated data show that BADE exhibits equal or better accuracy than existing methods, and visual tests on univariate and bivariate experimental data show that the results are also aesthetically pleasing. This is due in part to the use of a visual criterion for setting the smoothing level of the density estimate. Our results suggest that BADE offers an attractive new take on the fundamental density estimation problem in statistics. We have applied it on molecular dynamics simulations of membrane pore formation. We also expect BADE to be generally useful for low-dimensional applications in other statistical application domains such as bioinformatics, signal processing and econometrics.
Prolonged Instability Prior to a Regime Shift | Science ...
Regime shifts are generally defined as the point of ‘abrupt’ change in the state of a system. However, a seemingly abrupt transition can be the product of a system reorganization that has been ongoing much longer than is evident in statistical analysis of a single component of the system. Using both univariate and multivariate statistical methods, we tested a long-term high-resolution paleoecological dataset with a known change in species assemblage for a regime shift. Analysis of this dataset with Fisher Information and multivariate time series modeling showed that there was a∼2000 year period of instability prior to the regime shift. This period of instability and the subsequent regime shift coincide with regional climate change, indicating that the system is undergoing extrinsic forcing. Paleoecological records offer a unique opportunity to test tools for the detection of thresholds and stable-states, and thus to examine the long-term stability of ecosystems over periods of multiple millennia. This manuscript explores various methods of assessing the transition between alternative states in an ecological system described by a long-term high-resolution paleoecological dataset.
Estimation of value at risk and conditional value at risk using normal mixture distributions model
NASA Astrophysics Data System (ADS)
Kamaruzzaman, Zetty Ain; Isa, Zaidi
2013-04-01
Normal mixture distributions model has been successfully applied in financial time series analysis. In this paper, we estimate the return distribution, value at risk (VaR) and conditional value at risk (CVaR) for monthly and weekly rates of returns for FTSE Bursa Malaysia Kuala Lumpur Composite Index (FBMKLCI) from July 1990 until July 2010 using the two component univariate normal mixture distributions model. First, we present the application of normal mixture distributions model in empirical finance where we fit our real data. Second, we present the application of normal mixture distributions model in risk analysis where we apply the normal mixture distributions model to evaluate the value at risk (VaR) and conditional value at risk (CVaR) with model validation for both risk measures. The empirical results provide evidence that using the two components normal mixture distributions model can fit the data well and can perform better in estimating value at risk (VaR) and conditional value at risk (CVaR) where it can capture the stylized facts of non-normality and leptokurtosis in returns distribution.
Using MetaboAnalyst 3.0 for Comprehensive Metabolomics Data Analysis.
Xia, Jianguo; Wishart, David S
2016-09-07
MetaboAnalyst (http://www.metaboanalyst.ca) is a comprehensive Web application for metabolomic data analysis and interpretation. MetaboAnalyst handles most of the common metabolomic data types from most kinds of metabolomics platforms (MS and NMR) for most kinds of metabolomics experiments (targeted, untargeted, quantitative). In addition to providing a variety of data processing and normalization procedures, MetaboAnalyst also supports a number of data analysis and data visualization tasks using a range of univariate, multivariate methods such as PCA (principal component analysis), PLS-DA (partial least squares discriminant analysis), heatmap clustering and machine learning methods. MetaboAnalyst also offers a variety of tools for metabolomic data interpretation including MSEA (metabolite set enrichment analysis), MetPA (metabolite pathway analysis), and biomarker selection via ROC (receiver operating characteristic) curve analysis, as well as time series and power analysis. This unit provides an overview of the main functional modules and the general workflow of the latest version of MetaboAnalyst (MetaboAnalyst 3.0), followed by eight detailed protocols. © 2016 by John Wiley & Sons, Inc. Copyright © 2016 John Wiley & Sons, Inc.
Quantile Regression for Recurrent Gap Time Data
Luo, Xianghua; Huang, Chiung-Yu; Wang, Lan
2014-01-01
Summary Evaluating covariate effects on gap times between successive recurrent events is of interest in many medical and public health studies. While most existing methods for recurrent gap time analysis focus on modeling the hazard function of gap times, a direct interpretation of the covariate effects on the gap times is not available through these methods. In this article, we consider quantile regression that can provide direct assessment of covariate effects on the quantiles of the gap time distribution. Following the spirit of the weighted risk-set method by Luo and Huang (2011, Statistics in Medicine 30, 301–311), we extend the martingale-based estimating equation method considered by Peng and Huang (2008, Journal of the American Statistical Association 103, 637–649) for univariate survival data to analyze recurrent gap time data. The proposed estimation procedure can be easily implemented in existing software for univariate censored quantile regression. Uniform consistency and weak convergence of the proposed estimators are established. Monte Carlo studies demonstrate the effectiveness of the proposed method. An application to data from the Danish Psychiatric Central Register is presented to illustrate the methods developed in this article. PMID:23489055
Uveal Melanoma Regression after Brachytherapy: Relationship with Chromosome 3 Monosomy Status.
Salvi, Sachin M; Aziz, Hassan A; Dar, Suhail; Singh, Nakul; Hayden-Loreck, Brandy; Singh, Arun D
2017-07-01
The objective was to evaluate the relationship between the regression rate of ciliary body melanoma and choroidal melanoma after brachytherapy and chromosome 3 monosomy status. We conducted a prospective and consecutive case series of patients who underwent biopsy and brachytherapy for ciliary/choroidal melanoma. Tumor biopsy performed at the time of radiation plaque placement was analyzed with fluorescence in situ hybridization to determine the percentage of tumor cells with chromosome 3 monosomy. The regression rate was calculated as the percent change in tumor height at months 3, 6, and 12. The relationship between regression rate and tumor location, initial tumor height, and chromosome 3 monosomy (percentage) was assessed by univariate linear regression (R version 3.1.0). Of the 75 patients included in the study, 8 had ciliary body melanoma, and 67 were choroidal melanomas. The mean tumor height at the time of diagnosis was 5.2 mm (range: 1.90-13.00). The percentage composition of chromosome 3 monosomy ranged from 0-20% (n = 35) to 81-100% (n = 40). The regression of tumor height at months 3, 6, and 12 did not statistically correlate with tumor location (ciliary or choroidal), initial tumor height, or chromosome 3 monosomy (percentage). The regression rate of choroidal melanoma following brachytherapy did not correlate with chromosome 3 monosomy status.
Sang, Shaowei; Yin, Wenwu; Bi, Peng; Zhang, Honglong; Wang, Chenggang; Liu, Xiaobo; Chen, Bin; Yang, Weizhong; Liu, Qiyong
2014-01-01
Introduction Each year there are approximately 390 million dengue infections worldwide. Weather variables have a significant impact on the transmission of Dengue Fever (DF), a mosquito borne viral disease. DF in mainland China is characterized as an imported disease. Hence it is necessary to explore the roles of imported cases, mosquito density and climate variability in dengue transmission in China. The study was to identify the relationship between dengue occurrence and possible risk factors and to develop a predicting model for dengue’s control and prevention purpose. Methodology and Principal Findings Three traditional suburbs and one district with an international airport in Guangzhou city were selected as the study areas. Autocorrelation and cross-correlation analysis were used to perform univariate analysis to identify possible risk factors, with relevant lagged effects, associated with local dengue cases. Principal component analysis (PCA) was applied to extract principal components and PCA score was used to represent the original variables to reduce multi-collinearity. Combining the univariate analysis and prior knowledge, time-series Poisson regression analysis was conducted to quantify the relationship between weather variables, Breteau Index, imported DF cases and the local dengue transmission in Guangzhou, China. The goodness-of-fit of the constructed model was determined by pseudo-R2, Akaike information criterion (AIC) and residual test. There were a total of 707 notified local DF cases from March 2006 to December 2012, with a seasonal distribution from August to November. There were a total of 65 notified imported DF cases from 20 countries, with forty-six cases (70.8%) imported from Southeast Asia. The model showed that local DF cases were positively associated with mosquito density, imported cases, temperature, precipitation, vapour pressure and minimum relative humidity, whilst being negatively associated with air pressure, with different time lags. Conclusions Imported DF cases and mosquito density play a critical role in local DF transmission, together with weather variables. The establishment of an early warning system, using existing surveillance datasets will help to control and prevent dengue in Guangzhou, China. PMID:25019967
Sang, Shaowei; Yin, Wenwu; Bi, Peng; Zhang, Honglong; Wang, Chenggang; Liu, Xiaobo; Chen, Bin; Yang, Weizhong; Liu, Qiyong
2014-01-01
Each year there are approximately 390 million dengue infections worldwide. Weather variables have a significant impact on the transmission of Dengue Fever (DF), a mosquito borne viral disease. DF in mainland China is characterized as an imported disease. Hence it is necessary to explore the roles of imported cases, mosquito density and climate variability in dengue transmission in China. The study was to identify the relationship between dengue occurrence and possible risk factors and to develop a predicting model for dengue's control and prevention purpose. Three traditional suburbs and one district with an international airport in Guangzhou city were selected as the study areas. Autocorrelation and cross-correlation analysis were used to perform univariate analysis to identify possible risk factors, with relevant lagged effects, associated with local dengue cases. Principal component analysis (PCA) was applied to extract principal components and PCA score was used to represent the original variables to reduce multi-collinearity. Combining the univariate analysis and prior knowledge, time-series Poisson regression analysis was conducted to quantify the relationship between weather variables, Breteau Index, imported DF cases and the local dengue transmission in Guangzhou, China. The goodness-of-fit of the constructed model was determined by pseudo-R2, Akaike information criterion (AIC) and residual test. There were a total of 707 notified local DF cases from March 2006 to December 2012, with a seasonal distribution from August to November. There were a total of 65 notified imported DF cases from 20 countries, with forty-six cases (70.8%) imported from Southeast Asia. The model showed that local DF cases were positively associated with mosquito density, imported cases, temperature, precipitation, vapour pressure and minimum relative humidity, whilst being negatively associated with air pressure, with different time lags. Imported DF cases and mosquito density play a critical role in local DF transmission, together with weather variables. The establishment of an early warning system, using existing surveillance datasets will help to control and prevent dengue in Guangzhou, China.
Alternating event processes during lifetimes: population dynamics and statistical inference.
Shinohara, Russell T; Sun, Yifei; Wang, Mei-Cheng
2018-01-01
In the literature studying recurrent event data, a large amount of work has been focused on univariate recurrent event processes where the occurrence of each event is treated as a single point in time. There are many applications, however, in which univariate recurrent events are insufficient to characterize the feature of the process because patients experience nontrivial durations associated with each event. This results in an alternating event process where the disease status of a patient alternates between exacerbations and remissions. In this paper, we consider the dynamics of a chronic disease and its associated exacerbation-remission process over two time scales: calendar time and time-since-onset. In particular, over calendar time, we explore population dynamics and the relationship between incidence, prevalence and duration for such alternating event processes. We provide nonparametric estimation techniques for characteristic quantities of the process. In some settings, exacerbation processes are observed from an onset time until death; to account for the relationship between the survival and alternating event processes, nonparametric approaches are developed for estimating exacerbation process over lifetime. By understanding the population dynamics and within-process structure, the paper provide a new and general way to study alternating event processes.
NASA Astrophysics Data System (ADS)
Niedzielski, Tomasz; Kosek, Wiesław
2008-02-01
This article presents the application of a multivariate prediction technique for predicting universal time (UT1-UTC), length of day (LOD) and the axial component of atmospheric angular momentum (AAM χ 3). The multivariate predictions of LOD and UT1-UTC are generated by means of the combination of (1) least-squares (LS) extrapolation of models for annual, semiannual, 18.6-year, 9.3-year oscillations and for the linear trend, and (2) multivariate autoregressive (MAR) stochastic prediction of LS residuals (LS + MAR). The MAR technique enables the use of the AAM χ 3 time-series as the explanatory variable for the computation of LOD or UT1-UTC predictions. In order to evaluate the performance of this approach, two other prediction schemes are also applied: (1) LS extrapolation, (2) combination of LS extrapolation and univariate autoregressive (AR) prediction of LS residuals (LS + AR). The multivariate predictions of AAM χ 3 data, however, are computed as a combination of the extrapolation of the LS model for annual and semiannual oscillations and the LS + MAR. The AAM χ 3 predictions are also compared with LS extrapolation and LS + AR prediction. It is shown that the predictions of LOD and UT1-UTC based on LS + MAR taking into account the axial component of AAM are more accurate than the predictions of LOD and UT1-UTC based on LS extrapolation or on LS + AR. In particular, the UT1-UTC predictions based on LS + MAR during El Niño/La Niña events exhibit considerably smaller prediction errors than those calculated by means of LS or LS + AR. The AAM χ 3 time-series is predicted using LS + MAR with higher accuracy than applying LS extrapolation itself in the case of medium-term predictions (up to 100 days in the future). However, the predictions of AAM χ 3 reveal the best accuracy for LS + AR.
The relevance of timing in nonconvulsive status epilepticus: A series of 38 cases.
Gutiérrez-Viedma, Álvaro; Parejo-Carbonell, Beatriz; Cuadrado, María-Luz; Serrano-García, Irene; Abarrategui, Belén; García-Morales, Irene
2018-05-01
Timing in the management of nonconvulsive status epilepticus (NCSE) seems to be one of the most important modifiable prognostic factors. We aimed to determine the precise relationship between timing in NCSE management and its outcome. We performed a cross-sectional study in which clinical data were prospectively obtained from all consecutive adults with NCSE admitted to our hospital from 2014 to 2016. Univariate and multivariable regression analyses were performed to identify clinical and timing variables associated with NCSE prognosis. Among 38 NCSE cases, 59.9% were women, and 39.5% had prior epilepsy history. The median time to treatment (TTT) initiation and the median time to assessment by a neurologist (TTN) were 5h, and the median time to first electroencephalography assessment was 18.5h; in the cases with out-of-hospital onset (n=24), the median time to hospital (TTH) arrival was 2.8h. The median time to NCSE control (TTC) was 16.5h, and it positively correlated with both the TTH (Spearman's rho: 0.439) and the TTT (Spearman's rho: 0.683). In the multivariable regression analyses, the TTC was extended 1.7h for each hour of hospital arrival delay (p=0.01) and 2.7h for each hour of treatment delay (p<0.001). Recognition delay was more common in the episodes with in-hospital onset, which also had longer TTN and TTC, and increased morbidity. There were pervasive delays in all phases of NCSE management. Delays in hospital arrival or treatment initiation may result in prolonged TTC. Recognition of in-hospital episodes may be more delayed, which may lead to poorer prognosis in these cases. Copyright © 2018 Elsevier Inc. All rights reserved.
Biacchi, Daniele; Sammartino, Paolo; Sibio, Simone; Accarpio, Fabio; Cardi, Maurizio; Sapienza, Paolo; De Cesare, Alessandro; Atta, Joseph Maher Fouad; Impagnatiello, Alessio; Di Giorgio, Angelo
2016-02-01
Totally implantable venous access ports (TIVAP) are eventually explanted for various reasons, related or unrelated to the implantation technique used. Having more information on long-term explantation would help improve placement techniques. From a series of 1572 cancer patients who had TIVAPs implanted in our center with the cutdown technique or Seldinger technique, we studied the 542 patients who returned to us to have their TIVAP explanted after 70 days or more. As outcome measures we distinguished between TIVAPs explanted for long-term complications (infection, catheter-, reservoir-, and patient-related complications) and TIVAPs no longer needed. Univariate and multivariate analyses were run to investigate the reasons for explantation and their possible correlation with implantation techniques. The most common reason for explantation was infection (47.6 %), followed by catheter-related (20.8 %), patient-related (14.7 %), and reservoir-related complications (4.7 %). In the remaining 12.2 % of cases, the TIVAP was explanted complication free after the planned treatments ended. Infection correlated closely with longer TIVAP use. Univariate and multivariate analyses identified the Seldinger technique as a major risk factor for venous thrombosis and catheter dislocation. The need for long-term TIVAP explantation in about one-third of cancer patients is related to the implantation techniques used.
EMMPRIN/CD147 is an independent prognostic biomarker in cutaneous melanoma.
Caudron, Anne; Battistella, Maxime; Feugeas, Jean-Paul; Pages, Cécile; Basset-Seguin, Nicole; Mazouz Dorval, Sarra; Funck Brentano, Elisa; Sadoux, Aurélie; Podgorniak, Marie-Pierre; Menashi, Suzanne; Janin, Anne; Lebbé, Céleste; Mourah, Samia
2016-08-01
CD147 has been implicated in melanoma invasion and metastasis mainly through increasing metalloproteinase synthesis and regulating VEGF/VEGFR signalling. In this study, the prognostic value of CD147 expression was investigated in a cohort of 196 cutaneous melanomas including 136 consecutive primary malignant melanomas, 30 lymph nodes, 16 in-transit and 14 visceral metastases. A series of 10 normal skin, 10 blue nevi and 10 dermal nevi was used as control. CD147 expression was assessed by immunohistochemistry, and the association of its expression with the clinicopathological characteristics of patients and survival was evaluated using univariate and multivariate statistical analyses. Univariate analysis showed that high CD147 expression was significantly associated with metastatic potential and with a reduced overall survival (P < 0.05 for both) in primary melanoma patients. CD147 expression level was correlated with histological factors which were associated with prognosis: Clark level, ulceration status and more particularly with Breslow index (r = 0.7, P < 10(-8) ). Multivariate analysis retained CD147 expression level and ulceration status as predicting factors for metastasis and overall survival (P < 0.05 for both). CD147 emerges as an important factor in the aggressive behaviour of melanoma and deserves further evaluation as an independent prognostic biomarker. © 2016 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.
Dias, Adriana Neves; da Silva, Ana Cristine; Simão, Vanessa; Merib, Josias; Carasek, Eduardo
2015-08-12
This study describes the use of cork as a new coating for bar adsorptive microextraction (BAμE) and its application in determining benzophenone, triclocarban and parabens in aqueous samples by HPLC-DAD. In this study bars with 7.5 and 15 mm of length were used. The extraction and liquid desorption steps for BAμE were optimized employing multivariate and univariate procedures. The desorption time and solvent used for liquid desorption were optimized by univariate and multivariate studies, respectively. For the extraction step the sample pH was optimized by univariate experiments while the parameters extraction time and ionic strength were evaluated using the Doehlert design. The optimum extraction conditions were sample pH 5.5, NaCl concentration 25% and extraction time 90 min. Liquid desorption was carried out for 30 min with 250 μL (bar length of 15 mm) or 100 μL (bar length of 7.5 mm) of ACN:MeOH (50:50, v/v). The quantification limits varied between 1.6 and 20 μg L(-1) (bar length of 15 mm) and 0.64 and 8 μg L(-1) (bar length of 7.5 mm). The linear correlation coefficients were higher than 0.98 for both bars. The method with 7.5 mm bar length showed recovery values between 65 and 123%. The bar-to-bar reproducibility and the repeatability were lower than 13% (n = 2) and 14% (n = 3), respectively. Copyright © 2015 Elsevier B.V. All rights reserved.
The status of diabetes control in Kurdistan province, west of Iran.
Esmailnasab, Nader; Afkhamzadeh, Abdorrahim; Roshani, Daem; Moradi, Ghobad
2013-09-17
Based on some estimation more than two million peoples in Iran are affected by Type 2 diabetes. The present study was designed to evaluate the status of diabetes control among Type 2 diabetes patients in Kurdistan, west of Iran and its associated factors. In our cross sectional study conducted in 2010, 411 Type 2 diabetes patients were randomly recruited from Sanandaj, Capital of Kurdistan. Chi square test was used in univariate analysis to address the association between HgAlc and FBS status and other variables. The significant results from Univariate analysis were entered in multivariate analysis and multinomial logistic regression model. In 38% of patients, FBS was in normal range (70-130) and in 47% HgA1c was <7% which is normal range for HgA1c. In univariate analysis, FBS level was associated with educational levels (P=0.001), referral style (P=0.001), referral time (P=0.009), and insulin injection (P=0.016). In addition, HgA1c had a relationship with sex (P=0.023), age (P=0.035), education (P=0.001), referral style (P=0.001), and insulin injection (P=0.008). After using multinomial logistic regression for significant results of univariate analysis, it was found that FBS was significantly associated with referral style. In addition HgA1c was significantly associated with referral style and Insulin injection. Although some of patients were under the coverage of specialized cares, but their diabetes were not properly controlled.
Athens, Jessica K.; Remington, Patrick L.; Gangnon, Ronald E.
2015-01-01
Objectives The University of Wisconsin Population Health Institute has published the County Health Rankings since 2010. These rankings use population-based data to highlight health outcomes and the multiple determinants of these outcomes and to encourage in-depth health assessment for all United States counties. A significant methodological limitation, however, is the uncertainty of rank estimates, particularly for small counties. To address this challenge, we explore the use of longitudinal and pooled outcome data in hierarchical Bayesian models to generate county ranks with greater precision. Methods In our models we used pooled outcome data for three measure groups: (1) Poor physical and poor mental health days; (2) percent of births with low birth weight and fair or poor health prevalence; and (3) age-specific mortality rates for nine age groups. We used the fixed and random effects components of these models to generate posterior samples of rates for each measure. We also used time-series data in longitudinal random effects models for age-specific mortality. Based on the posterior samples from these models, we estimate ranks and rank quartiles for each measure, as well as the probability of a county ranking in its assigned quartile. Rank quartile probabilities for univariate, joint outcome, and/or longitudinal models were compared to assess improvements in rank precision. Results The joint outcome model for poor physical and poor mental health days resulted in improved rank precision, as did the longitudinal model for age-specific mortality rates. Rank precision for low birth weight births and fair/poor health prevalence based on the univariate and joint outcome models were equivalent. Conclusion Incorporating longitudinal or pooled outcome data may improve rank certainty, depending on characteristics of the measures selected. For measures with different determinants, joint modeling neither improved nor degraded rank precision. This approach suggests a simple way to use existing information to improve the precision of small-area measures of population health. PMID:26098858
NASA Astrophysics Data System (ADS)
Cannon, Alex J.
2018-01-01
Most bias correction algorithms used in climatology, for example quantile mapping, are applied to univariate time series. They neglect the dependence between different variables. Those that are multivariate often correct only limited measures of joint dependence, such as Pearson or Spearman rank correlation. Here, an image processing technique designed to transfer colour information from one image to another—the N-dimensional probability density function transform—is adapted for use as a multivariate bias correction algorithm (MBCn) for climate model projections/predictions of multiple climate variables. MBCn is a multivariate generalization of quantile mapping that transfers all aspects of an observed continuous multivariate distribution to the corresponding multivariate distribution of variables from a climate model. When applied to climate model projections, changes in quantiles of each variable between the historical and projection period are also preserved. The MBCn algorithm is demonstrated on three case studies. First, the method is applied to an image processing example with characteristics that mimic a climate projection problem. Second, MBCn is used to correct a suite of 3-hourly surface meteorological variables from the Canadian Centre for Climate Modelling and Analysis Regional Climate Model (CanRCM4) across a North American domain. Components of the Canadian Forest Fire Weather Index (FWI) System, a complicated set of multivariate indices that characterizes the risk of wildfire, are then calculated and verified against observed values. Third, MBCn is used to correct biases in the spatial dependence structure of CanRCM4 precipitation fields. Results are compared against a univariate quantile mapping algorithm, which neglects the dependence between variables, and two multivariate bias correction algorithms, each of which corrects a different form of inter-variable correlation structure. MBCn outperforms these alternatives, often by a large margin, particularly for annual maxima of the FWI distribution and spatiotemporal autocorrelation of precipitation fields.
Aspirin and Statin Nonuse Associated With Early Biochemical Failure After Prostate Radiation Therapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zaorsky, Nicholas G.; Buyyounouski, Mark K., E-mail: mark.buyyounouski@fccc.edu; Li, Tianyu
Purpose: To present the largest retrospective series investigating the effect of aspirin and statins, which are hypothesized to have antineoplastic properties, on biochemical failure (nadir plus 2 ng/mL) after prostate radiation therapy (RT). Methods and Materials: Between 1989 and 2006, 2051 men with clinically localized prostate cancer received definitive RT alone (median dose, 76 Gy). The rates of aspirin use and statin use (defined as any use at the time of RT or during follow-up) were 36% and 34%, respectively. The primary endpoint of the study was an interval to biochemical failure (IBF) of less than 18 months, which hasmore » been shown to be the single strongest predictor of distant metastasis, prostate cancer survival, and overall survival after RT. Patient demographic characteristics and tumor staging factors were assessed with regard to associations with the endpoint. Univariate analysis was performed with the {chi}{sup 2} test for categorical variables and the Wilcoxon test for continuous variables. Multivariable analysis was performed with a multiple logistic regression. Results: The median follow-up was 75 months. Univariate analysis showed that an IBF of less than 18 months was associated with aspirin nonuse (P<.0001), statin nonuse (P<.0001), anticoagulant nonuse (P=.0006), cardiovascular disease (P=.0008), and prostate-specific antigen (continuous) (P=.008) but not with Gleason score, age, RT dose, or T stage. On multivariate analysis, only aspirin nonuse (P=.0012; odds ratio, 2.052 [95% confidence interval, 1.328-3.172]) and statin nonuse (P=.0002; odds ratio, 2.465 [95% confidence interval, 1.529-3.974]) were associated with an IBF of less than 18 months. Conclusions: In patients who received RT for prostate cancer, aspirin or statin nonuse was associated with early biochemical failure, a harbinger of distant metastasis and death. Further study is needed to confirm these findings and to determine the optimal dosing and schedule, as well as the relative benefits and risks, of both therapies in combination with RT.« less
Silay, M S; Spinoit, A F; Undre, S; Fiala, V; Tandogdu, Z; Garmanova, T; Guttilla, A; Sancaktutar, A A; Haid, B; Waldert, M; Goyal, A; Serefoglu, E C; Baldassarre, E; Manzoni, G; Radford, A; Subramaniam, R; Cherian, A; Hoebeke, P; Jacobs, M; Rocco, B; Yuriy, R; Zattoni, Fabio; Kocvara, R; Koh, C J
2016-08-01
Minimally invasive pyeloplasty (MIP) for ureteropelvic junction (UPJ) obstruction in children has gained popularity over the past decade as an alternative to open surgery. The present study aimed to identify the factors affecting complication rates of MIP in children, and to compare the outcomes of laparoscopic (LP) and robotic-assisted laparoscopic pyeloplasty (RALP). The perioperative data of 783 pediatric patients (<18 years old) from 15 academic centers who underwent either LP or RALP with an Anderson Hynes dismembered pyeloplasty technique were retrospectively evaluated. Redo cases and patients with anatomic renal abnormalities were excluded. Demographics and operative data, including procedural factors, were collected. Complications were classified according to the Satava and modified Clavien systems. Failure was defined as any of the following: obstructive parameters on diuretic renal scintigraphy, decline in renal function, progressive hydronephrosis, or symptom relapse. Univariate and multivariate analysis were applied to identify factors affecting the complication rates. All parameters were compared between LP and RALP. A total of 575 children met the inclusion criteria. Laparoscopy, increased operative time, prolonged hospital stay, ureteral stenting technique, and time required for stenting were factors influencing complication rates on univariate analysis. None of those factors remained significant on multivariate analysis. Mean follow-up was 12.8 ± 9.8 months for RALP and 45.2 ± 33.8 months for LP (P = 0.001). Hospital stay and time for stenting were shorter for robotic pyeloplasty (P < 0.05 for both). Success rates were similar between RALP and LP (99.5% vs 97.3%, P = 0.11). The intraoperative complication rate was comparable between RALP and LP (3.8% vs 7.4%, P = 0.06). However, the postoperative complication rate was significantly higher in the LP group (3.2% for RALP and 7.7% for LP, P = 0.02). All complications were of no greater severity than Satava Grade IIa and Clavien Grade IIIb. This was the largest multicenter series of LP and RALP in the pediatric population. Limitations of the study included the retrospective design and lack of surgical experience as a confounder. Both minimally invasive approaches that were studied were safe and highly effective in treating UPJ obstruction in children in many centers globally. However, shorter hospitalization time and lower postoperative complication rates with RALP were noted. The aims of the study were met. Copyright © 2016 Journal of Pediatric Urology Company. Published by Elsevier Ltd. All rights reserved.
Sardo, Pedro Miguel Garcez; Guedes, Jenifer Adriana Domingues; Alvarelhão, José Joaquim Marques; Machado, Paulo Alexandre Puga; Melo, Elsa Maria Oliveira Pinheiro
2018-05-01
To study the influence of Braden subscales scores (at the first pressure ulcer risk assessment) on pressure ulcer incidence using a univariate and a multivariate time to event analysis. Retrospective cohort analysis of electronic health record database from adult patients admitted without pressure ulcer(s) to medical and surgical wards of a Portuguese hospital during 2012. The hazard ratio of developing a pressure ulcer during the length of inpatient stay was calculated by univariate Cox regression for each variable of interest and by multivariate Cox regression for the Braden subscales that were statistically significant. This study included a sample of 6552 participants. During the length of stay, 153 participants developed (at least) one pressure ulcer, giving a pressure ulcer incidence of 2.3%. The univariate time to event analysis showed that all Braden subscales, except "nutrition", were associated with the development of pressure ulcer. By multivariate analysis the scores for "mobility" and "activity" were independently predictive of the development of pressure ulcer(s) for all participants. (Im)"mobility" (the lack of ability to change and control body position) and (in)"activity" (the limited degree of physical activity) were the major risk factors assessed by Braden Scale for pressure ulcer development during the length of inpatient stay. Thus, the greatest efforts in managing pressure ulcer risk should be on "mobility" and "activity", independently of the total Braden Scale score. Copyright © 2018 Tissue Viability Society. Published by Elsevier Ltd. All rights reserved.
Wang, X; Jiao, Y; Tang, T; Wang, H; Lu, Z
2013-12-19
Intrinsic connectivity networks (ICNs) are composed of spatial components and time courses. The spatial components of ICNs were discovered with moderate-to-high reliability. So far as we know, few studies focused on the reliability of the temporal patterns for ICNs based their individual time courses. The goals of this study were twofold: to investigate the test-retest reliability of temporal patterns for ICNs, and to analyze these informative univariate metrics. Additionally, a correlation analysis was performed to enhance interpretability. Our study included three datasets: (a) short- and long-term scans, (b) multi-band echo-planar imaging (mEPI), and (c) eyes open or closed. Using dual regression, we obtained the time courses of ICNs for each subject. To produce temporal patterns for ICNs, we applied two categories of univariate metrics: network-wise complexity and network-wise low-frequency oscillation. Furthermore, we validated the test-retest reliability for each metric. The network-wise temporal patterns for most ICNs (especially for default mode network, DMN) exhibited moderate-to-high reliability and reproducibility under different scan conditions. Network-wise complexity for DMN exhibited fair reliability (ICC<0.5) based on eyes-closed sessions. Specially, our results supported that mEPI could be a useful method with high reliability and reproducibility. In addition, these temporal patterns were with physiological meanings, and certain temporal patterns were correlated to the node strength of the corresponding ICN. Overall, network-wise temporal patterns of ICNs were reliable and informative and could be complementary to spatial patterns of ICNs for further study. Copyright © 2013 IBRO. Published by Elsevier Ltd. All rights reserved.
Infusion phlebitis in post-operative patients: when and why.
Monreal, M; Oller, B; Rodriguez, N; Vega, J; Torres, T; Valero, P; Mach, G; Ruiz, A E; Roca, J
1999-01-01
The most common complication of intravenous therapy is infusion phlebitis. This study was done to prospectively assess its frequency in a series of consecutive patients who will undergo surgery, and to identify which variables may predict an increased risk for phlebitis. 400 consecutive patients who will undergo surgery in a general surgery department were included. Only the first catheter, inserted the day before surgery, was taken into account. Eighteen variables (from the infusion, the catheter and from the patient) were prospectively evaluated for their contribution to the occurrence of phlebitis. 60/400 patients (15%) developed phlebitis, and most of them needed insertion of a further catheter. The univariate analysis showed that patients who developed phlebitis were older, and their pre-operative levels of both blood haemoglobin and neutrophil cound were significantly higher than those in patients who did not develop phlebitis. However, the multivariate analysis only confirmed the association with blood haemoglobin levels: the risk of phlebitis sharply increased in the patients with the highest haemoglobin levels. As to the influence of time on phlebitis development, there was a significant decrease in the day-specific risk, from the 5th day on. In our series, blood haemoglobin levels were found to be the only variable associated to a higher risk of phlebitis. Besides, in contrast with the recommendations by the Centers for Disease Control, no significant increase in the day-specific risk of phlebitis was found. Thus, a guideline to select the type of catheter to be inserted in an individual patient is suggested. Copyright 2000 S. Karger AG, Basel
Formisano, Elia; De Martino, Federico; Valente, Giancarlo
2008-09-01
Machine learning and pattern recognition techniques are being increasingly employed in functional magnetic resonance imaging (fMRI) data analysis. By taking into account the full spatial pattern of brain activity measured simultaneously at many locations, these methods allow detecting subtle, non-strictly localized effects that may remain invisible to the conventional analysis with univariate statistical methods. In typical fMRI applications, pattern recognition algorithms "learn" a functional relationship between brain response patterns and a perceptual, cognitive or behavioral state of a subject expressed in terms of a label, which may assume discrete (classification) or continuous (regression) values. This learned functional relationship is then used to predict the unseen labels from a new data set ("brain reading"). In this article, we describe the mathematical foundations of machine learning applications in fMRI. We focus on two methods, support vector machines and relevance vector machines, which are respectively suited for the classification and regression of fMRI patterns. Furthermore, by means of several examples and applications, we illustrate and discuss the methodological challenges of using machine learning algorithms in the context of fMRI data analysis.
Compound summer temperature and precipitation extremes over central Europe
NASA Astrophysics Data System (ADS)
Sedlmeier, Katrin; Feldmann, H.; Schädler, G.
2018-02-01
Reliable knowledge of the near-future climate change signal of extremes is important for adaptation and mitigation strategies. Especially compound extremes, like heat and drought occurring simultaneously, may have a greater impact on society than their univariate counterparts and have recently become an active field of study. In this paper, we use a 12-member ensemble of high-resolution (7 km) regional climate simulations with the regional climate model COSMO-CLM over central Europe to analyze the climate change signal and its uncertainty for compound heat and drought extremes in summer by two different measures: one describing absolute (i.e., number of exceedances of absolute thresholds like hot days), the other relative (i.e., number of exceedances of time series intrinsic thresholds) compound extreme events. Changes are assessed between a reference period (1971-2000) and a projection period (2021-2050). Our findings show an increase in the number of absolute compound events for the whole investigation area. The change signal of relative extremes is more region-dependent, but there is a strong signal change in the southern and eastern parts of Germany and the neighboring countries. Especially the Czech Republic shows strong change in absolute and relative extreme events.
NASA Astrophysics Data System (ADS)
Cawiding, Olive R.; Natividad, Gina May R.; Bato, Crisostomo V.; Addawe, Rizavel C.
2017-11-01
The prevalence of typhoid fever in developing countries such as the Philippines calls for a need for accurate forecasting of the disease. This will be of great assistance in strategic disease prevention. This paper presents a development of useful models that predict the behavior of typhoid fever incidence based on the monthly incidence in the provinces of the Cordillera Administrative Region from 2010 to 2015 using univariate time series analysis. The data used was obtained from the Cordillera Office of the Department of Health (DOH-CAR). Seasonal autoregressive moving average (SARIMA) models were used to incorporate the seasonality of the data. A comparison of the results of the obtained models revealed that the SARIMA (1,1,7)(0,0,1)12 with a fixed coefficient at the seventh lag produces the smallest root mean square error (RMSE), mean absolute error (MAE), Akaike Information Criterion (AIC), and Bayesian Information Criterion (BIC). The model suggested that for the year 2016, the number of cases would increase from the months of July to September and have a drop in December. This was then validated using the data collected from January 2016 to December 2016.
Iturriaga, H; Hirsch, S; Bunout, D; Díaz, M; Kelly, M; Silva, G; de la Maza, M P; Petermann, M; Ugarte, G
1993-04-01
Looking for a noninvasive method to predict liver histologic alterations in alcoholic patients without clinical signs of liver failure, we studied 187 chronic alcoholics recently abstinent, divided in 2 series. In the model series (n = 94) several clinical variables and results of common laboratory tests were confronted to the findings of liver biopsies. These were classified in 3 groups: 1. Normal liver; 2. Moderate alterations; 3. Marked alterations, including alcoholic hepatitis and cirrhosis. Multivariate methods used were logistic regression analysis and a classification and regression tree (CART). Both methods entered gamma-glutamyltransferase (GGT), aspartate-aminotransferase (AST), weight and age as significant and independent variables. Univariate analysis with GGT and AST at different cutoffs were also performed. To predict the presence of any kind of damage (Groups 2 and 3), CART and AST > 30 IU showed the higher sensitivity, specificity and correct prediction, both in the model and validation series. For prediction of marked liver damage, a score based on logistic regression and GGT > 110 IU had the higher efficiencies. It is concluded that GGT and AST are good markers of alcoholic liver damage and that, using sample cutoffs, histologic diagnosis can be correctly predicted in 80% of recently abstinent asymptomatic alcoholics.
Relating annual increments of the endangered Blanding's turtle plastron growth to climate
Richard, Monik G; Laroque, Colin P; Herman, Thomas B
2014-01-01
This research is the first published study to report a relationship between climate variables and plastron growth increments of turtles, in this case the endangered Nova Scotia Blanding's turtle (Emydoidea blandingii). We used techniques and software common to the discipline of dendrochronology to successfully cross-date our growth increment data series, to detrend and average our series of 80 immature Blanding's turtles into one common chronology, and to seek correlations between the chronology and environmental temperature and precipitation variables. Our cross-dated chronology had a series intercorrelation of 0.441 (above 99% confidence interval), an average mean sensitivity of 0.293, and an average unfiltered autocorrelation of 0.377. Our master chronology represented increments from 1975 to 2007 (33 years), with index values ranging from a low of 0.688 in 2006 to a high of 1.303 in 1977. Univariate climate response function analysis on mean monthly air temperature and precipitation values revealed a positive correlation with the previous year's May temperature and current year's August temperature; a negative correlation with the previous year's October temperature; and no significant correlation with precipitation. These techniques for determining growth increment response to environmental variables should be applicable to other turtle species and merit further exploration. PMID:24963390
Relating annual increments of the endangered Blanding's turtle plastron growth to climate.
Richard, Monik G; Laroque, Colin P; Herman, Thomas B
2014-05-01
This research is the first published study to report a relationship between climate variables and plastron growth increments of turtles, in this case the endangered Nova Scotia Blanding's turtle (Emydoidea blandingii). We used techniques and software common to the discipline of dendrochronology to successfully cross-date our growth increment data series, to detrend and average our series of 80 immature Blanding's turtles into one common chronology, and to seek correlations between the chronology and environmental temperature and precipitation variables. Our cross-dated chronology had a series intercorrelation of 0.441 (above 99% confidence interval), an average mean sensitivity of 0.293, and an average unfiltered autocorrelation of 0.377. Our master chronology represented increments from 1975 to 2007 (33 years), with index values ranging from a low of 0.688 in 2006 to a high of 1.303 in 1977. Univariate climate response function analysis on mean monthly air temperature and precipitation values revealed a positive correlation with the previous year's May temperature and current year's August temperature; a negative correlation with the previous year's October temperature; and no significant correlation with precipitation. These techniques for determining growth increment response to environmental variables should be applicable to other turtle species and merit further exploration.
Boysen, William R; Akhavan, Ardavan; Ko, Joan; Ellison, Jonathan S; Lendvay, Thomas S; Huang, Jonathan; Garcia-Roig, Michael; Kirsch, Andrew; Koh, Chester J; Schulte, Marion; Noh, Paul; Monn, M Francesca; Whittam, Benjamin; Kawal, Trudy; Shukla, Aseem; Srinivasan, Arun; Gundeti, Mohan S
2018-02-20
Robot-assisted laparoscopic extravesical ureteral reimplantation (RALUR-EV) is a minimally invasive alternative to open surgery. We have previously reported retrospective outcomes from our study group, and herein provide an updated prospective analysis with a focus on success rate, surgical technique, and complications among surgeons who have overcome the initial learning curve. To assess the safety and efficacy of RALUR-EV in children, among experienced surgeons. We reviewed our prospective database of children undergoing RALUR-EV by pediatric urologists at eight academic centers from 2015 to 2017. Radiographic success was defined as absence of vesicoureteral reflux (VUR) on postoperative voiding cystourethrogram. Complications were graded using the Clavien scale. Univariate regression analysis was performed to assess for association among various patient and technical factors and radiographic failure. In total, 143 patients were treated with RALUR-EV for primary VUR (87 unilateral, 56 bilateral; 199 ureters). The majority of ureters (73.4%) had grade III or higher VUR preoperatively. Radiographic resolution was present in 93.8% of ureters, as shown in the summary table. Ureteral complications occurred in five ureters (2.5%) with mean follow-up of 7.4 months (SD 4.0). Transient urinary retention occurred in four patients following bilateral procedure (7.1%) and in no patients after unilateral. On univariate analysis, there were no patient or technical factors associated with increased odds of radiographic failure. We report a radiographic success rate of 93.8% overall, and 94.1% among children with grades III-V VUR. In contemporary series, alternate management options such as endoscopic injection and open UR have reported radiographic success rates of 90% and 93.5% respectively. We were unable to identify specific patient or technical factors that influenced outcomes, although immeasurable factors such as tissue handling and intraoperative decision-making could not be assessed. Ureteral complications requiring operative intervention were rare and occurred with the same incidence reported in a large open series. Limitations include lack of long-term follow-up and absence of radiographic follow-up on a subset of patients. Radiographic resolution of VUR following RALUR is on par with contemporary open series, and the incidence of ureteral complications is low. RALUR should be considered as one of several viable options for management of VUR in children. Copyright © 2018. Published by Elsevier Ltd.
Observed increase in extreme daily rainfall in the French Mediterranean
NASA Astrophysics Data System (ADS)
Ribes, Aurélien; Thao, Soulivanh; Vautard, Robert; Dubuisson, Brigitte; Somot, Samuel; Colin, Jeanne; Planton, Serge; Soubeyroux, Jean-Michel
2018-04-01
We examine long-term trends in the historical record of extreme precipitation events occurring over the French Mediterranean area. Extreme events are considered in terms of their intensity, frequency, extent and precipitated volume. Changes in intensity are analysed via an original statistical approach where the annual maximum rainfall amounts observed at each measurement station are aggregated into a univariate time-series according to their dependence. The mean intensity increase is significant and estimated at + 22% (+ 7 to + 39% at the 90% confidence level) over the 1961-2015 period. Given the observed warming over the considered area, this increase is consistent with a rate of about one to three times that implied by the Clausius-Clapeyron relationship. Changes in frequency and other spatial features are investigated through a Generalised Linear Model. Changes in frequency for events exceeding high thresholds (about 200 mm in 1 day) are found to be significant, typically near a doubling of the frequency, but with large uncertainties in this change ratio. The area affected by severe events and the water volume precipitated during those events also exhibit significant trends, with an increase by a factor of about 4 for a 200 mm threshold, again with large uncertainties. All diagnoses consistently point toward an intensification of the most extreme events over the last decades. We argue that it is difficult to explain the diagnosed trends without invoking the human influence on climate.
Chowdhury, Nilotpal; Sapru, Shantanu
2015-01-01
Microarray analysis has revolutionized the role of genomic prognostication in breast cancer. However, most studies are single series studies, and suffer from methodological problems. We sought to use a meta-analytic approach in combining multiple publicly available datasets, while correcting for batch effects, to reach a more robust oncogenomic analysis. The aim of the present study was to find gene sets associated with distant metastasis free survival (DMFS) in systemically untreated, node-negative breast cancer patients, from publicly available genomic microarray datasets. Four microarray series (having 742 patients) were selected after a systematic search and combined. Cox regression for each gene was done for the combined dataset (univariate, as well as multivariate - adjusted for expression of Cell cycle related genes) and for the 4 major molecular subtypes. The centre and microarray batch effects were adjusted by including them as random effects variables. The Cox regression coefficients for each analysis were then ranked and subjected to a Gene Set Enrichment Analysis (GSEA). Gene sets representing protein translation were independently negatively associated with metastasis in the Luminal A and Luminal B subtypes, but positively associated with metastasis in Basal tumors. Proteinaceous extracellular matrix (ECM) gene set expression was positively associated with metastasis, after adjustment for expression of cell cycle related genes on the combined dataset. Finally, the positive association of the proliferation-related genes with metastases was confirmed. To the best of our knowledge, the results depicting mixed prognostic significance of protein translation in breast cancer subtypes are being reported for the first time. We attribute this to our study combining multiple series and performing a more robust meta-analytic Cox regression modeling on the combined dataset, thus discovering 'hidden' associations. This methodology seems to yield new and interesting results and may be used as a tool to guide new research.
Chowdhury, Nilotpal; Sapru, Shantanu
2015-01-01
Introduction Microarray analysis has revolutionized the role of genomic prognostication in breast cancer. However, most studies are single series studies, and suffer from methodological problems. We sought to use a meta-analytic approach in combining multiple publicly available datasets, while correcting for batch effects, to reach a more robust oncogenomic analysis. Aim The aim of the present study was to find gene sets associated with distant metastasis free survival (DMFS) in systemically untreated, node-negative breast cancer patients, from publicly available genomic microarray datasets. Methods Four microarray series (having 742 patients) were selected after a systematic search and combined. Cox regression for each gene was done for the combined dataset (univariate, as well as multivariate – adjusted for expression of Cell cycle related genes) and for the 4 major molecular subtypes. The centre and microarray batch effects were adjusted by including them as random effects variables. The Cox regression coefficients for each analysis were then ranked and subjected to a Gene Set Enrichment Analysis (GSEA). Results Gene sets representing protein translation were independently negatively associated with metastasis in the Luminal A and Luminal B subtypes, but positively associated with metastasis in Basal tumors. Proteinaceous extracellular matrix (ECM) gene set expression was positively associated with metastasis, after adjustment for expression of cell cycle related genes on the combined dataset. Finally, the positive association of the proliferation-related genes with metastases was confirmed. Conclusion To the best of our knowledge, the results depicting mixed prognostic significance of protein translation in breast cancer subtypes are being reported for the first time. We attribute this to our study combining multiple series and performing a more robust meta-analytic Cox regression modeling on the combined dataset, thus discovering 'hidden' associations. This methodology seems to yield new and interesting results and may be used as a tool to guide new research. PMID:26080057
Kwei, Kimberly T; Liang, John; Wilson, Natalie; Tuhrim, Stanley; Dhamoon, Mandip
2018-05-01
Optimizing the time it takes to get a potential stroke patient to imaging is essential in a rapid stroke response. At our hospital, door-to-imaging time is comprised of 2 time periods: the time before a stroke is recognized, followed by the period after the stroke code is called during which the stroke team assesses and brings the patient to the computed tomography scanner. To control for delays due to triage, we isolated the time period after a potential stroke has been recognized, as few studies have examined the biases of stroke code responders. This "code-to-imaging time" (CIT) encompassed the time from stroke code activation to initial imaging, and we hypothesized that perception of stroke severity would affect how quickly stroke code responders act. In consecutively admitted ischemic stroke patients at The Mount Sinai Hospital emergency department, we tested associations between National Institutes of Health Stroke Scale scores (NIHSS), continuously and at different cutoffs, and CIT using spline regression, t tests for univariate analysis, and multivariable linear regression adjusting for age, sex, and race/ethnicity. In our study population, mean CIT was 26 minutes, and mean presentation NIHSS was 8. In univariate and multivariate analyses comparing CIT between mild and severe strokes, stroke scale scores <4 were associated with longer response times. Milder strokes are associated with a longer CIT with a threshold effect at a NIHSS of 4.
Monnier, Yan; Broome, Martin; Betz, Michael; Bouferrache, Kahina; Ozsahin, Mahmut; Jaques, Bertrand
2011-05-01
Mandibular osteoradionecrosis (ORN) is a serious complication of radiotherapy (RT) in head and neck cancer patients. The aim of this study was to analyze the incidence of and risk factors for mandibular ORN in squamous cell carcinoma (SCC) of the oral cavity and oropharynx. Case series with chart review. University tertiary care center for head and neck oncology. Seventy-three patients treated for stage I to IV SCC of the oral cavity and oropharynx between 2000 and 2007, with a minimum follow-up of 2 years, were included in the study. Treatment modalities included both RT with curative intent and adjuvant RT following tumor surgery. The log-rank test and Cox model were used for univariate and multivariate analyses. The incidence of mandibular ORN was 40% at 5 years. Using univariate analysis, the following risk factors were identified: oral cavity tumors (P < .01), bone invasion (P < .02), any surgery prior to RT (P < .04), and bone surgery (P < .0001). By multivariate analysis, mandibular surgery proved to be the most important risk factor and the only one reaching statistical significance (P < .0002). Mandibular ORN is a frequent long-term complication of RT for oral cavity and oropharynx cancers. Mandibular surgery before irradiation is the only independent risk factor. These aspects must be considered when planning treatment for these tumors.
[A SAS marco program for batch processing of univariate Cox regression analysis for great database].
Yang, Rendong; Xiong, Jie; Peng, Yangqin; Peng, Xiaoning; Zeng, Xiaomin
2015-02-01
To realize batch processing of univariate Cox regression analysis for great database by SAS marco program. We wrote a SAS macro program, which can filter, integrate, and export P values to Excel by SAS9.2. The program was used for screening survival correlated RNA molecules of ovarian cancer. A SAS marco program could finish the batch processing of univariate Cox regression analysis, the selection and export of the results. The SAS macro program has potential applications in reducing the workload of statistical analysis and providing a basis for batch processing of univariate Cox regression analysis.
Application of spatial methods to identify areas with lime requirement in eastern Croatia
NASA Astrophysics Data System (ADS)
Bogunović, Igor; Kisic, Ivica; Mesic, Milan; Zgorelec, Zeljka; Percin, Aleksandra; Pereira, Paulo
2016-04-01
With more than 50% of acid soils in all agricultural land in Croatia, soil acidity is recognized as a big problem. Low soil pH leads to a series of negative phenomena in plant production and therefore as a compulsory measure for reclamation of acid soils is liming, recommended on the base of soil analysis. The need for liming is often erroneously determined only on the basis of the soil pH, because the determination of cation exchange capacity, the hydrolytic acidity and base saturation is a major cost to producers. Therefore, in Croatia, as well as some other countries, the amount of liming material needed to ameliorate acid soils is calculated by considering their hydrolytic acidity. For this research, several interpolation methods were tested to identify the best spatial predictor of hidrolitic acidity. The purpose of this study was to: test several interpolation methods to identify the best spatial predictor of hidrolitic acidity; and to determine the possibility of using multivariate geostatistics in order to reduce the number of needed samples for determination the hydrolytic acidity, all with an aim that the accuracy of the spatial distribution of liming requirement is not significantly reduced. Soil pH (in KCl) and hydrolytic acidity (Y1) is determined in the 1004 samples (from 0-30 cm) randomized collected in agricultural fields near Orahovica in eastern Croatia. This study tested 14 univariate interpolation models (part of ArcGIS software package) in order to provide most accurate spatial map of hydrolytic acidity on a base of: all samples (Y1 100%), and the datasets with 15% (Y1 85%), 30% (Y1 70%) and 50% fewer samples (Y1 50%). Parallel to univariate interpolation methods, the precision of the spatial distribution of the Y1 was tested by the co-kriging method with exchangeable acidity (pH in KCl) as a covariate. The soils at studied area had an average pH (KCl) 4,81, while the average Y1 10,52 cmol+ kg-1. These data suggest that liming is necessary agrotechnical measure for soil conditioning. The results show that ordinary kriging was most accurate univariate interpolation method with smallest error (RMSE) in all four data sets, while the least precise showed Radial Basis Functions (Thin Plate Spline and Inverse Multiquadratic). Furthermore, it is noticeable a trend of increasing errors (RMSE) with a reduced number of samples tested on the most accurate univariate interpolation model: 3,096 (Y1 100%), 3,258 (Y1 85%), 3,317 (Y1 70%), 3,546 (Y1 50%). The best-fit semivariograms show a strong spatial dependence in Y1 100% (Nugget/Sill 20.19) and Y1 85% (Nugget/Sill 23.83), while a further reduction of the number of samples resulted with moderate spatial dependence (Y1 70% -35,85% and Y1 50% - 32,01). Co-kriging method resulted in a reduction in RMSE compared with univariate interpolation methods for each data set with: 2,054, 1,731 and 1,734 for Y1 85%, Y1 70%, Y1 50%, respectively. The results show the possibility for reducing sampling costs by using co-kriging method which is useful from the practical viewpoint. Reduced number of samples by half for determination of hydrolytic acidity in the interaction with the soil pH provides a higher precision for variable liming compared to the univariate interpolation methods of the entire set of data. These data provide new opportunities to reduce costs in the practical plant production in Croatia.
ERIC Educational Resources Information Center
Barton, Mitch; Yeatts, Paul E.; Henson, Robin K.; Martin, Scott B.
2016-01-01
There has been a recent call to improve data reporting in kinesiology journals, including the appropriate use of univariate and multivariate analysis techniques. For example, a multivariate analysis of variance (MANOVA) with univariate post hocs and a Bonferroni correction is frequently used to investigate group differences on multiple dependent…
ERIC Educational Resources Information Center
Sun, Anji; Valiga, Michael J.
In this study, the reliability of the American College Testing (ACT) Program's "Survey of Academic Advising" (SAA) was examined using both univariate and multivariate generalizability theory approaches. The primary purpose of the study was to compare the results of three generalizability theory models (a random univariate model, a mixed…
A Comparison of the Achievement of Statistics Students Enrolled in Online and Face-to-Face Settings
ERIC Educational Resources Information Center
Christmann, Edwin P.
2017-01-01
This study compared the achievement of male and female students who were enrolled in an online univariate statistics course to students enrolled in a traditional face-to-face univariate statistics course. The subjects, 47 graduate students enrolled in univariate statistics classes at a public, comprehensive university, were randomly assigned to…
Analysis of vector wind change with respect to time for Cape Kennedy, Florida
NASA Technical Reports Server (NTRS)
Adelfang, S. I.
1978-01-01
Multivariate analysis was used to determine the joint distribution of the four variables represented by the components of the wind vector at an initial time and after a specified elapsed time is hypothesized to be quadravariate normal; the fourteen statistics of this distribution, calculated from 15 years of twice-daily rawinsonde data are presented by monthly reference periods for each month from 0 to 27 km. The hypotheses that the wind component changes with respect to time is univariate normal, that the joint distribution of wind component change with respect to time is univariate normal, that the joint distribution of wind component changes is bivariate normal, and that the modulus of vector wind change is Rayleigh are tested by comparison with observed distributions. Statistics of the conditional bivariate normal distributions of vector wind at a future time given the vector wind at an initial time are derived. Wind changes over time periods from 1 to 5 hours, calculated from Jimsphere data, are presented. Extension of the theoretical prediction (based on rawinsonde data) of wind component change standard deviation to time periods of 1 to 5 hours falls (with a few exceptions) within the 95 percentile confidence band of the population estimate obtained from the Jimsphere sample data. The joint distributions of wind change components, conditional wind components, and 1 km vector wind shear change components are illustrated by probability ellipses at the 95 percentile level.
Campbell, Tonya J; Decloe, Melissa; Gill, Suzanne; Ho, Grace; McCready, Janine; Powis, Jeff
2017-01-01
The success of antimicrobial stewardship is dependent on how often it is completed and which antimicrobials are targeted. We evaluated the impact of an antimicrobial stewardship program (ASP) in three non-ICU settings where all systemic antibiotics, regardless of spectrum, were targeted on the first weekday after initiation. Prospective audit and feedback (PAAF) was initiated on the surgical, respiratory, and medical wards of a community hospital on July 1, 2010, October 1, 2010, and April 1, 2012, respectively. We evaluated rates of total antibiotic use, measured in days on therapy (DOTs), among all patients admitted to the wards before and after PAAF initiation using an interrupted time series analysis. Changes in antibiotic costs, rates of C. difficile infection (CDI), mortality, readmission, and length of stay were evaluated using univariate analyses. Time series modelling demonstrated that total antibiotic use decreased (± standard error) by 100 ± 51 DOTs/1,000 patient-days on the surgical wards (p = 0.049), 100 ± 46 DOTs/1,000 patient-days on the respiratory ward (p = 0.029), and 91 ± 33 DOTs/1,000 patient-days on the medical wards (p = 0.006) immediately following PAAF initiation. Reductions in antibiotic use were sustained up to 50 months after intervention initiation, and were accompanied by decreases in antibiotic costs. There were no significant changes to patient outcomes on the surgical and respiratory wards following intervention initiation. On the medical wards, however, readmission increased from 4.6 to 5.6 per 1,000 patient-days (p = 0.043), while mortality decreased from 7.4 to 5.0 per 1,000 patient-days (p = 0.001). CDI rates showed a non-significant declining trend after PAAF initiation. ASPs can lead to cost-effective, sustained reductions in total antibiotic use when interventions are conducted early in the course of therapy and target all antibiotics. Shifting to such a model may help strengthen the effectiveness of ASPs in non-ICU settings.
NASA Astrophysics Data System (ADS)
Serinaldi, Francesco; Kilsby, Chris G.
2013-06-01
The information contained in hyetographs and hydrographs is often synthesized by using key properties such as the peak or maximum value Xp, volume V, duration D, and average intensity I. These variables play a fundamental role in hydrologic engineering as they are used, for instance, to define design hyetographs and hydrographs as well as to model and simulate the rainfall and streamflow processes. Given their inherent variability and the empirical evidence of the presence of a significant degree of association, such quantities have been studied as correlated random variables suitable to be modeled by multivariate joint distribution functions. The advent of copulas in geosciences simplified the inference procedures allowing for splitting the analysis of the marginal distributions and the study of the so-called dependence structure or copula. However, the attention paid to the modeling task has overlooked a more thorough study of the true nature and origin of the relationships that link Xp,V,D, and I. In this study, we apply a set of ad hoc bootstrap algorithms to investigate these aspects by analyzing the hyetographs and hydrographs extracted from 282 daily rainfall series from central eastern Europe, three 5 min rainfall series from central Italy, 80 daily streamflow series from the continental United States, and two sets of 200 simulated universal multifractal time series. Our results show that all the pairwise dependence structures between Xp,V,D, and I exhibit some key properties that can be reproduced by simple bootstrap algorithms that rely on a standard univariate resampling without resort to multivariate techniques. Therefore, the strong similarities between the observed dependence structures and the agreement between the observed and bootstrap samples suggest the existence of a numerical generating mechanism based on the superposition of the effects of sampling data at finite time steps and the process of summing realizations of independent random variables over random durations. We also show that the pairwise dependence structures are weakly dependent on the internal patterns of the hyetographs and hydrographs, meaning that the temporal evolution of the rainfall and runoff events marginally influences the mutual relationships of Xp,V,D, and I. Finally, our findings point out that subtle and often overlooked deterministic relationships between the properties of the event hyetographs and hydrographs exist. Confusing these relationships with genuine stochastic relationships can lead to an incorrect application of multivariate distributions and copulas and to misleading results.
Nontraditional Student Withdrawal from Undergraduate Accounting Programmes: A Holistic Perspective
ERIC Educational Resources Information Center
Fortin, Anne; Sauvé, Louise; Viger, Chantal; Landry, France
2016-01-01
A collaborative project of several Quebec universities, this study investigates nontraditional student withdrawal from undergraduate accounting programmes. A nontraditional student is older than 24, or is a commuter or a part-time student, or combines some of these characteristics. Univariate and multivariate analyses of student dropout factors…
NASA Astrophysics Data System (ADS)
Cannon, Alex
2017-04-01
Estimating historical trends in short-duration rainfall extremes at regional and local scales is challenging due to low signal-to-noise ratios and the limited availability of homogenized observational data. In addition to being of scientific interest, trends in rainfall extremes are of practical importance, as their presence calls into question the stationarity assumptions that underpin traditional engineering and infrastructure design practice. Even with these fundamental challenges, increasingly complex questions are being asked about time series of extremes. For instance, users may not only want to know whether or not rainfall extremes have changed over time, they may also want information on the modulation of trends by large-scale climate modes or on the nonstationarity of trends (e.g., identifying hiatus periods or periods of accelerating positive trends). Efforts have thus been devoted to the development and application of more robust and powerful statistical estimators for regional and local scale trends. While a standard nonparametric method like the regional Mann-Kendall test, which tests for the presence of monotonic trends (i.e., strictly non-decreasing or non-increasing changes), makes fewer assumptions than parametric methods and pools information from stations within a region, it is not designed to visualize detected trends, include information from covariates, or answer questions about the rate of change in trends. As a remedy, monotone quantile regression (MQR) has been developed as a nonparametric alternative that can be used to estimate a common monotonic trend in extremes at multiple stations. Quantile regression makes efficient use of data by directly estimating conditional quantiles based on information from all rainfall data in a region, i.e., without having to precompute the sample quantiles. The MQR method is also flexible and can be used to visualize and analyze the nonlinearity of the detected trend. However, it is fundamentally a univariate technique, and cannot incorporate information from additional covariates, for example ENSO state or physiographic controls on extreme rainfall within a region. Here, the univariate MQR model is extended to allow the use of multiple covariates. Multivariate monotone quantile regression (MMQR) is based on a single hidden-layer feedforward network with the quantile regression error function and partial monotonicity constraints. The MMQR model is demonstrated via Monte Carlo simulations and the estimation and visualization of regional trends in moderate rainfall extremes based on homogenized sub-daily precipitation data at stations in Canada.
El-Husseini, Amr A; Foda, Mohamed A; Shokeir, Ahmed A; Shehab El-Din, Ahmed B; Sobh, Mohamed A; Ghoneim, Mohamed A
2005-12-01
To study the independent determinants of graft survival among pediatric and adolescent live donor kidney transplant recipients. Between March 1976 and March 2004, 1600 live donor kidney transplants were carried out in our center. Of them 284 were 20 yr old or younger (mean age 13.1 yr, ranging from 5 to 20 yr). Evaluation of the possible variables that may affect graft survival were carried out using univariate and multivariate analyses. Studied factors included age, gender, relation between donor and recipient, original kidney disease, ABO blood group, pretransplant blood transfusion, human leukocyte antigen (HLA) matching, pretransplant dialysis, height standard deviation score (SDS), pretransplant hypertension, cold ischemia time, number of renal arteries, ureteral anastomosis, time to diuresis, time of transplantation, occurrence of acute tubular necrosis (ATN), primary and secondary immunosuppression, total dose of steroids in the first 3 months, development of acute rejection and post-transplant hypertension. Using univariate analysis, the significant predictors for graft survival were HLA matching, type of primary urinary recontinuity, time to diuresis, ATN, acute rejection and post-transplant hypertension. The multivariate analysis restricted the significance to acute rejection and post-transplant hypertension. The independent determinants of graft survival in live-donor pediatric and adolescent renal transplant recipients are acute rejection and post-transplant hypertension.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gunn, Andrew J., E-mail: agunn@uabmc.edu; Sheth, Rahul A.; Luber, Brandon
2017-01-15
PurposeThe purpse of this study was to evaluate the ability of various radiologic response criteria to predict patient outcomes after trans-arterial chemo-embolization with drug-eluting beads (DEB-TACE) in patients with advanced-stage (BCLC C) hepatocellular carcinoma (HCC).Materials and methodsHospital records from 2005 to 2011 were retrospectively reviewed. Non-infiltrative lesions were measured at baseline and on follow-up scans after DEB-TACE according to various common radiologic response criteria, including guidelines of the World Health Organization (WHO), Response Evaluation Criteria in Solid Tumors (RECIST), the European Association for the Study of the Liver (EASL), and modified RECIST (mRECIST). Statistical analysis was performed to see which,more » if any, of the response criteria could be used as a predictor of overall survival (OS) or time-to-progression (TTP).Results75 patients met inclusion criteria. Median OS and TTP were 22.6 months (95 % CI 11.6–24.8) and 9.8 months (95 % CI 7.1–21.6), respectively. Univariate and multivariate Cox analyses revealed that none of the evaluated criteria had the ability to be used as a predictor for OS or TTP. Analysis of the C index in both univariate and multivariate models showed that the evaluated criteria were not accurate predictors of either OS (C-statistic range: 0.51–0.58 in the univariate model; range: 0.54–0.58 in the multivariate model) or TTP (C-statistic range: 0.55–0.59 in the univariate model; range: 0.57–0.61 in the multivariate model).ConclusionCurrent response criteria are not accurate predictors of OS or TTP in patients with advanced-stage HCC after DEB-TACE.« less
Gunn, Andrew J; Sheth, Rahul A; Luber, Brandon; Huynh, Minh-Huy; Rachamreddy, Niranjan R; Kalva, Sanjeeva P
2017-01-01
The purpse of this study was to evaluate the ability of various radiologic response criteria to predict patient outcomes after trans-arterial chemo-embolization with drug-eluting beads (DEB-TACE) in patients with advanced-stage (BCLC C) hepatocellular carcinoma (HCC). Hospital records from 2005 to 2011 were retrospectively reviewed. Non-infiltrative lesions were measured at baseline and on follow-up scans after DEB-TACE according to various common radiologic response criteria, including guidelines of the World Health Organization (WHO), Response Evaluation Criteria in Solid Tumors (RECIST), the European Association for the Study of the Liver (EASL), and modified RECIST (mRECIST). Statistical analysis was performed to see which, if any, of the response criteria could be used as a predictor of overall survival (OS) or time-to-progression (TTP). 75 patients met inclusion criteria. Median OS and TTP were 22.6 months (95 % CI 11.6-24.8) and 9.8 months (95 % CI 7.1-21.6), respectively. Univariate and multivariate Cox analyses revealed that none of the evaluated criteria had the ability to be used as a predictor for OS or TTP. Analysis of the C index in both univariate and multivariate models showed that the evaluated criteria were not accurate predictors of either OS (C-statistic range: 0.51-0.58 in the univariate model; range: 0.54-0.58 in the multivariate model) or TTP (C-statistic range: 0.55-0.59 in the univariate model; range: 0.57-0.61 in the multivariate model). Current response criteria are not accurate predictors of OS or TTP in patients with advanced-stage HCC after DEB-TACE.
Perinatal stroke and the risk of developing childhood epilepsy
Golomb, Meredith R.; Garg, Bhuwan P.; Carvalho, Karen S.; Johnson, Cynthia S.; Williams, Linda S.
2008-01-01
Objectives To describe the prevalence of epilepsy after 6 months-of-age in children with perinatal stroke and examine whether perinatal data predict epilepsy onset and resolution. Study design A retrospective review of 64 children with perinatal stroke. In children with at least 6 months of follow-up data, Kaplan-Meier curves, univariate log-rank tests, and Cox proportional hazards models were used to examine predictors of time to development of seizures, and time to resolution of seizures in children with epilepsy. The association of risk factors with the presence of epilepsy at any time after 6 months-of-age was examined using Fisher’s exact test. Results Forty-one of the 61 children with at least 6 months of follow-up data (67%) had epilepsy between 6 months-of-age and last follow-up, but in 13 of 41 seizures eventually resolved and anticonvulsants were discontinued. Infarct on prenatal ultrasound (p=0.0065) and family history of epilepsy (p=0.0093) were significantly associated with time to development of seizures after 6 months-of-age in the univariate analysis. No assessed variables were associated with time to resolution of epilepsy or with the presence of epilepsy after 6 months-of-age. Conclusions Childhood epilepsy is frequent after perinatal stroke. Evidence of infarction on prenatal ultrasound and a family history of epilepsy predict earlier onset of active seizures. PMID:17889079
Factors that impact turnaround time of surgical pathology specimens in an academic institution.
Patel, Samip; Smith, Jennifer B; Kurbatova, Ekaterina; Guarner, Jeannette
2012-09-01
Turnaround time of laboratory results is important for customer satisfaction. The College of American Pathologists' checklist requires an analytic turnaround time of 2 days or less for most routine cases and lets every hospital define what a routine specimen is. The objective of this study was to analyze which factors impact turnaround time of nonbiopsy surgical pathology specimens. We calculated the turnaround time from receipt to verification of results (adjusted for weekends and holidays) for all nonbiopsy surgical specimens during a 2-week period. Factors studied included tissue type, number of slides per case, decalcification, immunohistochemistry, consultations with other pathologists, and diagnosis. Univariate and multivariate analyses were performed. A total of 713 specimens were analyzed, 551 (77%) were verified within 2 days and 162 (23%) in 3 days or more. Lung, gastrointestinal, breast, and genitourinary specimens showed the highest percentage of cases being signed out in over 3 days. Diagnosis of malignancy (including staging of the neoplasia), consultation with other pathologists, having had a frozen section, and use of immunohistochemical stains were significantly associated with increased turnaround time in univariate analysis. Decalcification was not associated with increased turnaround time. In multivariate analysis, consultation with other pathologists, use of immunohistochemistry, diagnosis of malignancy, and the number of slides studied continued to be significantly associated with prolonged turnaround time. Our findings suggest that diagnosis of malignancy is central to significantly prolonging the turnaround time for surgical pathology specimens, thus institutions that serve cancer centers will have longer turnaround time than those that do not. Copyright © 2012 Elsevier Inc. All rights reserved.
Bayesian inference of Calibration curves: application to archaeomagnetism
NASA Astrophysics Data System (ADS)
Lanos, P.
2003-04-01
The range of errors that occur at different stages of the archaeomagnetic calibration process are modelled using a Bayesian hierarchical model. The archaeomagnetic data obtained from archaeological structures such as hearths, kilns or sets of bricks and tiles, exhibit considerable experimental errors and are typically more or less well dated by archaeological context, history or chronometric methods (14C, TL, dendrochronology, etc.). They can also be associated with stratigraphic observations which provide prior relative chronological information. The modelling we describe in this paper allows all these observations, on materials from a given period, to be linked together, and the use of penalized maximum likelihood for smoothing univariate, spherical or three-dimensional time series data allows representation of the secular variation of the geomagnetic field over time. The smooth curve we obtain (which takes the form of a penalized natural cubic spline) provides an adaptation to the effects of variability in the density of reference points over time. Since our model takes account of all the known errors in the archaeomagnetic calibration process, we are able to obtain a functional highest-posterior-density envelope on the new curve. With this new posterior estimate of the curve available to us, the Bayesian statistical framework then allows us to estimate the calendar dates of undated archaeological features (such as kilns) based on one, two or three geomagnetic parameters (inclination, declination and/or intensity). Date estimates are presented in much the same way as those that arise from radiocarbon dating. In order to illustrate the model and inference methods used, we will present results based on German archaeomagnetic data recently published by a German team.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bilbao, Cristina, E-mail: cbilbao@dbbf.ulpgc.e; Department of Radiation Oncology, Hospital Universitario de Gran Canaria Dr. Negrin, Las Palmas de Gran Canaria, Canary Islands; Lara, Pedro Carlos
Purpose: To elucidate whether microsatellite instability (MSI) predicts clinical outcome in radiation-treated endometrioid endometrial cancer (EEC). Methods and Materials: A consecutive series of 93 patients with EEC treated with extrafascial hysterectomy and postoperative radiotherapy was studied. The median clinical follow-up of patients was 138 months, with a maximum of 232 months. Five quasimonomorphic mononucleotide markers (BAT-25, BAT-26, NR21, NR24, and NR27) were used for MSI classification. Results: Twenty-five patients (22%) were classified as MSI. Both in the whole series and in early stages (I and II), univariate analysis showed a significant association between MSI and poorer 10-year local disease-free survival,more » disease-free survival, and cancer-specific survival. In multivariate analysis, MSI was excluded from the final regression model in the whole series, but in early stages MSI provided additional significant predictive information independent of traditional prognostic and predictive factors (age, stage, grade, and vascular invasion) for disease-free survival (hazard ratio [HR] 3.25, 95% confidence interval [CI] 1.01-10.49; p = 0.048) and cancer-specific survival (HR 4.20, 95% CI 1.23-14.35; p = 0.022) and was marginally significant for local disease-free survival (HR 3.54, 95% CI 0.93-13.46; p = 0.064). Conclusions: These results suggest that MSI may predict radiotherapy response in early-stage EEC.« less
USDA-ARS?s Scientific Manuscript database
Non-linear regression techniques are used widely to fit weed field emergence patterns to soil microclimatic indices using S-type functions. Artificial neural networks present interesting and alternative features for such modeling purposes. In this work, a univariate hydrothermal-time based Weibull m...
Faria, Eliney F; Caputo, Peter A; Wood, Christopher G; Karam, Jose A; Nogueras-González, Graciela M; Matin, Surena F
2014-02-01
Laparoscopic and robotic partial nephrectomy (LPN and RPN) are strongly related to influence of tumor complexity and learning curve. We analyzed a consecutive experience between RPN and LPN to discern if warm ischemia time (WIT) is in fact improved while accounting for these two confounding variables and if so by which particular aspect of WIT. This is a retrospective analysis of consecutive procedures performed by a single surgeon between 2002-2008 (LPN) and 2008-2012 (RPN). Specifically, individual steps, including tumor excision, suturing of intrarenal defect, and parenchyma, were recorded at the time of surgery. Multivariate and univariate analyzes were used to evaluate influence of learning curve, tumor complexity, and time kinetics of individual steps during WIT, to determine their influence in WIT. Additionally, we considered the effect of RPN on the learning curve. A total of 146 LPNs and 137 RPNs were included. Considering renal function, WIT, suturing time, renorrhaphy time were found statistically significant differences in favor of RPN (p < 0.05). In the univariate analysis, surgical procedure, learning curve, clinical tumor size, and RENAL nephrometry score were statistically significant predictors for WIT (p < 0.05). RPN decreased the WIT on average by approximately 7 min compared to LPN even when adjusting for learning curve, tumor complexity, and both together (p < 0.001). We found RPN was associated with a shorter WIT when controlling for influence of the learning curve and tumor complexity. The time required for tumor excision was not shortened but the time required for suturing steps was significantly shortened.
Time-Course Gene Set Analysis for Longitudinal Gene Expression Data
Hejblum, Boris P.; Skinner, Jason; Thiébaut, Rodolphe
2015-01-01
Gene set analysis methods, which consider predefined groups of genes in the analysis of genomic data, have been successfully applied for analyzing gene expression data in cross-sectional studies. The time-course gene set analysis (TcGSA) introduced here is an extension of gene set analysis to longitudinal data. The proposed method relies on random effects modeling with maximum likelihood estimates. It allows to use all available repeated measurements while dealing with unbalanced data due to missing at random (MAR) measurements. TcGSA is a hypothesis driven method that identifies a priori defined gene sets with significant expression variations over time, taking into account the potential heterogeneity of expression within gene sets. When biological conditions are compared, the method indicates if the time patterns of gene sets significantly differ according to these conditions. The interest of the method is illustrated by its application to two real life datasets: an HIV therapeutic vaccine trial (DALIA-1 trial), and data from a recent study on influenza and pneumococcal vaccines. In the DALIA-1 trial TcGSA revealed a significant change in gene expression over time within 69 gene sets during vaccination, while a standard univariate individual gene analysis corrected for multiple testing as well as a standard a Gene Set Enrichment Analysis (GSEA) for time series both failed to detect any significant pattern change over time. When applied to the second illustrative data set, TcGSA allowed the identification of 4 gene sets finally found to be linked with the influenza vaccine too although they were found to be associated to the pneumococcal vaccine only in previous analyses. In our simulation study TcGSA exhibits good statistical properties, and an increased power compared to other approaches for analyzing time-course expression patterns of gene sets. The method is made available for the community through an R package. PMID:26111374
VC-dimension of univariate decision trees.
Yildiz, Olcay Taner
2015-02-01
In this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension values of its subtrees and the number of inputs. Via a search algorithm that calculates the VC-dimension of univariate decision trees exhaustively, we show that our VC-dimension bounds are tight for simple trees. To verify that the VC-dimension bounds are useful, we also use them to get VC-generalization bounds for complexity control using structural risk minimization in decision trees, i.e., pruning. Our simulation results show that structural risk minimization pruning using the VC-dimension bounds finds trees that are more accurate as those pruned using cross validation.
Kaltenborn, Alexander; Bulling, Elke; Nitsche, Mirko; Carl, Ulrich Martin; Hermann, Robert Michael
2016-08-01
The purpose of this work was to evaluate the efficacy of low-dose radiotherapy (RT) for thumb carpometacarpal osteoarthritis (rhizarthrosis). The responses of 84 patients (n = 101 joints) were analyzed 3 months after therapy (n = 65) and at 12 months (n = 27). Patients were treated with 6 fractions of 1 Gy, two times a week, with a linear accelerator. At the end of therapy, about 70 % of patients reported a response (partial remission or complete remission), 3 months later about 60 %, and 1 year after treatment 70 %. In univariate regression analysis, higher patient age and field size greater than 6 × 4 cm were associated with response to treatment, while initial increase of pain under treatment was predictive for treatment failure. Duration of RT series (more than 18 days), gender, time of symptoms before RT, stress pain or rest pain, or prior ortheses use, injections, or surgery of the joint were not associated with treatment efficacy. In multivariate regression analysis, only field size and initial pain increase were highly correlated with treatment outcome. In conclusion, RT represents a useful treatment option for patients suffering from carpometacarpal osteoarthritis. In contrast to other benign indications, a larger field size (>6 × 4 cm) seems to be more effective than smaller fields and should be evaluated in further prospective studies.
Liver Transplantation for Fulminant Hepatic Failure
Farmer, Douglas G.; Anselmo, Dean M.; Ghobrial, R. Mark; Yersiz, Hasan; McDiarmid, Suzanne V.; Cao, Carlos; Weaver, Michael; Figueroa, Jesus; Khan, Khurram; Vargas, Jorge; Saab, Sammy; Han, Steven; Durazo, Francisco; Goldstein, Leonard; Holt, Curtis; Busuttil, Ronald W.
2003-01-01
Objective To analyze outcomes after liver transplantation (LT) in patients with fulminant hepatic failure (FHF) with emphasis on pretransplant variables that can potentially help predict posttransplant outcome. Summary Background Data FHF is a formidable clinical problem associated with a high mortality rate. While LT is the treatment of choice for irreversible FHF, few investigations have examined pretransplant variables that can potentially predict outcome after LT. Methods A retrospective review was undertaken of all patients undergoing LT for FHF at a single transplant center. The median follow-up was 41 months. Thirty-five variables were analyzed by univariate and multivariate analysis to determine their impact on patient and graft survival. Results Two hundred four patients (60% female, median age 20.2 years) required urgent LT for FHF. Before LT, the majority of patients were comatose (76%), on hemodialysis (16%), and ICU-bound. The 1- and 5-year survival rates were 73% and 67% (patient) and 63% and 57% (graft). The primary cause of patient death was sepsis, and the primary cause of graft failure was primary graft nonfunction. Univariate analysis of pre-LT variables revealed that 19 variables predicted survival. From these results, multivariate analysis determined that the serum creatinine was the single most important prognosticator of patient survival. Conclusions This study, representing one of the largest published series on LT for FHF, demonstrates a long-term survival of nearly 70% and develops a clinically applicable and readily measurable set of pretransplant factors that determine posttransplant outcome. PMID:12724633
Multivariate Analyses of Rotator Cuff Pathologies in Shoulder Disability
Henseler, Jan F.; Raz, Yotam; Nagels, Jochem; van Zwet, Erik W.; Raz, Vered; Nelissen, Rob G. H. H.
2015-01-01
Background Disability of the shoulder joint is often caused by a tear in the rotator cuff (RC) muscles. Four RC muscles coordinate shoulder movement and stability, among them the supraspinatus and infraspinatus muscle which are predominantly torn. The contribution of each RC muscle to tear pathology is not fully understood. We hypothesized that muscle atrophy and fatty infiltration, features of RC muscle degeneration, are predictive of superior humeral head translation and shoulder functional disability. Methods Shoulder features, including RC muscle surface area and fatty infiltration, superior humeral translation and RC tear size were obtained from a consecutive series of Magnetic Resonance Imaging with arthrography (MRA). We investigated patients with superior (supraspinatus, n = 39) and posterosuperior (supraspinatus and infraspinatus, n = 30) RC tears, and patients with an intact RC (n = 52) as controls. The individual or combinatorial contribution of RC measures to superior humeral translation, as a sign of RC dysfunction, was investigated with univariate or multivariate models, respectively. Results Using the univariate model the infraspinatus surface area and fatty infiltration in both the supraspinatus and infraspinatus had a significant contribution to RC dysfunction. With the multivariate model, however, the infraspinatus surface area only affected superior humeral translation (p<0.001) and discriminated between superior and posterosuperior tears. In contrast neither tear size nor fatty infiltration of the supraspinatus or infraspinatus contributed to superior humeral translation. Conclusion Our study reveals that infraspinatus atrophy has the strongest contribution to RC tear pathologies. This suggests a pivotal role for the infraspinatus in preventing shoulder disability. PMID:25710703
TATES: Efficient Multivariate Genotype-Phenotype Analysis for Genome-Wide Association Studies
van der Sluis, Sophie; Posthuma, Danielle; Dolan, Conor V.
2013-01-01
To date, the genome-wide association study (GWAS) is the primary tool to identify genetic variants that cause phenotypic variation. As GWAS analyses are generally univariate in nature, multivariate phenotypic information is usually reduced to a single composite score. This practice often results in loss of statistical power to detect causal variants. Multivariate genotype–phenotype methods do exist but attain maximal power only in special circumstances. Here, we present a new multivariate method that we refer to as TATES (Trait-based Association Test that uses Extended Simes procedure), inspired by the GATES procedure proposed by Li et al (2011). For each component of a multivariate trait, TATES combines p-values obtained in standard univariate GWAS to acquire one trait-based p-value, while correcting for correlations between components. Extensive simulations, probing a wide variety of genotype–phenotype models, show that TATES's false positive rate is correct, and that TATES's statistical power to detect causal variants explaining 0.5% of the variance can be 2.5–9 times higher than the power of univariate tests based on composite scores and 1.5–2 times higher than the power of the standard MANOVA. Unlike other multivariate methods, TATES detects both genetic variants that are common to multiple phenotypes and genetic variants that are specific to a single phenotype, i.e. TATES provides a more complete view of the genetic architecture of complex traits. As the actual causal genotype–phenotype model is usually unknown and probably phenotypically and genetically complex, TATES, available as an open source program, constitutes a powerful new multivariate strategy that allows researchers to identify novel causal variants, while the complexity of traits is no longer a limiting factor. PMID:23359524
Keogh, Brad; Culliford, David; Guerrero-Ludueña, Richard; Monks, Thomas
2018-05-24
To quantify the effect of intrahospital patient flow on emergency department (ED) performance targets and indicate if the expectations set by the National Health Service (NHS) England 5-year forward review are realistic in returning emergency services to previous performance levels. Linear regression analysis of routinely reported trust activity and performance data using a series of cross-sectional studies. NHS trusts in England submitting routine nationally reported measures to NHS England. 142 acute non-specialist trusts operating in England between 2012 and 2016. The primary outcome measures were proportion of 4-hour waiting time breaches and cancelled elective operations. Univariate and multivariate linear regression models were used to show relationships between the outcome measures and various measures of trust activity including empty day beds, empty night beds, day bed to night bed ratio, ED conversion ratio and delayed transfers of care. Univariate regression results using the outcome of 4-hour breaches showed clear relationships with empty night beds and ED conversion ratio between 2012 and 2016. The day bed to night bed ratio showed an increasing ability to explain variation in performance between 2015 and 2016. Delayed transfers of care showed little evidence of an association. Multivariate model results indicated that the ability of patient flow variables to explain 4-hour target performance had reduced between 2012 and 2016 (19% to 12%), and had increased in explaining cancelled elective operations (7% to 17%). The flow of patients through trusts is shown to influence ED performance; however, performance has become less explainable by intratrust patient flow between 2012 and 2016. Some commonly stated explanatory factors such as delayed transfers of care showed limited evidence of being related. The results indicate some of the measures proposed by NHS England to reduce pressure on EDs may not have the desired impact on returning services to previous performance levels. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Rojas-Campos, Enrique; Alcántar-Medina, Mario; Cortés-Sanabria, Laura; Martínez-Ramírez, Héctor R; Camarena, José L; Chávez, Salvador; Flores, Antonio; Nieves, Juan J; Monteón, Francisco; Gómez-Navarro, Benjamin; Cueto-Manzano, Alfonso M
2007-01-01
In Mexico, CAPD survival has been analyzed in few studies from the center of the country. However, there are concerns that such results may not represent what occurs in other province centers of our country, particularly in our geographical area. To evaluate the patient and technique survival on CAPD of a single center of the west of Mexico, and compare them with other reported series. Retrospective cohort study. Tertiary care, teaching hospital located in Guadalajara, Jalisco. Patients from our CAPD program (1999-2002) were retrospectively studied. Interventions. Clinical and biochemical variables at the start of dialysis and at the end of the follow-up were recorded and considered in the analysis of risk factors. Endpoints were patient (alive, dead or lost to follow-up) and technique status at the end of the study (June 2002). 49 patients were included. Mean patient survival (+/- SE) was 3.32 +/- 0.22 years (CI 95%: 2.9-3.8 years). Patients in the present study were younger (39 +/- 17yrs), had larger body surface area (1.72 +/- 0.22 m2), lower hematocrit (25.4 +/- 5.2%), albumin (2.6 +/- 0.6g/dL), and cholesterol (173 +/- 44 mg/dL), and higher urea (300 +/- 93 mg/dL) and creatinine (14.9 +/- 5.6 mg/ dL) than those in other Mexican series. In univariate analysis, the following variables were associated (p < 0.05) to mortality: pre-dialysis age and creatinine clearance, and serum albumin and cholesterol at the end of follow-up. In multivariate analysis, only pre-dialysis creatinine clearance (RR 0.66, p = 0.03) and age (RR 1.08, p = 0.005) significantly predicted mortality. Mean technique survival was 2.83 +/- 0.24 years (CI 95%: 2.4-3.3). Pre-dialysis age (p < 0.05), peritonitis rate (p < 0.05), and serum phosphorus at the end of follow-up (p < 0.05) were associated with technique failure in univariate analysis, while in multivariate analysis, only pre-dialysis age (RR 1.07, p = 0.001) and peritonitis rate (RR 481, p < 0.0001) were technique failure predictors. Patients from this single center of the west of Mexico were younger, had higher body surface area and initiated peritoneal dialysis with a more deteriorated general status than patients reported in other Mexican series; in spite of the latter, patient and technique survival were not different. In our setting, pre-dialysis older age and lower CrCl significantly predicted mortality, while older predialysis age and higher peritonitis rate predicted technique failure.
Davis, Tyler; LaRocque, Karen F.; Mumford, Jeanette; Norman, Kenneth A.; Wagner, Anthony D.; Poldrack, Russell A.
2014-01-01
Multi-voxel pattern analysis (MVPA) has led to major changes in how fMRI data are analyzed and interpreted. Many studies now report both MVPA results and results from standard univariate voxel-wise analysis, often with the goal of drawing different conclusions from each. Because MVPA results can be sensitive to latent multidimensional representations and processes whereas univariate voxel-wise analysis cannot, one conclusion that is often drawn when MVPA and univariate results differ is that the activation patterns underlying MVPA results contain a multidimensional code. In the current study, we conducted simulations to formally test this assumption. Our findings reveal that MVPA tests are sensitive to the magnitude of voxel-level variability in the effect of a condition within subjects, even when the same linear relationship is coded in all voxels. We also find that MVPA is insensitive to subject-level variability in mean activation across an ROI, which is the primary variance component of interest in many standard univariate tests. Together, these results illustrate that differences between MVPA and univariate tests do not afford conclusions about the nature or dimensionality of the neural code. Instead, targeted tests of the informational content and/or dimensionality of activation patterns are critical for drawing strong conclusions about the representational codes that are indicated by significant MVPA results. PMID:24768930
Vassilakopoulou, Maria; Parisi, Fabio; Siddiqui, Summar; England, Allison M; Zarella, Elizabeth R; Anagnostou, Valsamo; Kluger, Yuval; Hicks, David G; Rimm, David L; Neumeister, Veronique M
2015-03-01
Individualized targeted therapies for cancer patients require accurate and reproducible assessment of biomarkers to be able to plan treatment accordingly. Recent studies have shown highly variable effects of preanalytical variables on gene expression profiling and protein levels of different tissue types. Several publications have described protein degradation of tissue samples as a direct result of delay of formalin fixation of the tissue. Phosphorylated proteins are more labile and epitope degradation can happen within 30 min of cold ischemic time. To address this issue, we evaluated the change in antigenicity of a series of phosphoproteins in paraffin-embedded samples from breast tumors as a function of time to formalin fixation. A tissue microarray consisting of 93 breast cancer specimens with documented time-to-fixation was used to evaluate changes in antigenicity of 12 phosphoepitopes frequently used in research settings as a function of cold ischemic time. Analysis was performed in a quantitative manner using the AQUA technology for quantitative immunofluorescence. For each marker, least squares univariate linear regression was performed and confidence intervals were computed using bootstrapping. The majority of the epitopes tested revealed changes in expression levels with increasing time to formalin fixation. Some phosphorylated proteins, such as phospho-HSP27 and phospho-S6 RP, involved in post-translational modification and stress response pathways increased in expression or phosphorylation levels. Others (like phospho-AKT, phosphor-ERK1/2, phospho-Tyrosine, phospho-MET, and others) are quite labile and loss of antigenicity can be reported within 1-2 h of cold ischemic time. Therefore specimen collection should be closely monitored and subjected to quality control measures to ensure accurate measurement of these epitopes. However, a few phosphoepitopes (like phospho-JAK2 and phospho-ER) are sufficiently robust for routine usage in companion diagnostic testing.
Baradez, Marc-Olivier; Biziato, Daniela; Hassan, Enas; Marshall, Damian
2018-01-01
Cell therapies offer unquestionable promises for the treatment, and in some cases even the cure, of complex diseases. As we start to see more of these therapies gaining market authorization, attention is turning to the bioprocesses used for their manufacture, in particular the challenge of gaining higher levels of process control to help regulate cell behavior, manage process variability, and deliver product of a consistent quality. Many processes already incorporate the measurement of key markers such as nutrient consumption, metabolite production, and cell concentration, but these are often performed off-line and only at set time points in the process. Having the ability to monitor these markers in real-time using in-line sensors would offer significant advantages, allowing faster decision-making and a finer level of process control. In this study, we use Raman spectroscopy as an in-line optical sensor for bioprocess monitoring of an autologous T-cell immunotherapy model produced in a stirred tank bioreactor system. Using reference datasets generated on a standard bioanalyzer, we develop chemometric models from the Raman spectra for glucose, glutamine, lactate, and ammonia. These chemometric models can accurately monitor donor-specific increases in nutrient consumption and metabolite production as the primary T-cell transition from a recovery phase and begin proliferating. Using a univariate modeling approach, we then show how changes in peak intensity within the Raman spectra can be correlated with cell concentration and viability. These models, which act as surrogate markers, can be used to monitor cell behavior including cell proliferation rates, proliferative capacity, and transition of the cells to a quiescent phenotype. Finally, using the univariate models, we also demonstrate how Raman spectroscopy can be applied for real-time monitoring. The ability to measure these key parameters using an in-line Raman optical sensor makes it possible to have immediate feedback on process performance. This could help significantly improve cell therapy bioprocessing by allowing proactive decision-making based on real-time process data. Going forward, these types of in-line sensors also open up opportunities to improve bioprocesses further through concepts such as adaptive manufacturing. PMID:29556497
Baradez, Marc-Olivier; Biziato, Daniela; Hassan, Enas; Marshall, Damian
2018-01-01
Cell therapies offer unquestionable promises for the treatment, and in some cases even the cure, of complex diseases. As we start to see more of these therapies gaining market authorization, attention is turning to the bioprocesses used for their manufacture, in particular the challenge of gaining higher levels of process control to help regulate cell behavior, manage process variability, and deliver product of a consistent quality. Many processes already incorporate the measurement of key markers such as nutrient consumption, metabolite production, and cell concentration, but these are often performed off-line and only at set time points in the process. Having the ability to monitor these markers in real-time using in-line sensors would offer significant advantages, allowing faster decision-making and a finer level of process control. In this study, we use Raman spectroscopy as an in-line optical sensor for bioprocess monitoring of an autologous T-cell immunotherapy model produced in a stirred tank bioreactor system. Using reference datasets generated on a standard bioanalyzer, we develop chemometric models from the Raman spectra for glucose, glutamine, lactate, and ammonia. These chemometric models can accurately monitor donor-specific increases in nutrient consumption and metabolite production as the primary T-cell transition from a recovery phase and begin proliferating. Using a univariate modeling approach, we then show how changes in peak intensity within the Raman spectra can be correlated with cell concentration and viability. These models, which act as surrogate markers, can be used to monitor cell behavior including cell proliferation rates, proliferative capacity, and transition of the cells to a quiescent phenotype. Finally, using the univariate models, we also demonstrate how Raman spectroscopy can be applied for real-time monitoring. The ability to measure these key parameters using an in-line Raman optical sensor makes it possible to have immediate feedback on process performance. This could help significantly improve cell therapy bioprocessing by allowing proactive decision-making based on real-time process data. Going forward, these types of in-line sensors also open up opportunities to improve bioprocesses further through concepts such as adaptive manufacturing.
Analysis of Cross-Sectional Univariate Measurements for Family Dyads Using Linear Mixed Modeling
Knafl, George J.; Dixon, Jane K.; O'Malley, Jean P.; Grey, Margaret; Deatrick, Janet A.; Gallo, Agatha M.; Knafl, Kathleen A.
2010-01-01
Outcome measurements from members of the same family are likely correlated. Such intrafamilial correlation (IFC) is an important dimension of the family as a unit but is not always accounted for in analyses of family data. This article demonstrates the use of linear mixed modeling to account for IFC in the important special case of univariate measurements for family dyads collected at a single point in time. Example analyses of data from partnered parents having a child with a chronic condition on their child's adaptation to the condition and on the family's general functioning and management of the condition are provided. Analyses of this kind are reasonably straightforward to generate with popular statistical tools. Thus, it is recommended that IFC be reported as standard practice reflecting the fact that a family dyad is more than just the aggregate of two individuals. Moreover, not accounting for IFC can affect the conclusions. PMID:19307316
Determinants of survival after liver resection for metastatic colorectal carcinoma.
Parau, Angela; Todor, Nicolae; Vlad, Liviu
2015-01-01
Prognostic factors for survival after liver resection for metastatic colorectal cancer identified up to date are quite inconsistent with a great inter-study variability. In this study we aimed to identify predictors of outcome in our patient population. A series of 70 consecutive patients from the oncological hepatobiliary database, who had undergone curative hepatic surgical resection for hepatic metastases of colorectal origin, operated between 2006 and 2011, were identified. At 44.6 months (range 13.7-73), 30 of 70 patients (42.85%) were alive. Patient demographics, primary tumor and liver tumor factors, operative factors, pathologic findings, recurrence patterns, disease-free survival (DFS), overall survival (OS) and cancer-specific survival (CSS) were analyzed. Clinicopathologic variables were tested using univariate and multivariate analyses. The 3-year CSS after first hepatic resection was 54%. Median CSS survival after first hepatic resection was 40.2 months. Median CSS after second hepatic resection was 24.2 months. The 3-year DFS after first hepatic resection was 14%. Median disease free survival after first hepatic resection was 18 months. The 3-year DFS after second hepatic resection was 27% and median DFS after second hepatic resection 12 months. The 30-day mortality and morbidity rate after first hepatic resection was 5.71% and 12.78%, respectively. In univariate analysis CSS was significantly reduced for the following factors: age >53 years, advanced T stage of primary tumor, moderately- poorly differentiated tumor, positive and narrow resection margin, preoperative CEA level >30 ng/ml, DFS <18 months. Perioperative chemotherapy related to metastasectomy showed a trend in improving CSS (p=0.07). Perioperative chemotherapy improved DFS in a statistically significant way (p=0.03). Perioperative chemotherapy and achievement of resection margins beyond 1 mm were the major determinants of both CSS and DFS after first liver resection in multivariate analysis. In our series predictors of outcome in multivariate analysis were resection margins beyond 1mm and perioperative chemotherapy. Studies on larger population and analyses of additional clinicopathologic factors like genetic markers could contribute to development of clinical scoring models to assess the risk of relapse and survival.
Customer Service Analysis of Tactical Air Command Base Level Supply Support
1990-09-01
function. A large number of respondents described customer service as an activity such as order processing , handling of complaints, or troubleshooting...thru 14 General Service .69 19 thru 28 Demeanor of Supply .86 Representatives 29 thru 36 Order Processing .82 37 thru 40 Order Cycle Time .84 41 thru...Representatives 23 thru 30 Order Processing .83 31 thru 34 Order Cycle Time .75 35 thru 39 Item Availability .80 40 thru 45 Responsiveness .86 Univariate
Multifractal analysis of visibility graph-based Ito-related connectivity time series.
Czechowski, Zbigniew; Lovallo, Michele; Telesca, Luciano
2016-02-01
In this study, we investigate multifractal properties of connectivity time series resulting from the visibility graph applied to normally distributed time series generated by the Ito equations with multiplicative power-law noise. We show that multifractality of the connectivity time series (i.e., the series of numbers of links outgoing any node) increases with the exponent of the power-law noise. The multifractality of the connectivity time series could be due to the width of connectivity degree distribution that can be related to the exit time of the associated Ito time series. Furthermore, the connectivity time series are characterized by persistence, although the original Ito time series are random; this is due to the procedure of visibility graph that, connecting the values of the time series, generates persistence but destroys most of the nonlinear correlations. Moreover, the visibility graph is sensitive for detecting wide "depressions" in input time series.
ERIC Educational Resources Information Center
Mittelstaedt, H. Fred; Morris, Michael H.
2017-01-01
This study shows that graduates from nonprofit educational institutions outperform graduates from for-profit institutions on the four sections of the certified public accountant (CPA) exam. Specifically, it (1) documents univariate differences in CPA exam scores, score distributions, pass rates, and time to complete the CPA exam; (2) investigates…
Curriculum for Junior High School Students: Dairy Food Consumption Self-Efficacy
ERIC Educational Resources Information Center
Pobocik, Rebecca S.; Haar, Christine M.; Dawson, Erin E.; Coleman, Priscilla; Bakies, Karen
2009-01-01
A nutrition education module for a family and consumer sciences curriculum was developed and evaluated with junior high school students (n = 63) using a quasi-experimental design. The multivariate interaction between time of measurement and intervention was significant, F (2, 50) = 8.68, p = 0.001.The univariate interaction between pre and post…
Chest wall recurrence after mastectomy does not always portend a dismal outcome.
Chagpar, Anees; Meric-Bernstam, Funda; Hunt, Kelly K; Ross, Merrick I; Cristofanilli, Massimo; Singletary, S Eva; Buchholz, Thomas A; Ames, Frederick C; Marcy, Sylvie; Babiera, Gildy V; Feig, Barry W; Hortobagyi, Gabriel N; Kuerer, Henry M
2003-07-01
Chest wall recurrence (CWR) after mastectomy often forecasts a grim prognosis. Predictors of outcome after CWR, however, are not clear. From 1988 to 1998, 130 patients with isolated CWRs were seen at our center. Clinicopathologic factors were studied by univariate and multivariate analyses for distant metastasis-free survival after CWR. The median post-CWR follow-up was 37 months. Initial nodal status was the strongest predictor of outcome by univariate analysis. Other significant factors included initial T4 disease, primary lymphovascular invasion, treatment of the primary tumor with neoadjuvant therapy or radiation, time to CWR >24 months, and treatment for CWR (surgery, radiation, or multimodality therapy). Multivariate analysis also found initial nodal status to have the greatest effect; time to CWR and use of radiation for CWR were also independent predictors. Three groups of patients were identified. Low risk was defined by initial node-negative disease, time to CWR >24 months, and radiation for CWR; intermediate risk had one or two favorable features; and high risk had none. The median distant metastasis-free survival after CWR was significantly different among these groups (P <.0001). Patients with CWR are a heterogeneous population. Patients with initial node-negative disease who develop CWR after 24 months have an optimistic prognosis, especially if they are treated with radiation.
Mo, Xiaoliang; Qin, Guirong; Zhou, Zhoulin; Jiang, Xiaoli
2017-10-03
To explore the risk factors for intrauterine adhesions in patients with artificial abortion and clinical efficacy of hysteroscopic dissection. 1500 patients undergoing artificial abortion between January 2014 and June 2015 were enrolled into this study. The patients were divided into two groups with or without intrauterine adhesions. Univariate and Multiple logistic regression were conducted to assess the effects of multiple factors on the development of intrauterine adhesions following induced abortion. The incidence rate for intrauterine adhesions following induced abortion is 17.0%. Univariate showed that preoperative inflammation, multiple pregnancies and suction evacuation time are the influence risk factors of intrauterine adhesions. Multiple logistic regression demonstrates that multiple pregnancies, high intrauterine negative pressure, and long suction evacuation time are independent risk factors for the development of intrauterine adhesions following induced abortion. Additionally, intrauterine adhesions were observed in 105 mild, 80 moderate, and 70 severe cases. The cure rates for these three categories of intrauterine adhesions by hysteroscopic surgery were 100.0%, 93.8%, and 85.7%, respectively. Multiple pregnancies, high negative pressure suction evacuation and long suction evacuation time are independent risk factors for the development of intrauterine adhesions following induced abortions. Hysteroscopic surgery substantially improves the clinical outcomes of intrauterine adhesions.
[Risk factors for anorexia in children].
Liu, Wei-Xiao; Lang, Jun-Feng; Zhang, Qin-Feng
2016-11-01
To investigate the risk factors for anorexia in children, and to reduce the prevalence of anorexia in children. A questionnaire survey and a case-control study were used to collect the general information of 150 children with anorexia (case group) and 150 normal children (control group). Univariate analysis and multivariate logistic stepwise regression analysis were performed to identify the risk factors for anorexia in children. The results of the univariate analysis showed significant differences between the case and control groups in the age in months when supplementary food were added, feeding pattern, whether they liked meat, vegetables and salty food, whether they often took snacks and beverages, whether they liked to play while eating, and whether their parents asked them to eat food on time (P<0.05). The results of the multivariate logistic regression analysis showed that late addition of supplementary food (OR=5.408), high frequency of taking snacks and/or drinks (OR=11.813), and eating while playing (OR=6.654) were major risk factors for anorexia in children. Liking of meat (OR=0.093) and vegetables (OR=0.272) and eating on time required by parents (OR=0.079) were protective factors against anorexia in children. Timely addition of supplementary food, a proper diet, and development of children's proper eating and living habits can reduce the incidence of anorexia in children.
Causes of visual disability among Central Africans with diabetes mellitus.
Mvitu Muaka, M; Longo-Mbenza, B
2012-06-01
Diabetic Retinopathy (DR) remains a common and one of the major causes of blindness in the developed and western societies. The same situation is shown in emerging economic areas (5,6). In sub-Saharan Africa (SSA) however, the issues of visual disability due to diabetes mellitus (DM) are overshadowed by the presence of the prevalent and common nutritional deficiency diseases and eye infections This clinic-based study was conducted to determine whether diabetic retinopathy is independently related to visual disability in black patients with diabetes mellitus (DM) from Kinshasa, Congo. A total of 299 urban patients with DM and low income including 108 cases of visual disability and matched for time admission and DM type to 191 controls, were assessed. Demographic, clinical, and ophthalmic data were assessed using univariate and multivariate analyses. Age ≥60 years, female sex, presence of diabetic retinopathy (DR), proliferative DR, shorter DM duration, glaucoma, macular oedema, diabetic nephropathy were the univariate risk factors of visual disability. Using logistic regression model, visual disability was significantly associated with female sex and diabetic retinopathy. The risk of visual disability is 4 times higher in patients with diabetic retinopathy and 2 times higher in females with DM. Therefore, to prevent further increase of visual disability, the Congolese Ministry of Health should prioritize the eye care in patients with DM.
Diaz Beveridge, Robert; Alcolea, Vicent; Aparicio, Jorge; Segura, Ángel; García, Jose; Corbellas, Miguel; Fonfría, María; Giménez, Alejandra; Montalar, Joaquin
2014-01-10
The combination of gemcitabine and erlotinib is a standard first-line treatment for unresectable, locally advanced or metastatic pancreatic cancer. We reviewed our single centre experience to assess its efficacy and toxicity in clinical practice. Clinical records of patients with unresectable, locally advanced or metastatic pancreatic cancer who were treated with the combination of gemcitabine and erlotinib were reviewed. Univariate survival analysis and multivariate analysis were carried out to indentify independent predictors factors of overall survival. Our series included 55 patients. Overall disease control rate was 47%: 5% of patients presented complete response, 20% partial response and 22% stable disease. Median overall survival was 8.3 months). Cox regression analysis indicated that performance status and locally advanced versus metastatic disease were independent factors of overall survival. Patients who developed acne-like rash toxicity, related to erlotinib administration, presented a higher survival than those patients who did not develop this toxicity. Gemcitabine plus erlotinib doublet is active in our series of patients with advanced pancreatic cancer. This study provides efficacy and safety results similar to those of the pivotal phase III clinical trial that tested the same combination.
GPS Position Time Series @ JPL
NASA Technical Reports Server (NTRS)
Owen, Susan; Moore, Angelyn; Kedar, Sharon; Liu, Zhen; Webb, Frank; Heflin, Mike; Desai, Shailen
2013-01-01
Different flavors of GPS time series analysis at JPL - Use same GPS Precise Point Positioning Analysis raw time series - Variations in time series analysis/post-processing driven by different users. center dot JPL Global Time Series/Velocities - researchers studying reference frame, combining with VLBI/SLR/DORIS center dot JPL/SOPAC Combined Time Series/Velocities - crustal deformation for tectonic, volcanic, ground water studies center dot ARIA Time Series/Coseismic Data Products - Hazard monitoring and response focused center dot ARIA data system designed to integrate GPS and InSAR - GPS tropospheric delay used for correcting InSAR - Caltech's GIANT time series analysis uses GPS to correct orbital errors in InSAR - Zhen Liu's talking tomorrow on InSAR Time Series analysis
Laser-Induced Breakdown Spectroscopy (LIBS) Measurement of Uranium in Molten Salt.
Williams, Ammon; Phongikaroon, Supathorn
2018-01-01
In this current study, the molten salt aerosol-laser-induced breakdown spectroscopy (LIBS) system was used to measure the uranium (U) content in a ternary UCl 3 -LiCl-KCl salt to investigate and assess a near real-time analytical approach for material safeguards and accountability. Experiments were conducted using five different U concentrations to determine the analytical figures of merit for the system with respect to U. In the analysis, three U lines were used to develop univariate calibration curves at the 367.01 nm, 385.96 nm, and 387.10 nm lines. The 367.01 nm line had the lowest limit of detection (LOD) of 0.065 wt% U. The 385.96 nm line had the best root mean square error of cross-validation (RMSECV) of 0.20 wt% U. In addition to the univariate calibration approach, a multivariate partial least squares (PLS) model was developed to further analyze the data. Using partial least squares (PLS) modeling, an RMSECV of 0.085 wt% U was determined. The RMSECV from the multivariate approach was significantly better than the univariate case and the PLS model is recommended for future LIBS analysis. Overall, the aerosol-LIBS system performed well in monitoring the U concentration and it is expected that the system could be used to quantitatively determine the U compositions within the normal operational concentrations of U in pyroprocessing molten salts.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhan, Xianyuan; Aziz, H. M. Abdul; Ukkusuri, Satish V.
Our study investigates the Multivariate Poisson-lognormal (MVPLN) model that jointly models crash frequency and severity accounting for correlations. The ordinary univariate count models analyze crashes of different severity level separately ignoring the correlations among severity levels. The MVPLN model is capable to incorporate the general correlation structure and takes account of the over dispersion in the data that leads to a superior data fitting. But, the traditional estimation approach for MVPLN model is computationally expensive, which often limits the use of MVPLN model in practice. In this work, a parallel sampling scheme is introduced to improve the original Markov Chainmore » Monte Carlo (MCMC) estimation approach of the MVPLN model, which significantly reduces the model estimation time. Two MVPLN models are developed using the pedestrian vehicle crash data collected in New York City from 2002 to 2006, and the highway-injury data from Washington State (5-year data from 1990 to 1994) The Deviance Information Criteria (DIC) is used to evaluate the model fitting. The estimation results show that the MVPLN models provide a superior fit over univariate Poisson-lognormal (PLN), univariate Poisson, and Negative Binomial models. Moreover, the correlations among the latent effects of different severity levels are found significant in both datasets that justifies the importance of jointly modeling crash frequency and severity accounting for correlations.« less
Risk factors for tube exposure as a late complication of glaucoma drainage implant surgery.
Chaku, Meenakshi; Netland, Peter A; Ishida, Kyoko; Rhee, Douglas J
2016-01-01
The purpose of this study was to evaluate the risk factors for tube exposure after glaucoma drainage implant surgery. This was a retrospective case-controlled observational study of 64 eyes from 64 patients. Thirty-two eyes of 32 patients with tube erosion requiring surgical revision were compared with 32 matched control eyes of 32 patients. Univariate and multivariate risk factor analyses were performed. Mean age was significantly younger in the tube exposure group compared with the control group (48.2±28.1 years versus 67.3±18.0 years, respectively; P=0.003). The proportion of diabetic patients (12.5%) in the tube exposure group was significantly less (P=0.041) compared with the control group (37.5%). Comparisons of the type and position of the drainage implant were not significantly different between the two groups. The average time to tube exposure was 17.2±18.0 months after implantation of the drainage device. In both univariate and multivariate analyses, younger age (P=0.005 and P=0.027) and inflammation prior to tube exposure (P≤0.001 and P=0.004) were significant risk factors. Diabetes was a significant risk factor only in the univariate analysis (P=0.027). Younger age and inflammation were significant risk factors for tube exposure after drainage implant surgery.
Zhan, Xianyuan; Aziz, H. M. Abdul; Ukkusuri, Satish V.
2015-11-19
Our study investigates the Multivariate Poisson-lognormal (MVPLN) model that jointly models crash frequency and severity accounting for correlations. The ordinary univariate count models analyze crashes of different severity level separately ignoring the correlations among severity levels. The MVPLN model is capable to incorporate the general correlation structure and takes account of the over dispersion in the data that leads to a superior data fitting. But, the traditional estimation approach for MVPLN model is computationally expensive, which often limits the use of MVPLN model in practice. In this work, a parallel sampling scheme is introduced to improve the original Markov Chainmore » Monte Carlo (MCMC) estimation approach of the MVPLN model, which significantly reduces the model estimation time. Two MVPLN models are developed using the pedestrian vehicle crash data collected in New York City from 2002 to 2006, and the highway-injury data from Washington State (5-year data from 1990 to 1994) The Deviance Information Criteria (DIC) is used to evaluate the model fitting. The estimation results show that the MVPLN models provide a superior fit over univariate Poisson-lognormal (PLN), univariate Poisson, and Negative Binomial models. Moreover, the correlations among the latent effects of different severity levels are found significant in both datasets that justifies the importance of jointly modeling crash frequency and severity accounting for correlations.« less
Kiani, Behzad; Bagheri, Nasser; Tara, Ahmad; Hoseini, Benyamin; Tabesh, Hamed; Tara, Mahmood
2017-11-07
Poor access to haemodialysis facilities is associated with high mortality and morbidity rates. This study investigated factors affecting revealed access to the haemodialysis facilities considering patients living in rural and urban areas without any haemodialysis facility (Group A) and those living urban areas with haemodialysis facilities (Group B). This study is based on selfreported Actual Access Time (AAT) to referred haemodialysis facilities and other information regarding travel to haemodialysis facilities from patients. All significant variables on univariate analysis were entered into a univariate general linear model in order to identify factors associated with AAT. Both spatial (driving time and distance) and non-spatial factors (sex, income level, caregivers, transportation mode, education level, ethnicity and personal vehicle ownership) influenced the revealed access identified in Group A. The non-spatial factors for Group B patients were the same as for Group A, but no spatial factor was identified in Group B. It was found that accessibility is strongly underestimated when driving time is chosen as accessibility measure to haemodialysis facilities. Analysis of revealed access determinants provides policymakers with an appropriate decision base for making appropriate decisions and finding solutions to decrease the access time for patients under haemodialysis therapy. Driving time alone is not a good proxy for measuring access to haemodialysis facilities as there are many other potential obstacles, such as women's special travel problems, poor other transportation possibilities, ethnicity disparities, low education levels, low caregiver status and low-income.
FUNSTAT and statistical image representations
NASA Technical Reports Server (NTRS)
Parzen, E.
1983-01-01
General ideas of functional statistical inference analysis of one sample and two samples, univariate and bivariate are outlined. ONESAM program is applied to analyze the univariate probability distributions of multi-spectral image data.
Lara, Juan A; Lizcano, David; Pérez, Aurora; Valente, Juan P
2014-10-01
There are now domains where information is recorded over a period of time, leading to sequences of data known as time series. In many domains, like medicine, time series analysis requires to focus on certain regions of interest, known as events, rather than analyzing the whole time series. In this paper, we propose a framework for knowledge discovery in both one-dimensional and multidimensional time series containing events. We show how our approach can be used to classify medical time series by means of a process that identifies events in time series, generates time series reference models of representative events and compares two time series by analyzing the events they have in common. We have applied our framework on time series generated in the areas of electroencephalography (EEG) and stabilometry. Framework performance was evaluated in terms of classification accuracy, and the results confirmed that the proposed schema has potential for classifying EEG and stabilometric signals. The proposed framework is useful for discovering knowledge from medical time series containing events, such as stabilometric and electroencephalographic time series. These results would be equally applicable to other medical domains generating iconographic time series, such as, for example, electrocardiography (ECG). Copyright © 2014 Elsevier Inc. All rights reserved.
Detection of a sudden change of the field time series based on the Lorenz system.
Da, ChaoJiu; Li, Fang; Shen, BingLu; Yan, PengCheng; Song, Jian; Ma, DeShan
2017-01-01
We conducted an exploratory study of the detection of a sudden change of the field time series based on the numerical solution of the Lorenz system. First, the time when the Lorenz path jumped between the regions on the left and right of the equilibrium point of the Lorenz system was quantitatively marked and the sudden change time of the Lorenz system was obtained. Second, the numerical solution of the Lorenz system was regarded as a vector; thus, this solution could be considered as a vector time series. We transformed the vector time series into a time series using the vector inner product, considering the geometric and topological features of the Lorenz system path. Third, the sudden change of the resulting time series was detected using the sliding t-test method. Comparing the test results with the quantitatively marked time indicated that the method could detect every sudden change of the Lorenz path, thus the method is effective. Finally, we used the method to detect the sudden change of the pressure field time series and temperature field time series, and obtained good results for both series, which indicates that the method can apply to high-dimension vector time series. Mathematically, there is no essential difference between the field time series and vector time series; thus, we provide a new method for the detection of the sudden change of the field time series.
Volatility of linear and nonlinear time series
NASA Astrophysics Data System (ADS)
Kalisky, Tomer; Ashkenazy, Yosef; Havlin, Shlomo
2005-07-01
Previous studies indicated that nonlinear properties of Gaussian distributed time series with long-range correlations, ui , can be detected and quantified by studying the correlations in the magnitude series ∣ui∣ , the “volatility.” However, the origin for this empirical observation still remains unclear and the exact relation between the correlations in ui and the correlations in ∣ui∣ is still unknown. Here we develop analytical relations between the scaling exponent of linear series ui and its magnitude series ∣ui∣ . Moreover, we find that nonlinear time series exhibit stronger (or the same) correlations in the magnitude time series compared with linear time series with the same two-point correlations. Based on these results we propose a simple model that generates multifractal time series by explicitly inserting long range correlations in the magnitude series; the nonlinear multifractal time series is generated by multiplying a long-range correlated time series (that represents the magnitude series) with uncorrelated time series [that represents the sign series sgn(ui) ]. We apply our techniques on daily deep ocean temperature records from the equatorial Pacific, the region of the El-Ninõ phenomenon, and find: (i) long-range correlations from several days to several years with 1/f power spectrum, (ii) significant nonlinear behavior as expressed by long-range correlations of the volatility series, and (iii) broad multifractal spectrum.
Duality between Time Series and Networks
Campanharo, Andriana S. L. O.; Sirer, M. Irmak; Malmgren, R. Dean; Ramos, Fernando M.; Amaral, Luís A. Nunes.
2011-01-01
Studying the interaction between a system's components and the temporal evolution of the system are two common ways to uncover and characterize its internal workings. Recently, several maps from a time series to a network have been proposed with the intent of using network metrics to characterize time series. Although these maps demonstrate that different time series result in networks with distinct topological properties, it remains unclear how these topological properties relate to the original time series. Here, we propose a map from a time series to a network with an approximate inverse operation, making it possible to use network statistics to characterize time series and time series statistics to characterize networks. As a proof of concept, we generate an ensemble of time series ranging from periodic to random and confirm that application of the proposed map retains much of the information encoded in the original time series (or networks) after application of the map (or its inverse). Our results suggest that network analysis can be used to distinguish different dynamic regimes in time series and, perhaps more importantly, time series analysis can provide a powerful set of tools that augment the traditional network analysis toolkit to quantify networks in new and useful ways. PMID:21858093
The Effect of Patient and Surgical Characteristics on Renal Function After Partial Nephrectomy.
Winer, Andrew G; Zabor, Emily C; Vacchio, Michael J; Hakimi, A Ari; Russo, Paul; Coleman, Jonathan A; Jaimes, Edgar A
2018-06-01
The purpose of the study was to identify patient and disease characteristics that have an adverse effect on renal function after partial nephrectomy. We conducted a retrospective review of 387 patients who underwent partial nephrectomy for renal tumors between 2006 and 2014. A line plot with a locally weighted scatterplot smoothing was generated to visually assess renal function over time. Univariable and multivariable longitudinal regression analyses incorporated a random intercept and slope to evaluate the association between patient and disease characteristics with renal function after surgery. Median age was 60 years and most patients were male (255 patients [65.9%]) and white (343 patients [88.6%]). In univariable analysis, advanced age at surgery, larger tumor size, male sex, longer ischemia time, history of smoking, and hypertension were significantly associated with lower preoperative estimated glomerular filtration rate (eGFR). In multivariable analysis, independent predictors of reduced renal function after surgery included advanced age, lower preoperative eGFR, and longer ischemia time. Length of time from surgery was strongly associated with improvement in renal function among all patients. Independent predictors of postoperative decline in renal function include advanced age, lower preoperative eGFR, and longer ischemia time. A substantial number of subjects had recovery in renal function over time after surgery, which continued past the 12-month mark. These findings suggest that patients who undergo partial nephrectomy can experience long-term improvement in renal function. This improvement is most pronounced among younger patients with higher preoperative eGFR. Copyright © 2017 Elsevier Inc. All rights reserved.
Kryklywy, James H; Macpherson, Ewan A; Mitchell, Derek G V
2018-04-01
Emotion can have diverse effects on behaviour and perception, modulating function in some circumstances, and sometimes having little effect. Recently, it was identified that part of the heterogeneity of emotional effects could be due to a dissociable representation of emotion in dual pathway models of sensory processing. Our previous fMRI experiment using traditional univariate analyses showed that emotion modulated processing in the auditory 'what' but not 'where' processing pathway. The current study aims to further investigate this dissociation using a more recently emerging multi-voxel pattern analysis searchlight approach. While undergoing fMRI, participants localized sounds of varying emotional content. A searchlight multi-voxel pattern analysis was conducted to identify activity patterns predictive of sound location and/or emotion. Relative to the prior univariate analysis, MVPA indicated larger overlapping spatial and emotional representations of sound within early secondary regions associated with auditory localization. However, consistent with the univariate analysis, these two dimensions were increasingly segregated in late secondary and tertiary regions of the auditory processing streams. These results, while complimentary to our original univariate analyses, highlight the utility of multiple analytic approaches for neuroimaging, particularly for neural processes with known representations dependent on population coding.
Factors Influencing Cecal Intubation Time during Retrograde Approach Single-Balloon Enteroscopy
Chen, Peng-Jen; Shih, Yu-Lueng; Huang, Hsin-Hung; Hsieh, Tsai-Yuan
2014-01-01
Background and Aim. The predisposing factors for prolonged cecal intubation time (CIT) during colonoscopy have been well identified. However, the factors influencing CIT during retrograde SBE have not been addressed. The aim of this study was to determine the factors influencing CIT during retrograde SBE. Methods. We investigated patients who underwent retrograde SBE at a medical center from January 2011 to March 2014. The medical charts and SBE reports were reviewed. The patients' characteristics and procedure-associated data were recorded. These data were analyzed with univariate analysis as well as multivariate logistic regression analysis to identify the possible predisposing factors. Results. We enrolled 66 patients into this study. The median CIT was 17.4 minutes. With univariate analysis, there was no statistical difference in age, sex, BMI, or history of abdominal surgery, except for bowel preparation (P = 0.021). Multivariate logistic regression analysis showed that inadequate bowel preparation (odds ratio 30.2, 95% confidence interval 4.63–196.54; P < 0.001) was the independent predisposing factors for prolonged CIT during retrograde SBE. Conclusions. For experienced endoscopist, inadequate bowel preparation was the independent predisposing factor for prolonged CIT during retrograde SBE. PMID:25505904
Multiple Hypothesis Testing for Experimental Gingivitis Based on Wilcoxon Signed Rank Statistics
Preisser, John S.; Sen, Pranab K.; Offenbacher, Steven
2011-01-01
Dental research often involves repeated multivariate outcomes on a small number of subjects for which there is interest in identifying outcomes that exhibit change in their levels over time as well as to characterize the nature of that change. In particular, periodontal research often involves the analysis of molecular mediators of inflammation for which multivariate parametric methods are highly sensitive to outliers and deviations from Gaussian assumptions. In such settings, nonparametric methods may be favored over parametric ones. Additionally, there is a need for statistical methods that control an overall error rate for multiple hypothesis testing. We review univariate and multivariate nonparametric hypothesis tests and apply them to longitudinal data to assess changes over time in 31 biomarkers measured from the gingival crevicular fluid in 22 subjects whereby gingivitis was induced by temporarily withholding tooth brushing. To identify biomarkers that can be induced to change, multivariate Wilcoxon signed rank tests for a set of four summary measures based upon area under the curve are applied for each biomarker and compared to their univariate counterparts. Multiple hypothesis testing methods with choice of control of the false discovery rate or strong control of the family-wise error rate are examined. PMID:21984957
Risk factors for granuloma formation in children induced by tracheobronchial foreign bodies.
Huang, Zhenghua; Zhou, Ai; Zhang, Jianya; Xie, Lisheng; Li, Qi
2015-12-01
The aim of this study was to analyze the risk factors for granuloma formation caused by plant-based tracheobronchial foreign bodies in children, and investigate the underlying pathogenesis. In this retrospective analysis of 153 cases with tracheobronchial foreign bodies (peanuts and watermelon seeds), 35 cases of granuloma formation as granulation group (G), and 118 cases of no granuloma formation as non-granulation group (NG) were studied. Clinical data pertaining to sex (S), age (A), foreign body surface smoothness (SF), foreign body shape (SH), foreign body oil release state (O), the location of foreign bodies (L), and foreign body retention time (T) were collected for statistical analysis. Univariate analysis showed no significant difference between the two groups (G and NG) with respect to S, A, SH and L. Significant factors based on univariate analysis included SF, O and T. Multivariate logistic regression analysis revealed that SF and T were independent risk factors associated with development of granuloma. SF, O and T had relationship with the granuloma formation. Local trauma caused by an irregular and sharp foreign body, and extended period of time represent the main factors causing granuloma formation. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
The Impact of a Sports Vision Training Program in Youth Field Hockey Players
Schwab, Sebastian; Memmert, Daniel
2012-01-01
The aim of this study was to investigate whether a sports vision training program improves the visual performance of youth male field hockey players, ages 12 to 16 years, after an intervention of six weeks compared to a control group with no specific sports vision training. The choice reaction time task at the D2 board (Learning Task I), the functional field of view task (Learning Task II) and the multiple object tracking (MOT) task (Transfer Task) were assessed before and after the intervention and again six weeks after the second test. Analyzes showed significant differences between the two groups for the choice reaction time task at the D2 board and the functional field of view task, with significant improvements for the intervention group and none for the control group. For the transfer task, we could not find statistically significant improvements for either group. The results of this study are discussed in terms of theoretical and practical implications. Key pointsPerceptual training with youth field hockey playersCan a sports vision training program improve the visual performance of youth male field hockey players, ages 12 to 16 years, after an intervention of six weeks compared to a control group with no specific sports vision training?The intervention was performed in the “VisuLab” as DynamicEye® SportsVision Training at the German Sport University Cologne.We ran a series of 3 two-factor univariate analysis of variance (ANOVA) with repeated measures on both within subject independent variables (group; measuring point) to examine the effects on central perception, peripheral perception and choice reaction time.The present study shows an improvement of certain visual abilities with the help of the sports vision training program. PMID:24150071
Detection of a sudden change of the field time series based on the Lorenz system
Li, Fang; Shen, BingLu; Yan, PengCheng; Song, Jian; Ma, DeShan
2017-01-01
We conducted an exploratory study of the detection of a sudden change of the field time series based on the numerical solution of the Lorenz system. First, the time when the Lorenz path jumped between the regions on the left and right of the equilibrium point of the Lorenz system was quantitatively marked and the sudden change time of the Lorenz system was obtained. Second, the numerical solution of the Lorenz system was regarded as a vector; thus, this solution could be considered as a vector time series. We transformed the vector time series into a time series using the vector inner product, considering the geometric and topological features of the Lorenz system path. Third, the sudden change of the resulting time series was detected using the sliding t-test method. Comparing the test results with the quantitatively marked time indicated that the method could detect every sudden change of the Lorenz path, thus the method is effective. Finally, we used the method to detect the sudden change of the pressure field time series and temperature field time series, and obtained good results for both series, which indicates that the method can apply to high-dimension vector time series. Mathematically, there is no essential difference between the field time series and vector time series; thus, we provide a new method for the detection of the sudden change of the field time series. PMID:28141832
Yellman, Merissa A; Rodriguez, Marissa A; Colunga, Maria Isabel; McCoy, Mary A; Stephens-Stidham, Shelli; Brown, L Steven; Istre, Gregory R
2018-05-19
This study evaluated the effectiveness of a series of 1-year multifaceted school-based programs aimed at increasing booster seat use among urban children 4-7 years of age in economically disadvantaged areas. During 4 consecutive school years, 2011-2015, the Give Kids a Boost (GKB) program was implemented in a total of 8 schools with similar demographics in Dallas County. Observational surveys were conducted at project schools before project implementation (P 0 ), 1-4 weeks after the completion of project implementation (P 1 ), and 4-5 months later (P 2 ). Changes in booster seat use for the 3 time periods were compared for the 8 project and 14 comparison schools that received no intervention using a nonrandomized trial process. The intervention included (1) train-the-trainer sessions with teachers and parents; (2) presentations about booster seat safety; (3) tailored communication to parents; (4) distribution of fact sheets/resources; (5) walk-around education; and (6) booster seat inspections. The association between the GKB intervention and proper booster seat use was determined initially using univariate analysis. The association was also estimated using a generalized linear mixed model predicting a binomial outcome (booster seat use) for those aged 4 to 7 years, adjusted for child-level variables (age, sex, race/ethnicity) and car-level variables (vehicle type). The model incorporated the effects of clustering by site and by collection date to account for the possibility of repeated sampling. In the 8 project schools, booster seat use for children 4-7 years of age increased an average of 20.9 percentage points between P 0 and P 1 (P 0 = 4.8%, P 1 = 25.7%; odds ratio [OR] = 6.9; 95% confidence interval [CI], 5.5, 8.7; P < .001) and remained at that level in the P 2 time period (P 2 = 25.7%; P < .001, for P 0 vs. P 2 ) in the univariate analysis. The 14 comparison schools had minimal change in booster seat use. The multivariable model showed that children at the project schools were significantly more likely to be properly restrained in a booster seat after the intervention (OR = 2.7; 95% CI, 2.2, 3.3) compared to the P 0 time period and compared to the comparison schools. Despite study limitations, the GKB program was positively associated with an increase in proper booster seat use for children 4-7 years of age in school settings among diverse populations in economically disadvantaged areas. These increases persisted into the following school year in a majority of the project schools. The GKB model may be a replicable strategy to increase booster seat use among school-age children in similar urban settings.
Relevant Feature Set Estimation with a Knock-out Strategy and Random Forests
Ganz, Melanie; Greve, Douglas N.; Fischl, Bruce; Konukoglu, Ender
2015-01-01
Group analysis of neuroimaging data is a vital tool for identifying anatomical and functional variations related to diseases as well as normal biological processes. The analyses are often performed on a large number of highly correlated measurements using a relatively smaller number of samples. Despite the correlation structure, the most widely used approach is to analyze the data using univariate methods followed by post-hoc corrections that try to account for the data’s multivariate nature. Although widely used, this approach may fail to recover from the adverse effects of the initial analysis when local effects are not strong. Multivariate pattern analysis (MVPA) is a powerful alternative to the univariate approach for identifying relevant variations. Jointly analyzing all the measures, MVPA techniques can detect global effects even when individual local effects are too weak to detect with univariate analysis. Current approaches are successful in identifying variations that yield highly predictive and compact models. However, they suffer from lessened sensitivity and instabilities in identification of relevant variations. Furthermore, current methods’ user-defined parameters are often unintuitive and difficult to determine. In this article, we propose a novel MVPA method for group analysis of high-dimensional data that overcomes the drawbacks of the current techniques. Our approach explicitly aims to identify all relevant variations using a “knock-out” strategy and the Random Forest algorithm. In evaluations with synthetic datasets the proposed method achieved substantially higher sensitivity and accuracy than the state-of-the-art MVPA methods, and outperformed the univariate approach when the effect size is low. In experiments with real datasets the proposed method identified regions beyond the univariate approach, while other MVPA methods failed to replicate the univariate results. More importantly, in a reproducibility study with the well-known ADNI dataset the proposed method yielded higher stability and power than the univariate approach. PMID:26272728
Multivariate and Multiscale Data Assimilation in Terrestrial Systems: A Review
Montzka, Carsten; Pauwels, Valentijn R. N.; Franssen, Harrie-Jan Hendricks; Han, Xujun; Vereecken, Harry
2012-01-01
More and more terrestrial observational networks are being established to monitor climatic, hydrological and land-use changes in different regions of the World. In these networks, time series of states and fluxes are recorded in an automated manner, often with a high temporal resolution. These data are important for the understanding of water, energy, and/or matter fluxes, as well as their biological and physical drivers and interactions with and within the terrestrial system. Similarly, the number and accuracy of variables, which can be observed by spaceborne sensors, are increasing. Data assimilation (DA) methods utilize these observations in terrestrial models in order to increase process knowledge as well as to improve forecasts for the system being studied. The widely implemented automation in observing environmental states and fluxes makes an operational computation more and more feasible, and it opens the perspective of short-time forecasts of the state of terrestrial systems. In this paper, we review the state of the art with respect to DA focusing on the joint assimilation of observational data precedents from different spatial scales and different data types. An introduction is given to different DA methods, such as the Ensemble Kalman Filter (EnKF), Particle Filter (PF) and variational methods (3/4D-VAR). In this review, we distinguish between four major DA approaches: (1) univariate single-scale DA (UVSS), which is the approach used in the majority of published DA applications, (2) univariate multiscale DA (UVMS) referring to a methodology which acknowledges that at least some of the assimilated data are measured at a different scale than the computational grid scale, (3) multivariate single-scale DA (MVSS) dealing with the assimilation of at least two different data types, and (4) combined multivariate multiscale DA (MVMS). Finally, we conclude with a discussion on the advantages and disadvantages of the assimilation of multiple data types in a simulation model. Existing approaches can be used to simultaneously update several model states and model parameters if applicable. In other words, the basic principles for multivariate data assimilation are already available. We argue that a better understanding of the measurement errors for different observation types, improved estimates of observation bias and improved multiscale assimilation methods for data which scale nonlinearly is important to properly weight them in multiscale multivariate data assimilation. In this context, improved cross-validation of different data types, and increased ground truth verification of remote sensing products are required. PMID:23443380
Multivariate and multiscale data assimilation in terrestrial systems: a review.
Montzka, Carsten; Pauwels, Valentijn R N; Franssen, Harrie-Jan Hendricks; Han, Xujun; Vereecken, Harry
2012-11-26
More and more terrestrial observational networks are being established to monitor climatic, hydrological and land-use changes in different regions of the World. In these networks, time series of states and fluxes are recorded in an automated manner, often with a high temporal resolution. These data are important for the understanding of water, energy, and/or matter fluxes, as well as their biological and physical drivers and interactions with and within the terrestrial system. Similarly, the number and accuracy of variables, which can be observed by spaceborne sensors, are increasing. Data assimilation (DA) methods utilize these observations in terrestrial models in order to increase process knowledge as well as to improve forecasts for the system being studied. The widely implemented automation in observing environmental states and fluxes makes an operational computation more and more feasible, and it opens the perspective of short-time forecasts of the state of terrestrial systems. In this paper, we review the state of the art with respect to DA focusing on the joint assimilation of observational data precedents from different spatial scales and different data types. An introduction is given to different DA methods, such as the Ensemble Kalman Filter (EnKF), Particle Filter (PF) and variational methods (3/4D-VAR). In this review, we distinguish between four major DA approaches: (1) univariate single-scale DA (UVSS), which is the approach used in the majority of published DA applications, (2) univariate multiscale DA (UVMS) referring to a methodology which acknowledges that at least some of the assimilated data are measured at a different scale than the computational grid scale, (3) multivariate single-scale DA (MVSS) dealing with the assimilation of at least two different data types, and (4) combined multivariate multiscale DA (MVMS). Finally, we conclude with a discussion on the advantages and disadvantages of the assimilation of multiple data types in a simulation model. Existing approaches can be used to simultaneously update several model states and model parameters if applicable. In other words, the basic principles for multivariate data assimilation are already available. We argue that a better understanding of the measurement errors for different observation types, improved estimates of observation bias and improved multiscale assimilation methods for data which scale nonlinearly is important to properly weight them in multiscale multivariate data assimilation. In this context, improved cross-validation of different data types, and increased ground truth verification of remote sensing products are required.
Multiple Indicator Stationary Time Series Models.
ERIC Educational Resources Information Center
Sivo, Stephen A.
2001-01-01
Discusses the propriety and practical advantages of specifying multivariate time series models in the context of structural equation modeling for time series and longitudinal panel data. For time series data, the multiple indicator model specification improves on classical time series analysis. For panel data, the multiple indicator model…
Prognostic Factors in Glioblastoma: Is There a Role for Epilepsy?
DOBRAN, Mauro; NASI, Davide; CHIRIATTI, Stefano; GLADI, Maurizio; di SOMMA, Lucia; IACOANGELI, Maurizio; SCERRATI, Massimo
2018-01-01
The prognostic relevance of epilepsy at glioblastoma (GBMs) onset is still under debate. In this study, we analyzed the value of epilepsy and other prognostic factors on GBMs survival. We retrospectively analyzed the clinical, radiological, surgical and histological data in 139 GBMs. Seizures were the presenting symptoms in 50 patients out of 139 (35.9%). 123 patients (88%) were treated with craniotomy and tumor resection while 16 (12%) with biopsy. The median overall survival was 9.9 months from surgery. At univariable Cox regression, the factors that significantly improved survival were age less than 65 years (P = 0.0015), focal without impairment of consciousness seizures at presentation (P = 0.043), complete surgical resection (P < 0.001), pre-operative Karnofsky performance status (KPS) > 70 (P = 0.015), frontal location (P < 0.001), radiotherapy (XRT) plus concomitant and adjuvant TMZ (P < 0.001). A multivariable Cox regression showed that the complete surgical resection (P < 0.0001), age less than 65 years (P = 0.008), frontal location (P = 0.0001) and XRT adjuvant temozolomide (TMZ) (P < 0.0001) were independent factors on longer survival. In our series epilepsy at presentation is not an independent prognostic factor for longer survival in GBM patients. Only in the subgroup of patients with focal seizures without impairment of consciousness, epilepsy was associated with an increased significant overall survival at univariate analysis (P = 0.043). Main independent factors for relatively favorable GBMs outcome are complete tumor resection plus combined XRT-TMZ, frontal location and patient age below 65 years old. PMID:29343677
Scenario generation for stochastic optimization problems via the sparse grid method
Chen, Michael; Mehrotra, Sanjay; Papp, David
2015-04-19
We study the use of sparse grids in the scenario generation (or discretization) problem in stochastic programming problems where the uncertainty is modeled using a continuous multivariate distribution. We show that, under a regularity assumption on the random function involved, the sequence of optimal objective function values of the sparse grid approximations converges to the true optimal objective function values as the number of scenarios increases. The rate of convergence is also established. We treat separately the special case when the underlying distribution is an affine transform of a product of univariate distributions, and show how the sparse grid methodmore » can be adapted to the distribution by the use of quadrature formulas tailored to the distribution. We numerically compare the performance of the sparse grid method using different quadrature rules with classic quasi-Monte Carlo (QMC) methods, optimal rank-one lattice rules, and Monte Carlo (MC) scenario generation, using a series of utility maximization problems with up to 160 random variables. The results show that the sparse grid method is very efficient, especially if the integrand is sufficiently smooth. In such problems the sparse grid scenario generation method is found to need several orders of magnitude fewer scenarios than MC and QMC scenario generation to achieve the same accuracy. As a result, it is indicated that the method scales well with the dimension of the distribution--especially when the underlying distribution is an affine transform of a product of univariate distributions, in which case the method appears scalable to thousands of random variables.« less
SALAH, SAMER; TOUBASI, SAMAR
2015-01-01
Pulmonary metastasectomy (PM) is associated with improved survival of patients with metastatic osteosarcoma; however, the factors affecting survival following achievement of complete surgical remission remain controversial. The main objective of this study was to report the outcomes and prognostic factors of osteosarcoma patients who achieved complete remission (CR) following PM. We analyzed the effect of demographic and disease-related characteristics on the overall survival (OS) of consecutive patients with metastatic osteosarcoma who were treated at a single institution and achieved CR following PM, through univariate and multivariate analyses. Between January, 2000 and August, 2013, 62 patients with metastatic osteosarcoma were treated and followed up at our institution. A total of 25 patients achieved CR following PM and were included in this analysis. The 5-year OS and disease-free survival following PM were 30 and 21%, respectively. The factors correlated with inferior OS in the univariate analysis included chondroblastic subtype, post-chemotherapy necrosis <90% in the primary tumor, metastasis detected during neoadjuvant or adjuvant chemotherapy and pathological identification of tumor cells reaching the visceral pleural surface of any of the resected nodules. In the multivariate analysis, the chondroblastic subtype was the sole independent adverse prognostic factor (HR=4.6, 95% CI: 1.0–21.3, P=0.044). Therefore, factors associated with tumor biology, including poor tumor necrosis in the primary tumor and detection of metastasis during primary chemotherapy, are associated with poor post-metastasectomy survival. In addition, chondroblastic subtype and visceral pleural involvement predicted poor prognosis in our series. PMID:25469287
Does buccal cancer have worse prognosis than other oral cavity cancers?
Camilon, P Ryan; Stokes, William A; Fuller, Colin W; Nguyen, Shaun A; Lentsch, Eric J
2014-06-01
To determine whether buccal squamous cell carcinoma has worse overall survival (OS) and disease-specific survival (DSS) than cancers in the rest of the oral cavity. Retrospective analysis of a large population database. We began with a Kaplan-Meier analysis of OS and DSS for buccal versus nonbuccal tumors with unmatched data, followed by an analysis of cases matched for race, age at diagnosis, stage at diagnosis, and treatment modality. This was supported by a univariate Cox regression comparing buccal cancer to nonbuccal cancer, followed by a multivariate Cox regression that included all significant variables studied. With unmatched data, buccal cancer had significantly lesser OS and DSS values than cancers in the rest of the oral cavity (P < .001). After case matching, the differences between OS and DSS for buccal cancer versus nonbuccal oral cancer were no longer significant. Univariate Cox regression models with respect to OS and DSS showed a significant difference between buccal cancer and nonbuccal cancer. However, with multivariate analysis, buccal hazard ratios for OS and DSS were not significant. With the largest series of buccal carcinoma to date, our study concludes that the OS and DSS of buccal cancer are similar to those of cancers in other oral cavity sites once age at diagnosis, tumor stage, treatment, and race are taken into consideration. The previously perceived poor prognosis of buccal carcinoma may be due to variations in tumor presentation, such as later stage and older patient age. 2b. © 2014 The American Laryngological, Rhinological and Otological Society, Inc.
Efficient Algorithms for Segmentation of Item-Set Time Series
NASA Astrophysics Data System (ADS)
Chundi, Parvathi; Rosenkrantz, Daniel J.
We propose a special type of time series, which we call an item-set time series, to facilitate the temporal analysis of software version histories, email logs, stock market data, etc. In an item-set time series, each observed data value is a set of discrete items. We formalize the concept of an item-set time series and present efficient algorithms for segmenting a given item-set time series. Segmentation of a time series partitions the time series into a sequence of segments where each segment is constructed by combining consecutive time points of the time series. Each segment is associated with an item set that is computed from the item sets of the time points in that segment, using a function which we call a measure function. We then define a concept called the segment difference, which measures the difference between the item set of a segment and the item sets of the time points in that segment. The segment difference values are required to construct an optimal segmentation of the time series. We describe novel and efficient algorithms to compute segment difference values for each of the measure functions described in the paper. We outline a dynamic programming based scheme to construct an optimal segmentation of the given item-set time series. We use the item-set time series segmentation techniques to analyze the temporal content of three different data sets—Enron email, stock market data, and a synthetic data set. The experimental results show that an optimal segmentation of item-set time series data captures much more temporal content than a segmentation constructed based on the number of time points in each segment, without examining the item set data at the time points, and can be used to analyze different types of temporal data.
A novel weight determination method for time series data aggregation
NASA Astrophysics Data System (ADS)
Xu, Paiheng; Zhang, Rong; Deng, Yong
2017-09-01
Aggregation in time series is of great importance in time series smoothing, predicting and other time series analysis process, which makes it crucial to address the weights in times series correctly and reasonably. In this paper, a novel method to obtain the weights in time series is proposed, in which we adopt induced ordered weighted aggregation (IOWA) operator and visibility graph averaging (VGA) operator and linearly combine the weights separately generated by the two operator. The IOWA operator is introduced to the weight determination of time series, through which the time decay factor is taken into consideration. The VGA operator is able to generate weights with respect to the degree distribution in the visibility graph constructed from the corresponding time series, which reflects the relative importance of vertices in time series. The proposed method is applied to two practical datasets to illustrate its merits. The aggregation of Construction Cost Index (CCI) demonstrates the ability of proposed method to smooth time series, while the aggregation of The Taiwan Stock Exchange Capitalization Weighted Stock Index (TAIEX) illustrate how proposed method maintain the variation tendency of original data.
Decloe, Melissa; Gill, Suzanne; Ho, Grace; McCready, Janine; Powis, Jeff
2017-01-01
Background The success of antimicrobial stewardship is dependent on how often it is completed and which antimicrobials are targeted. We evaluated the impact of an antimicrobial stewardship program (ASP) in three non-ICU settings where all systemic antibiotics, regardless of spectrum, were targeted on the first weekday after initiation. Methods Prospective audit and feedback (PAAF) was initiated on the surgical, respiratory, and medical wards of a community hospital on July 1, 2010, October 1, 2010, and April 1, 2012, respectively. We evaluated rates of total antibiotic use, measured in days on therapy (DOTs), among all patients admitted to the wards before and after PAAF initiation using an interrupted time series analysis. Changes in antibiotic costs, rates of C. difficile infection (CDI), mortality, readmission, and length of stay were evaluated using univariate analyses. Results Time series modelling demonstrated that total antibiotic use decreased (± standard error) by 100 ± 51 DOTs/1,000 patient-days on the surgical wards (p = 0.049), 100 ± 46 DOTs/1,000 patient-days on the respiratory ward (p = 0.029), and 91 ± 33 DOTs/1,000 patient-days on the medical wards (p = 0.006) immediately following PAAF initiation. Reductions in antibiotic use were sustained up to 50 months after intervention initiation, and were accompanied by decreases in antibiotic costs. There were no significant changes to patient outcomes on the surgical and respiratory wards following intervention initiation. On the medical wards, however, readmission increased from 4.6 to 5.6 per 1,000 patient-days (p = 0.043), while mortality decreased from 7.4 to 5.0 per 1,000 patient-days (p = 0.001). CDI rates showed a non-significant declining trend after PAAF initiation. Conclusions ASPs can lead to cost-effective, sustained reductions in total antibiotic use when interventions are conducted early in the course of therapy and target all antibiotics. Shifting to such a model may help strengthen the effectiveness of ASPs in non-ICU settings. PMID:28562638
Solving large mixed linear models using preconditioned conjugate gradient iteration.
Strandén, I; Lidauer, M
1999-12-01
Continuous evaluation of dairy cattle with a random regression test-day model requires a fast solving method and algorithm. A new computing technique feasible in Jacobi and conjugate gradient based iterative methods using iteration on data is presented. In the new computing technique, the calculations in multiplication of a vector by a matrix were recorded to three steps instead of the commonly used two steps. The three-step method was implemented in a general mixed linear model program that used preconditioned conjugate gradient iteration. Performance of this program in comparison to other general solving programs was assessed via estimation of breeding values using univariate, multivariate, and random regression test-day models. Central processing unit time per iteration with the new three-step technique was, at best, one-third that needed with the old technique. Performance was best with the test-day model, which was the largest and most complex model used. The new program did well in comparison to other general software. Programs keeping the mixed model equations in random access memory required at least 20 and 435% more time to solve the univariate and multivariate animal models, respectively. Computations of the second best iteration on data took approximately three and five times longer for the animal and test-day models, respectively, than did the new program. Good performance was due to fast computing time per iteration and quick convergence to the final solutions. Use of preconditioned conjugate gradient based methods in solving large breeding value problems is supported by our findings.
The Effects of a 6-Week Plyometric Training Program on Agility
Miller, Michael G.; Herniman, Jeremy J.; Ricard, Mark D.; Cheatham, Christopher C.; Michael, Timothy J.
2006-01-01
The purpose of the study was to determine if six weeks of plyometric training can improve an athlete's agility. Subjects were divided into two groups, a plyometric training and a control group. The plyometric training group performed in a six week plyometric training program and the control group did not perform any plyometric training techniques. All subjects participated in two agility tests: T-test and Illinois Agility Test, and a force plate test for ground reaction times both pre and post testing. Univariate ANCOVAs were conducted to analyze the change scores (post - pre) in the independent variables by group (training or control) with pre scores as covariates. The Univariate ANCOVA revealed a significant group effect F2,26 = 25.42, p=0.0000 for the T-test agility measure. For the Illinois Agility test, a significant group effect F2,26 = 27.24, p = 0.000 was also found. The plyometric training group had quicker posttest times compared to the control group for the agility tests. A significant group effect F2,26 = 7.81, p = 0.002 was found for the Force Plate test. The plyometric training group reduced time on the ground on the posttest compared to the control group. The results of this study show that plyometric training can be an effective training technique to improve an athlete's agility. Key Points Plyometric training can enhance agility of athletes. 6 weeks of plyometric training is sufficient to see agility results. Ground reaction times are decreased with plyometric training PMID:24353464
Association mining of dependency between time series
NASA Astrophysics Data System (ADS)
Hafez, Alaaeldin
2001-03-01
Time series analysis is considered as a crucial component of strategic control over a broad variety of disciplines in business, science and engineering. Time series data is a sequence of observations collected over intervals of time. Each time series describes a phenomenon as a function of time. Analysis on time series data includes discovering trends (or patterns) in a time series sequence. In the last few years, data mining has emerged and been recognized as a new technology for data analysis. Data Mining is the process of discovering potentially valuable patterns, associations, trends, sequences and dependencies in data. Data mining techniques can discover information that many traditional business analysis and statistical techniques fail to deliver. In this paper, we adapt and innovate data mining techniques to analyze time series data. By using data mining techniques, maximal frequent patterns are discovered and used in predicting future sequences or trends, where trends describe the behavior of a sequence. In order to include different types of time series (e.g. irregular and non- systematic), we consider past frequent patterns of the same time sequences (local patterns) and of other dependent time sequences (global patterns). We use the word 'dependent' instead of the word 'similar' for emphasis on real life time series where two time series sequences could be completely different (in values, shapes, etc.), but they still react to the same conditions in a dependent way. In this paper, we propose the Dependence Mining Technique that could be used in predicting time series sequences. The proposed technique consists of three phases: (a) for all time series sequences, generate their trend sequences, (b) discover maximal frequent trend patterns, generate pattern vectors (to keep information of frequent trend patterns), use trend pattern vectors to predict future time series sequences.
Cohort study comparing prostate photovaporisation with XPS 180W and HPS 120W laser.
López, B; Capitán, C; Hernández, V; de la Peña, E; Jiménez-Valladolid, I; Guijarro, A; Pérez-Fernández, E; Llorente, C
2016-01-01
Prostate photovaporisation with Greenlight laser for the surgical treatment of benign prostate hyperplasia has rapidly evolve to the new XPS 180W. We have previously demonstrated the safety and efficacy of the HPS 120W. The aim of this study was to assess the functional and safety results, with a year of follow-up, of photovaporisation using the XPS 180W laser compared with its predecessor. A cohort study was conducted with a series of 191 consecutive patients who underwent photovaporisation between 1/2008 and 5/2013. The inclusion criteria were an international prostate symptom score (IPSS) >15 after medical failure, a prostate volume <80 cm(3) and a maximum flow <15 mL/s. We assessed preoperative and intraoperative variables (energy used, laser time and total surgical time), complications, catheter hours, length of stay and functional results (maximum flow, IPSS, prostate-specific antigen and prostate volume) at 3, 6 and 12 months. We analysed the homogeneity in preoperative characteristics of the 2 groups through univariate analysis techniques. The postoperative functional results were assessed through an analysis of variance of repeated measures with mixed models. A total of 109 (57.1%) procedures were performed using HPS 120W, and 82 (42.9%) were performed using XPS. There were no differences between the preoperative characteristics. We observed significant differences both in the surgical time and effective laser time in favour of the XPS system. This advantage was 11% (48 ± 15.7 vs. 53.8 ± 16.2, p<.05) and 9% (32.8 ± 11.7 vs. 36 ± 11.6, p<.05), respectively. There were no statistically significant differences in the rest of the analysed parameters. The technical improvements in the XPS 180W system help reduce surgical time, maintaining the safety and efficacy profile offered by the HPS 120W system, with completely superimposable results at 1 year of follow-up. Copyright © 2015 AEU. Publicado por Elsevier España, S.L.U. All rights reserved.
Mousavizadeh, A; Dastoorpoor, M; Naimi, E; Dohrabpour, K
2018-01-01
This study was designed and implemented to assess the current situation and to estimate the time trend of multiple sclerosis (MS), as well as to explain potential factors associated with such a trend. This longitudinal study was carried out based on analysis of the data from the monitoring and treatment surveillance system for 421 patients with MS in Kohgiluyeh and Boyer-Ahmad Province, Iran, from 1990 to 2015. To this end, curve estimation approach was used to investigate the changes in prevalence and incidence of the disease, and univariate time series model analysis was applied in order to estimate the disease incidence in the next 10 years. The mean and standard deviation of age were 29.78 and 8.5 years at the time of diagnosis, and the mean and 95% confidence interval of age were 29.18 (28.86-30.77) and 29.68 (28.06-31.30) at the time of diagnosis for women and men, respectively. The sex ratio (males to females) was estimated as 3.3, and the prevalence of the disease was estimated as 60.14 in 100,000 people. The diagram of the 35-year trend of the disease indicated three distinct patterns with a tendency to increase in recent years. The prevalence and incidence trend of the disease in the study population is consistent with regional and global changes. Climatic and environmental factors such as extreme weather changes, dust particles, expansion of the application of new industrial materials, and regional wars with potential use of banned weapons are among the issues that may, in part, be able to justify the global and regional changes of the disease. Predictive models indicate a growing trend of the disease, highlighting the need for more regular monitoring of the disease trend in upcoming years. Copyright © 2017 The Royal Society for Public Health. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Hermosilla, Txomin; Wulder, Michael A.; White, Joanne C.; Coops, Nicholas C.; Hobart, Geordie W.
2017-12-01
The use of time series satellite data allows for the temporally dense, systematic, transparent, and synoptic capture of land dynamics over time. Subsequent to the opening of the Landsat archive, several time series approaches for characterizing landscape change have been developed, often representing a particular analytical time window. The information richness and widespread utility of these time series data have created a need to maintain the currency of time series information via the addition of new data, as it becomes available. When an existing time series is temporally extended, it is critical that previously generated change information remains consistent, thereby not altering reported change statistics or science outcomes based on that change information. In this research, we investigate the impacts and implications of adding additional years to an existing 29-year annual Landsat time series for forest change. To do so, we undertook a spatially explicit comparison of the 29 overlapping years of a time series representing 1984-2012, with a time series representing 1984-2016. Surface reflectance values, and presence, year, and type of change were compared. We found that the addition of years to extend the time series had minimal effect on the annual surface reflectance composites, with slight band-specific differences (r ≥ 0.1) in the final years of the original time series being updated. The area of stand replacing disturbances and determination of change year are virtually unchanged for the overlapping period between the two time-series products. Over the overlapping temporal period (1984-2012), the total area of change differs by 0.53%, equating to an annual difference in change area of 0.019%. Overall, the spatial and temporal agreement of the changes detected by both time series was 96%. Further, our findings suggest that the entire pre-existing historic time series does not need to be re-processed during the update process. Critically, given the time series change detection and update approach followed here, science outcomes or reports representing one temporal epoch can be considered stable and will not be altered when a time series is updated with newly available data.
Highly comparative time-series analysis: the empirical structure of time series and their methods.
Fulcher, Ben D; Little, Max A; Jones, Nick S
2013-06-06
The process of collecting and organizing sets of observations represents a common theme throughout the history of science. However, despite the ubiquity of scientists measuring, recording and analysing the dynamics of different processes, an extensive organization of scientific time-series data and analysis methods has never been performed. Addressing this, annotated collections of over 35 000 real-world and model-generated time series, and over 9000 time-series analysis algorithms are analysed in this work. We introduce reduced representations of both time series, in terms of their properties measured by diverse scientific methods, and of time-series analysis methods, in terms of their behaviour on empirical time series, and use them to organize these interdisciplinary resources. This new approach to comparing across diverse scientific data and methods allows us to organize time-series datasets automatically according to their properties, retrieve alternatives to particular analysis methods developed in other scientific disciplines and automate the selection of useful methods for time-series classification and regression tasks. The broad scientific utility of these tools is demonstrated on datasets of electroencephalograms, self-affine time series, heartbeat intervals, speech signals and others, in each case contributing novel analysis techniques to the existing literature. Highly comparative techniques that compare across an interdisciplinary literature can thus be used to guide more focused research in time-series analysis for applications across the scientific disciplines.
Highly comparative time-series analysis: the empirical structure of time series and their methods
Fulcher, Ben D.; Little, Max A.; Jones, Nick S.
2013-01-01
The process of collecting and organizing sets of observations represents a common theme throughout the history of science. However, despite the ubiquity of scientists measuring, recording and analysing the dynamics of different processes, an extensive organization of scientific time-series data and analysis methods has never been performed. Addressing this, annotated collections of over 35 000 real-world and model-generated time series, and over 9000 time-series analysis algorithms are analysed in this work. We introduce reduced representations of both time series, in terms of their properties measured by diverse scientific methods, and of time-series analysis methods, in terms of their behaviour on empirical time series, and use them to organize these interdisciplinary resources. This new approach to comparing across diverse scientific data and methods allows us to organize time-series datasets automatically according to their properties, retrieve alternatives to particular analysis methods developed in other scientific disciplines and automate the selection of useful methods for time-series classification and regression tasks. The broad scientific utility of these tools is demonstrated on datasets of electroencephalograms, self-affine time series, heartbeat intervals, speech signals and others, in each case contributing novel analysis techniques to the existing literature. Highly comparative techniques that compare across an interdisciplinary literature can thus be used to guide more focused research in time-series analysis for applications across the scientific disciplines. PMID:23554344
Online probabilistic learning with an ensemble of forecasts
NASA Astrophysics Data System (ADS)
Thorey, Jean; Mallet, Vivien; Chaussin, Christophe
2016-04-01
Our objective is to produce a calibrated weighted ensemble to forecast a univariate time series. In addition to a meteorological ensemble of forecasts, we rely on observations or analyses of the target variable. The celebrated Continuous Ranked Probability Score (CRPS) is used to evaluate the probabilistic forecasts. However applying the CRPS on weighted empirical distribution functions (deriving from the weighted ensemble) may introduce a bias because of which minimizing the CRPS does not produce the optimal weights. Thus we propose an unbiased version of the CRPS which relies on clusters of members and is strictly proper. We adapt online learning methods for the minimization of the CRPS. These methods generate the weights associated to the members in the forecasted empirical distribution function. The weights are updated before each forecast step using only past observations and forecasts. Our learning algorithms provide the theoretical guarantee that, in the long run, the CRPS of the weighted forecasts is at least as good as the CRPS of any weighted ensemble with weights constant in time. In particular, the performance of our forecast is better than that of any subset ensemble with uniform weights. A noteworthy advantage of our algorithm is that it does not require any assumption on the distributions of the observations and forecasts, both for the application and for the theoretical guarantee to hold. As application example on meteorological forecasts for photovoltaic production integration, we show that our algorithm generates a calibrated probabilistic forecast, with significant performance improvements on probabilistic diagnostic tools (the CRPS, the reliability diagram and the rank histogram).
Alvarez, Hector; Corvalan, Alejandro; Roa, Juan C; Argani, Pedram; Murillo, Francisco; Edwards, Jennifer; Beaty, Robert; Feldmann, Georg; Hong, Seung-Mo; Mullendore, Michael; Roa, Ivan; Ibañez, Luis; Pimentel, Fernando; Diaz, Alfonso; Riggins, Gregory J; Maitra, Anirban
2008-05-01
Gallbladder cancer (GBC) is an uncommon neoplasm in the United States, but one with high mortality rates. This malignancy remains largely understudied at the molecular level such that few targeted therapies or predictive biomarkers exist. We built the first series of serial analysis of gene expression (SAGE) libraries from GBC and nonneoplastic gallbladder mucosa, composed of 21-bp long-SAGE tags. SAGE libraries were generated from three stage-matched GBC patients (representing Hispanic/Latino, Native American, and Caucasian ethnicities, respectively) and one histologically alithiasic gallbladder. Real-time quantitative PCR was done on microdissected epithelium from five matched GBC and corresponding nonneoplastic gallbladder mucosa. Immunohistochemical analysis was done on a panel of 182 archival GBC in high-throughput tissue microarray format. SAGE tags corresponding to connective tissue growth factor (CTGF) transcripts were identified as differentially overexpressed in all pairwise comparisons of GBC (P < 0.001). Real-time quantitative PCR confirmed significant overexpression of CTGF transcripts in microdissected primary GBC (P < 0.05), but not in metastatic GBC, compared with nonneoplastic gallbladder epithelium. By immunohistochemistry, 66 of 182 (36%) GBC had high CTGF antigen labeling, which was significantly associated with better survival on univariate analysis (P = 0.0069, log-rank test). An unbiased analysis of the GBC transcriptome by SAGE has identified CTGF expression as a predictive biomarker of favorable prognosis in this malignancy. The SAGE libraries from GBC and nonneoplastic gallbladder mucosa are publicly available at the Cancer Genome Anatomy Project web site and should facilitate much needed research into this lethal neoplasm.
ERIC Educational Resources Information Center
Ngan, Chun-Kit
2013-01-01
Making decisions over multivariate time series is an important topic which has gained significant interest in the past decade. A time series is a sequence of data points which are measured and ordered over uniform time intervals. A multivariate time series is a set of multiple, related time series in a particular domain in which domain experts…
Analysis of risk factors for central venous port failure in cancer patients
Hsieh, Ching-Chuan; Weng, Hsu-Huei; Huang, Wen-Shih; Wang, Wen-Ke; Kao, Chiung-Lun; Lu, Ming-Shian; Wang, Chia-Siu
2009-01-01
AIM: To analyze the risk factors for central port failure in cancer patients administered chemotherapy, using univariate and multivariate analyses. METHODS: A total of 1348 totally implantable venous access devices (TIVADs) were implanted into 1280 cancer patients in this cohort study. A Cox proportional hazard model was applied to analyze risk factors for failure of TIVADs. Log-rank test was used to compare actuarial survival rates. Infection, thrombosis, and surgical complication rates (χ2 test or Fisher’s exact test) were compared in relation to the risk factors. RESULTS: Increasing age, male gender and open-ended catheter use were significant risk factors reducing survival of TIVADs as determined by univariate and multivariate analyses. Hematogenous malignancy decreased the survival time of TIVADs; this reduction was not statistically significant by univariate analysis [hazard ratio (HR) = 1.336, 95% CI: 0.966-1.849, P = 0.080)]. However, it became a significant risk factor by multivariate analysis (HR = 1.499, 95% CI: 1.079-2.083, P = 0.016) when correlated with variables of age, sex and catheter type. Close-ended (Groshong) catheters had a lower thrombosis rate than open-ended catheters (2.5% vs 5%, P = 0.015). Hematogenous malignancy had higher infection rates than solid malignancy (10.5% vs 2.5%, P < 0.001). CONCLUSION: Increasing age, male gender, open-ended catheters and hematogenous malignancy were risk factors for TIVAD failure. Close-ended catheters had lower thrombosis rates and hematogenous malignancy had higher infection rates. PMID:19787834
A Review of Subsequence Time Series Clustering
Teh, Ying Wah
2014-01-01
Clustering of subsequence time series remains an open issue in time series clustering. Subsequence time series clustering is used in different fields, such as e-commerce, outlier detection, speech recognition, biological systems, DNA recognition, and text mining. One of the useful fields in the domain of subsequence time series clustering is pattern recognition. To improve this field, a sequence of time series data is used. This paper reviews some definitions and backgrounds related to subsequence time series clustering. The categorization of the literature reviews is divided into three groups: preproof, interproof, and postproof period. Moreover, various state-of-the-art approaches in performing subsequence time series clustering are discussed under each of the following categories. The strengths and weaknesses of the employed methods are evaluated as potential issues for future studies. PMID:25140332
A review of subsequence time series clustering.
Zolhavarieh, Seyedjamal; Aghabozorgi, Saeed; Teh, Ying Wah
2014-01-01
Clustering of subsequence time series remains an open issue in time series clustering. Subsequence time series clustering is used in different fields, such as e-commerce, outlier detection, speech recognition, biological systems, DNA recognition, and text mining. One of the useful fields in the domain of subsequence time series clustering is pattern recognition. To improve this field, a sequence of time series data is used. This paper reviews some definitions and backgrounds related to subsequence time series clustering. The categorization of the literature reviews is divided into three groups: preproof, interproof, and postproof period. Moreover, various state-of-the-art approaches in performing subsequence time series clustering are discussed under each of the following categories. The strengths and weaknesses of the employed methods are evaluated as potential issues for future studies.
Multiscale structure of time series revealed by the monotony spectrum.
Vamoş, Călin
2017-03-01
Observation of complex systems produces time series with specific dynamics at different time scales. The majority of the existing numerical methods for multiscale analysis first decompose the time series into several simpler components and the multiscale structure is given by the properties of their components. We present a numerical method which describes the multiscale structure of arbitrary time series without decomposing them. It is based on the monotony spectrum defined as the variation of the mean amplitude of the monotonic segments with respect to the mean local time scale during successive averagings of the time series, the local time scales being the durations of the monotonic segments. The maxima of the monotony spectrum indicate the time scales which dominate the variations of the time series. We show that the monotony spectrum can correctly analyze a diversity of artificial time series and can discriminate the existence of deterministic variations at large time scales from the random fluctuations. As an application we analyze the multifractal structure of some hydrological time series.
Wang, Yong; Yao, Xiaomei; Parthasarathy, Ranganathan
2008-01-01
Fourier transform infrared (FTIR) chemical imaging can be used to investigate molecular chemical features of the adhesive/dentin interfaces. However, the information is not straightforward, and is not easily extracted. The objective of this study was to use multivariate analysis methods, principal component analysis and fuzzy c-means clustering, to analyze spectral data in comparison with univariate analysis. The spectral imaging data collected from both the adhesive/healthy dentin and adhesive/caries-affected dentin specimens were used and compared. The univariate statistical methods such as mapping of intensities of specific functional group do not always accurately identify functional group locations and concentrations due to more or less band overlapping in adhesive and dentin. Apart from the ease with which information can be extracted, multivariate methods highlight subtle and often important changes in the spectra that are difficult to observe using univariate methods. The results showed that the multivariate methods gave more satisfactory, interpretable results than univariate methods and were conclusive in showing that they can discriminate and classify differences between healthy dentin and caries-affected dentin within the interfacial regions. It is demonstrated that the multivariate FTIR imaging approaches can be used in the rapid characterization of heterogeneous, complex structure. PMID:18980198
NASA Astrophysics Data System (ADS)
Sella, Lisa; Vivaldo, Gianna; Ghil, Michael; Groth, Andreas
2010-05-01
The present work applies several advanced spectral methods (Ghil et al., Rev. Geophys., 2002) to the analysis of macroeconomic fluctuations in Italy, The Netherlands, and the United Kingdom. These methods provide valuable time-and-frequency-domain tools that complement traditional time-domain analysis, and are thus fairly well known by now in the geosciences and life sciences, but not yet widespread in quantitative economics. In particular, they enable the identification and characterization of nonlinear trends and dominant cycles --- including low-frequency and seasonal components --- that characterize the behavior of each time series. We explore five fundamental indicators of the real (i.e., non-monetary), aggregate economy --- namely gross domestic product (GDP), consumption, fixed investments, exports and imports --- in a univariate as well as multivariate setting. A single-channel analysis by means of three independent spectral methods --- singular spectrum analysis (SSA), the multi-taper method (MTM), and the maximum-entropy method (MEM) --- reveals very similar near-annual cycles, as well as several longer periodicities, in the macroeconomic indicators of all the countries analyzed. Since each indicator represents different features of an economic system, we combine them to infer if common oscillatory modes are present, either among different indicators within the same country or among the same indicators across different countries. Multichannel-SSA (M-SSA) reinforces the previous results, and shows that the common modes agree in character with solutions of a non-equilibrium dynamic model (NEDyM) that produces endogenous business cycles (Hallegatte et al., JEBO, 2008). The presence of these modes in NEDyM results from adjustment delays and other nonequilibrium effects that were added to a neoclassical Solow (Q. J. Econ., 1956) growth model. Their confirmation by the present analysis has important consequences for the net impact of natural disasters on the economy of a country: Hallegatte and Ghil (Ecol. Econ., 2008) have shown that the presence of business cycles modifies substantially this impact with respect to their impact on an economy in or near equilibrium. The present work concludes with a study of the synchronization of economic fluctuations, which follows a similar study of macroeconomic indicators for the United States, presented in a nearby poster. Since business cycles are not country-specific phenomena, but show common characteristics across countries, our aim is to uncover hidden global behavior across the European economies (cf. Mazzi and Savio, Macmillan, 2006).
Reliability Stress-Strength Models for Dependent Observations with Applications in Clinical Trials
NASA Technical Reports Server (NTRS)
Kushary, Debashis; Kulkarni, Pandurang M.
1995-01-01
We consider the applications of stress-strength models in studies involving clinical trials. When studying the effects and side effects of certain procedures (treatments), it is often the case that observations are correlated due to subject effect, repeated measurements and observing many characteristics simultaneously. We develop maximum likelihood estimator (MLE) and uniform minimum variance unbiased estimator (UMVUE) of the reliability which in clinical trial studies could be considered as the chances of increased side effects due to a particular procedure compared to another. The results developed apply to both univariate and multivariate situations. Also, for the univariate situations we develop simple to use lower confidence bounds for the reliability. Further, we consider the cases when both stress and strength constitute time dependent processes. We define the future reliability and obtain methods of constructing lower confidence bounds for this reliability. Finally, we conduct simulation studies to evaluate all the procedures developed and also to compare the MLE and the UMVUE.
Cervantes-Sanchez, Fernando; Hernandez-Aguirre, Arturo; Solorio-Meza, Sergio; Ornelas-Rodriguez, Manuel; Torres-Cisneros, Miguel
2016-01-01
This paper presents a novel method for improving the training step of the single-scale Gabor filters by using the Boltzmann univariate marginal distribution algorithm (BUMDA) in X-ray angiograms. Since the single-scale Gabor filters (SSG) are governed by three parameters, the optimal selection of the SSG parameters is highly desirable in order to maximize the detection performance of coronary arteries while reducing the computational time. To obtain the best set of parameters for the SSG, the area (A z) under the receiver operating characteristic curve is used as fitness function. Moreover, to classify vessel and nonvessel pixels from the Gabor filter response, the interclass variance thresholding method has been adopted. The experimental results using the proposed method obtained the highest detection rate with A z = 0.9502 over a training set of 40 images and A z = 0.9583 with a test set of 40 images. In addition, the experimental results of vessel segmentation provided an accuracy of 0.944 with the test set of angiograms. PMID:27738422
[Prevalence and determinants of exclusive breastfeeding in the city of Serrana, São Paulo, Brazil].
Queluz, Mariângela Carletti; Pereira, Maria José Bistafa; dos Santos, Claudia Benedita; Leite, Adriana Moraes; Ricco, Rubens Garcia
2012-06-01
The objective of this cross-sectional and quantitative study was to identify the prevalence and determinants of exclusive breastfeeding among infants less than six months of age in the city of Serrana, Sao Paulo, Brazil in 2009. A validated semi-structured questionnaire was administered to the guardians of the children less than six months of age who attended the second phase of a Brazilian vaccination campaign against polio. Univariate and multivariate analysis presented in odds ratios and confidence intervals was accomplished. Of the total of 275 infant participants, only 29.8% were exclusively breastfed. Univariate analysis revealed that mothers who work outside the home without maternity leave, mothers who did not work outside the home, adolescent mothers, and the use of pacifiers have a greater chance of interrupting exclusive breastfeeding. In the multivariate analysis, mothers who work outside the home without maternity leave are three times more likely to wean their children early. Results provide suggestions for the redirection and planning of interventions targeting breastfeeding.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-02-07
... Series, adjusted option series and any options series until the time to expiration for such series is... time to expiration for such series is less than nine months be treated differently. Specifically, under... series until the time to expiration for such series is less than nine months. Accordingly, the...
Marwaha, Puneeta; Sunkaria, Ramesh Kumar
2016-09-01
The sample entropy (SampEn) has been widely used to quantify the complexity of RR-interval time series. It is a fact that higher complexity, and hence, entropy is associated with the RR-interval time series of healthy subjects. But, SampEn suffers from the disadvantage that it assigns higher entropy to the randomized surrogate time series as well as to certain pathological time series, which is a misleading observation. This wrong estimation of the complexity of a time series may be due to the fact that the existing SampEn technique updates the threshold value as a function of long-term standard deviation (SD) of a time series. However, time series of certain pathologies exhibits substantial variability in beat-to-beat fluctuations. So the SD of the first order difference (short term SD) of the time series should be considered while updating threshold value, to account for period-to-period variations inherited in a time series. In the present work, improved sample entropy (I-SampEn), a new methodology has been proposed in which threshold value is updated by considering the period-to-period variations of a time series. The I-SampEn technique results in assigning higher entropy value to age-matched healthy subjects than patients suffering atrial fibrillation (AF) and diabetes mellitus (DM). Our results are in agreement with the theory of reduction in complexity of RR-interval time series in patients suffering from chronic cardiovascular and non-cardiovascular diseases.
Maisa, Anna; Brockmann, Ansgar; Renken, Frank; Lück, Christian; Pleischl, Stefan; Exner, Martin; Daniels-Haardt, Inka; Jurke, Annette
2015-01-01
Between 1 August and 6 September 2013, an outbreak of Legionnaires' disease (LD) with 159 suspected cases occurred in Warstein, North Rhine-Westphalia, Germany. The outbreak consisted of 78 laboratory-confirmed cases of LD, including one fatality, with a case fatality rate of 1%. Legionella pneumophila, serogroup 1, subtype Knoxville, sequence type 345, was identified as the epidemic strain. A case-control study was conducted to identify possible sources of infection. In univariable analysis, cases were almost five times more likely to smoke than controls (odds ratio (OR): 4.81; 95% confidence interval (CI): 2.33-9.93; p < 0.0001). Furthermore, cases were twice as likely to live within a 3 km distance from one identified infection source as controls (OR: 2.14; 95% CI: 1.09-4.20; p < 0.027). This is the largest outbreak of LD in Germany to date. Due to a series of uncommon events, this outbreak was most likely caused by multiple sources involving industrial cooling towers. Quick epidemiological assessment, source tracing and shutting down of potential sources as well as rapid laboratory testing and early treatment are necessary to reduce morbidity and mortality. Maintenance of cooling towers must be carried out according to specification to prevent similar LD outbreaks in the future.
Bueno, Irene; Redig, Patrick T; Rendahl, Aaron K
2015-11-15
To evaluate the outcome of the application of an external skeletal fixator intramedullary pin tie-in (TIF) to tibiotarsal fractures in raptors. Retrospective case series. Thirty-four raptors with 37 tibiotarsal fractures. Medical records and radiographs for raptors with tibiotarsal fractures that were treated at The Raptor Center at the University of Minnesota between 1995 and 2011 were reviewed. Descriptive statistics were generated and univariate logistic regression analyses were used to assess whether age, sex, body weight, location and nature of the fracture, and type of surgical reduction were significantly associated with whether the fracture healed following surgical reduction and TIF application. 31 of 37 (84%) tibiotarsal fractures successfully healed following surgical reduction and TIF application. The mean healing time was 38 days (range, 15 to 70 days). None of the variables assessed were significantly associated with whether the tibiotarsal fracture healed. Twenty of the 34 (59%) raptors were eventually rehabilitated and released. Results indicated that most tibiotarsal fractures were successfully managed by surgical reduction and stabilization with a TIF. However, other comorbidities (eg, systemic infections and visual deficits) negatively affected the rehabilitation of raptors and sometimes resulted in euthanasia despite the fact that the tibiotarsal fracture had healed, and those comorbidities, along with the variables evaluated (eg, age, sex, and nature of the fracture), should be used as triage criteria and prognostic indicators.
Hubble, Michael W; Richards, Michael E; Wilfong, Denise A
2008-01-01
To estimate the cost-effectiveness of continuous positive airway pressure (CPAP) in managing prehospital acute pulmonary edema in an urban EMS system. Using estimates from published reports on prehospital and emergency department CPAP, a cost-effectiveness model of implementing CPAP in a typical urban EMS system was derived from the societal perspective as well as the perspective of the implementing EMS system. To assess the robustness of the model, a series of univariate and multivariate sensitivity analyses was performed on the input variables. The cost of consumables, equipment, and training yielded a total cost of $89 per CPAP application. The theoretical system would be expected to use CPAP 4 times per 1000 EMS patients and is expected to save 0.75 additional lives per 1000 EMS patients at a cost of $490 per life saved. CPAP is also expected to result in approximately one less intubation per 6 CPAP applications and reduce hospitalization costs by $4075 per year for each CPAP application. Through sensitivity analyses the model was verified to be robust across a wide range of input variable assumptions. Previous studies have demonstrated the clinical effectiveness of CPAP in the management of acute pulmonary edema. Through a theoretical analysis which modeled the costs and clinical benefits of implementing CPAP in an urban EMS system, prehospital CPAP appears to be a cost-effective treatment.
Climate variability, weather and enteric disease incidence in New Zealand: time series analysis.
Lal, Aparna; Ikeda, Takayoshi; French, Nigel; Baker, Michael G; Hales, Simon
2013-01-01
Evaluating the influence of climate variability on enteric disease incidence may improve our ability to predict how climate change may affect these diseases. To examine the associations between regional climate variability and enteric disease incidence in New Zealand. Associations between monthly climate and enteric diseases (campylobacteriosis, salmonellosis, cryptosporidiosis, giardiasis) were investigated using Seasonal Auto Regressive Integrated Moving Average (SARIMA) models. No climatic factors were significantly associated with campylobacteriosis and giardiasis, with similar predictive power for univariate and multivariate models. Cryptosporidiosis was positively associated with average temperature of the previous month (β = 0.130, SE = 0.060, p <0.01) and inversely related to the Southern Oscillation Index (SOI) two months previously (β = -0.008, SE = 0.004, p <0.05). By contrast, salmonellosis was positively associated with temperature (β = 0.110, SE = 0.020, p<0.001) of the current month and SOI of the current (β = 0.005, SE = 0.002, p<0.050) and previous month (β = 0.005, SE = 0.002, p<0.05). Forecasting accuracy of the multivariate models for cryptosporidiosis and salmonellosis were significantly higher. Although spatial heterogeneity in the observed patterns could not be assessed, these results suggest that temporally lagged relationships between climate variables and national communicable disease incidence data can contribute to disease prediction models and early warning systems.
Modeling jointly low, moderate, and heavy rainfall intensities without a threshold selection
NASA Astrophysics Data System (ADS)
Naveau, Philippe; Huser, Raphael; Ribereau, Pierre; Hannart, Alexis
2016-04-01
In statistics, extreme events are often defined as excesses above a given large threshold. This definition allows hydrologists and flood planners to apply Extreme-Value Theory (EVT) to their time series of interest. Even in the stationary univariate context, this approach has at least two main drawbacks. First, working with excesses implies that a lot of observations (those below the chosen threshold) are completely disregarded. The range of precipitation is artificially shopped down into two pieces, namely large intensities and the rest, which necessarily imposes different statistical models for each piece. Second, this strategy raises a nontrivial and very practical difficultly: how to choose the optimal threshold which correctly discriminates between low and heavy rainfall intensities. To address these issues, we propose a statistical model in which EVT results apply not only to heavy, but also to low precipitation amounts (zeros excluded). Our model is in compliance with EVT on both ends of the spectrum and allows a smooth transition between the two tails, while keeping a low number of parameters. In terms of inference, we have implemented and tested two classical methods of estimation: likelihood maximization and probability weighed moments. Last but not least, there is no need to choose a threshold to define low and high excesses. The performance and flexibility of this approach are illustrated on simulated and hourly precipitation recorded in Lyon, France.
Perceived somatotype as related to self-concept in Nigerian adolescent high school students.
Olu, S S
1991-12-01
Some 288 Nigerian high school male adolescents in two different age brackets were administered the perceived somatotype scale and the Tennessee self-concept scale to determine the trend of self-concept ratings among subjects differentiated according to perceived somatotype-self, ideal and self-ideal discrepancy. Series of one-way ANOVA revealed significant differences in global self-concept among the perceived somatotype-self (PSS), the perceived somatotype-ideal (PSI) and perceived somatotype discrepancy (PSD) groups. Univariate analyses showed that the PSS groups differ significantly in all dimensions of self-concept except moral-ethical dimension. The PSI groups differed significantly on all self-concept dimensions except the behavioural and the moral-ethical self subscales. This study, which also revealed that perceived-ideal somatotype and self-ideal somatotype discrepancy differentiated markedly among measures of self-concept, is consistent with those obtained from other parts of the world.
Cluster-based exposure variation analysis
2013-01-01
Background Static posture, repetitive movements and lack of physical variation are known risk factors for work-related musculoskeletal disorders, and thus needs to be properly assessed in occupational studies. The aims of this study were (i) to investigate the effectiveness of a conventional exposure variation analysis (EVA) in discriminating exposure time lines and (ii) to compare it with a new cluster-based method for analysis of exposure variation. Methods For this purpose, we simulated a repeated cyclic exposure varying within each cycle between “low” and “high” exposure levels in a “near” or “far” range, and with “low” or “high” velocities (exposure change rates). The duration of each cycle was also manipulated by selecting a “small” or “large” standard deviation of the cycle time. Theses parameters reflected three dimensions of exposure variation, i.e. range, frequency and temporal similarity. Each simulation trace included two realizations of 100 concatenated cycles with either low (ρ = 0.1), medium (ρ = 0.5) or high (ρ = 0.9) correlation between the realizations. These traces were analyzed by conventional EVA, and a novel cluster-based EVA (C-EVA). Principal component analysis (PCA) was applied on the marginal distributions of 1) the EVA of each of the realizations (univariate approach), 2) a combination of the EVA of both realizations (multivariate approach) and 3) C-EVA. The least number of principal components describing more than 90% of variability in each case was selected and the projection of marginal distributions along the selected principal component was calculated. A linear classifier was then applied to these projections to discriminate between the simulated exposure patterns, and the accuracy of classified realizations was determined. Results C-EVA classified exposures more correctly than univariate and multivariate EVA approaches; classification accuracy was 49%, 47% and 52% for EVA (univariate and multivariate), and C-EVA, respectively (p < 0.001). All three methods performed poorly in discriminating exposure patterns differing with respect to the variability in cycle time duration. Conclusion While C-EVA had a higher accuracy than conventional EVA, both failed to detect differences in temporal similarity. The data-driven optimality of data reduction and the capability of handling multiple exposure time lines in a single analysis are the advantages of the C-EVA. PMID:23557439
Univariate normalization of bispectrum using Hölder's inequality.
Shahbazi, Forooz; Ewald, Arne; Nolte, Guido
2014-08-15
Considering that many biological systems including the brain are complex non-linear systems, suitable methods capable of detecting these non-linearities are required to study the dynamical properties of these systems. One of these tools is the third order cummulant or cross-bispectrum, which is a measure of interfrequency interactions between three signals. For convenient interpretation, interaction measures are most commonly normalized to be independent of constant scales of the signals such that its absolute values are bounded by one, with this limit reflecting perfect coupling. Although many different normalization factors for cross-bispectra were suggested in the literature these either do not lead to bounded measures or are themselves dependent on the coupling and not only on the scale of the signals. In this paper we suggest a normalization factor which is univariate, i.e., dependent only on the amplitude of each signal and not on the interactions between signals. Using a generalization of Hölder's inequality it is proven that the absolute value of this univariate bicoherence is bounded by zero and one. We compared three widely used normalizations to the univariate normalization concerning the significance of bicoherence values gained from resampling tests. Bicoherence values are calculated from real EEG data recorded in an eyes closed experiment from 10 subjects. The results show slightly more significant values for the univariate normalization but in general, the differences are very small or even vanishing in some subjects. Therefore, we conclude that the normalization factor does not play an important role in the bicoherence values with regard to statistical power, although a univariate normalization is the only normalization factor which fulfills all the required conditions of a proper normalization. Copyright © 2014 Elsevier B.V. All rights reserved.
2011-01-01
Background Comparing biological time series data across different conditions, or different specimens, is a common but still challenging task. Algorithms aligning two time series represent a valuable tool for such comparisons. While many powerful computation tools for time series alignment have been developed, they do not provide significance estimates for time shift measurements. Results Here, we present an extended version of the original DTW algorithm that allows us to determine the significance of time shift estimates in time series alignments, the DTW-Significance (DTW-S) algorithm. The DTW-S combines important properties of the original algorithm and other published time series alignment tools: DTW-S calculates the optimal alignment for each time point of each gene, it uses interpolated time points for time shift estimation, and it does not require alignment of the time-series end points. As a new feature, we implement a simulation procedure based on parameters estimated from real time series data, on a series-by-series basis, allowing us to determine the false positive rate (FPR) and the significance of the estimated time shift values. We assess the performance of our method using simulation data and real expression time series from two published primate brain expression datasets. Our results show that this method can provide accurate and robust time shift estimates for each time point on a gene-by-gene basis. Using these estimates, we are able to uncover novel features of the biological processes underlying human brain development and maturation. Conclusions The DTW-S provides a convenient tool for calculating accurate and robust time shift estimates at each time point for each gene, based on time series data. The estimates can be used to uncover novel biological features of the system being studied. The DTW-S is freely available as an R package TimeShift at http://www.picb.ac.cn/Comparative/data.html. PMID:21851598
Yuan, Yuan; Chen, Yi-Ping Phoebe; Ni, Shengyu; Xu, Augix Guohua; Tang, Lin; Vingron, Martin; Somel, Mehmet; Khaitovich, Philipp
2011-08-18
Comparing biological time series data across different conditions, or different specimens, is a common but still challenging task. Algorithms aligning two time series represent a valuable tool for such comparisons. While many powerful computation tools for time series alignment have been developed, they do not provide significance estimates for time shift measurements. Here, we present an extended version of the original DTW algorithm that allows us to determine the significance of time shift estimates in time series alignments, the DTW-Significance (DTW-S) algorithm. The DTW-S combines important properties of the original algorithm and other published time series alignment tools: DTW-S calculates the optimal alignment for each time point of each gene, it uses interpolated time points for time shift estimation, and it does not require alignment of the time-series end points. As a new feature, we implement a simulation procedure based on parameters estimated from real time series data, on a series-by-series basis, allowing us to determine the false positive rate (FPR) and the significance of the estimated time shift values. We assess the performance of our method using simulation data and real expression time series from two published primate brain expression datasets. Our results show that this method can provide accurate and robust time shift estimates for each time point on a gene-by-gene basis. Using these estimates, we are able to uncover novel features of the biological processes underlying human brain development and maturation. The DTW-S provides a convenient tool for calculating accurate and robust time shift estimates at each time point for each gene, based on time series data. The estimates can be used to uncover novel biological features of the system being studied. The DTW-S is freely available as an R package TimeShift at http://www.picb.ac.cn/Comparative/data.html.
Retention of community college students in online courses
NASA Astrophysics Data System (ADS)
Krajewski, Sarah
The issue of attrition in online courses at higher learning institutions remains a high priority in the United States. A recent rapid growth of online courses at community colleges has been instigated by student demand, as they meet the time constraints many nontraditional community college students have as a result of the need to work and care for dependents. Failure in an online course can cause students to become frustrated with the college experience, financially burdened, or to even give up and leave college. Attrition could be avoided by proper guidance of who is best suited for online courses. This study examined factors related to retention (i.e., course completion) and success (i.e., receiving a C or better) in an online biology course at a community college in the Midwest by operationalizing student characteristics (age, race, gender), student skills (whether or not the student met the criteria to be placed in an AFP course), and external factors (Pell recipient, full/part time status, first term) from the persistence model developed by Rovai. Internal factors from this model were not included in this study. Both univariate analyses and multivariate logistic regression were used to analyze the variables. Results suggest that race and Pell recipient were both predictive of course completion on univariate analyses. However, multivariate analyses showed that age, race, academic load and first term were predictive of completion and Pell recipient was no longer predictive. The univariate results for the C or better showed that age, race, Pell recipient, academic load, and meeting AFP criteria were predictive of success. Multivariate analyses showed that only age, race, and Pell recipient were significant predictors of success. Both regression models explained very little (<15%) of the variability within the outcome variables of retention and success. Therefore, although significant predictors were identified for course completion and retention, there are still many factors that remain unaccounted for in both regression models. Further research into the operationalization of Rovai's model, including internal factors, to predict completion and success is necessary.
1988-10-01
meteorologists’ rule-of-thumb that climatic drift manifests itself in periods greater than 30 years. For a fractionally-differenced model with our...estimates in a univariate ARIMA (p, d, q) with I d I< 0.5 has been derived by Li and McLrjd (1986). The model used by I-Iaslett an Raftery can be viewed as...Reply to the Discussion of "Space-time Modelling with Long-mnmory cDependence: Assessing Ireland’s Wind Resource" cJohn Haslett Department of
Ciatto, Stefano; Bonardi, Rita; Lombardi, Claudio; Zappa, Marco; Gervasi, Ginetta
2002-01-01
To evaluate the sensitivity at transrectal ultrasonography (TRUS) for prostate cancer. A consecutive series of 170 prostate cancers identified by matching local cancer registry and TRUS archives at the Centro per lo Studio e la Prevenzione Oncologica of Florence. TRUS sensitivity was determined as the ratio of TRUS positive to total prostate cancers occurring at different intervals from TRUS date. Univariate and multivariate analyses of sensitivity determinants were performed. Sensitivity at 6 months, 1, 2 and 3 years after the test was 94.1% (95% CI, 90-98), 89.8% (95% CI, 85-95), 80.4% (95% CI, 74-87) and 74.1% (95% CI, 68-81%), respectively. A higher sensitivity (statistically significant) of TRUS was observed only if digital rectal examination was suspicious, whereas no association to sensitivity was observed for age, prostate-specific antigen or prostate-specific antigen density. The study provided a reliable estimate of TRUS sensitivity, particularly reliable being checked against a cancer registry: observed sensitivity was high, at least of the same magnitude of other cancer screening tests. TRUS, which is known to allow for considerable diagnostic anticipation and is more specific than prostate-specific antigen, might still be considered for its contribution to a screening approach.
Zhou, Renjie; Yang, Chen; Wan, Jian; Zhang, Wei; Guan, Bo; Xiong, Naixue
2017-01-01
Measurement of time series complexity and predictability is sometimes the cornerstone for proposing solutions to topology and congestion control problems in sensor networks. As a method of measuring time series complexity and predictability, multiscale entropy (MSE) has been widely applied in many fields. However, sample entropy, which is the fundamental component of MSE, measures the similarity of two subsequences of a time series with either zero or one, but without in-between values, which causes sudden changes of entropy values even if the time series embraces small changes. This problem becomes especially severe when the length of time series is getting short. For solving such the problem, we propose flexible multiscale entropy (FMSE), which introduces a novel similarity function measuring the similarity of two subsequences with full-range values from zero to one, and thus increases the reliability and stability of measuring time series complexity. The proposed method is evaluated on both synthetic and real time series, including white noise, 1/f noise and real vibration signals. The evaluation results demonstrate that FMSE has a significant improvement in reliability and stability of measuring complexity of time series, especially when the length of time series is short, compared to MSE and composite multiscale entropy (CMSE). The proposed method FMSE is capable of improving the performance of time series analysis based topology and traffic congestion control techniques. PMID:28383496
Zhou, Renjie; Yang, Chen; Wan, Jian; Zhang, Wei; Guan, Bo; Xiong, Naixue
2017-04-06
Measurement of time series complexity and predictability is sometimes the cornerstone for proposing solutions to topology and congestion control problems in sensor networks. As a method of measuring time series complexity and predictability, multiscale entropy (MSE) has been widely applied in many fields. However, sample entropy, which is the fundamental component of MSE, measures the similarity of two subsequences of a time series with either zero or one, but without in-between values, which causes sudden changes of entropy values even if the time series embraces small changes. This problem becomes especially severe when the length of time series is getting short. For solving such the problem, we propose flexible multiscale entropy (FMSE), which introduces a novel similarity function measuring the similarity of two subsequences with full-range values from zero to one, and thus increases the reliability and stability of measuring time series complexity. The proposed method is evaluated on both synthetic and real time series, including white noise, 1/f noise and real vibration signals. The evaluation results demonstrate that FMSE has a significant improvement in reliability and stability of measuring complexity of time series, especially when the length of time series is short, compared to MSE and composite multiscale entropy (CMSE). The proposed method FMSE is capable of improving the performance of time series analysis based topology and traffic congestion control techniques.
NASA Astrophysics Data System (ADS)
Das Bhowmik, R.; Arumugam, S.
2015-12-01
Multivariate downscaling techniques exhibited superiority over univariate regression schemes in terms of preserving cross-correlations between multiple variables- precipitation and temperature - from GCMs. This study focuses on two aspects: (a) develop an analytical solutions on estimating biases in cross-correlations from univariate downscaling approaches and (b) quantify the uncertainty in land-surface states and fluxes due to biases in cross-correlations in downscaled climate forcings. Both these aspects are evaluated using climate forcings available from both historical climate simulations and CMIP5 hindcasts over the entire US. The analytical solution basically relates the univariate regression parameters, co-efficient of determination of regression and the co-variance ratio between GCM and downscaled values. The analytical solutions are compared with the downscaled univariate forcings by choosing the desired p-value (Type-1 error) in preserving the observed cross-correlation. . For quantifying the impacts of biases on cross-correlation on estimating streamflow and groundwater, we corrupt the downscaled climate forcings with different cross-correlation structure.
Empirical method to measure stochasticity and multifractality in nonlinear time series
NASA Astrophysics Data System (ADS)
Lin, Chih-Hao; Chang, Chia-Seng; Li, Sai-Ping
2013-12-01
An empirical algorithm is used here to study the stochastic and multifractal nature of nonlinear time series. A parameter can be defined to quantitatively measure the deviation of the time series from a Wiener process so that the stochasticity of different time series can be compared. The local volatility of the time series under study can be constructed using this algorithm, and the multifractal structure of the time series can be analyzed by using this local volatility. As an example, we employ this method to analyze financial time series from different stock markets. The result shows that while developed markets evolve very much like an Ito process, the emergent markets are far from efficient. Differences about the multifractal structures and leverage effects between developed and emergent markets are discussed. The algorithm used here can be applied in a similar fashion to study time series of other complex systems.
Multiscale Poincaré plots for visualizing the structure of heartbeat time series.
Henriques, Teresa S; Mariani, Sara; Burykin, Anton; Rodrigues, Filipa; Silva, Tiago F; Goldberger, Ary L
2016-02-09
Poincaré delay maps are widely used in the analysis of cardiac interbeat interval (RR) dynamics. To facilitate visualization of the structure of these time series, we introduce multiscale Poincaré (MSP) plots. Starting with the original RR time series, the method employs a coarse-graining procedure to create a family of time series, each of which represents the system's dynamics in a different time scale. Next, the Poincaré plots are constructed for the original and the coarse-grained time series. Finally, as an optional adjunct, color can be added to each point to represent its normalized frequency. We illustrate the MSP method on simulated Gaussian white and 1/f noise time series. The MSP plots of 1/f noise time series reveal relative conservation of the phase space area over multiple time scales, while those of white noise show a marked reduction in area. We also show how MSP plots can be used to illustrate the loss of complexity when heartbeat time series from healthy subjects are compared with those from patients with chronic (congestive) heart failure syndrome or with atrial fibrillation. This generalized multiscale approach to Poincaré plots may be useful in visualizing other types of time series.
Sun, Tao; Liu, Hongbo; Yu, Hong; Chen, C L Philip
2016-06-28
The central time series crystallizes the common patterns of the set it represents. In this paper, we propose a global constrained degree-pruning dynamic programming (g(dp)²) approach to obtain the central time series through minimizing dynamic time warping (DTW) distance between two time series. The DTW matching path theory with global constraints is proved theoretically for our degree-pruning strategy, which is helpful to reduce the time complexity and computational cost. Our approach can achieve the optimal solution between two time series. An approximate method to the central time series of multiple time series [called as m_g(dp)²] is presented based on DTW barycenter averaging and our g(dp)² approach by considering hierarchically merging strategy. As illustrated by the experimental results, our approaches provide better within-group sum of squares and robustness than other relevant algorithms.
Univariate Probability Distributions
ERIC Educational Resources Information Center
Leemis, Lawrence M.; Luckett, Daniel J.; Powell, Austin G.; Vermeer, Peter E.
2012-01-01
We describe a web-based interactive graphic that can be used as a resource in introductory classes in mathematical statistics. This interactive graphic presents 76 common univariate distributions and gives details on (a) various features of the distribution such as the functional form of the probability density function and cumulative distribution…
[Risk factors for elevated serum total bile acid in preterm infants].
Song, Yan-Ting; Wang, Yong-Qin; Zhao, Yue-Hua; Zhu, Hai-Ling; Liu, Qian; Zhang, Xiao; Gao, Yi-Wen; Zhang, Wei-Ye; Sang, Yu-Tong
2018-03-01
To study the risk factors for elevated serum total bile acid (TBA) in preterm infants. A retrospective analysis was performed for the clinical data of 216 preterm infants who were admitted to the neonatal intensive care unit. According to the presence or absence of elevated TBA (TBA >24.8 μmol/L), the preterm infants were divided into elevated TBA group with 53 infants and non-elevated TBA group with 163 infants. A univariate analysis and an unconditional multivariate logistic regression analysis were used to investigate the risk factors for elevated TBA. The univariate analysis showed that there were significant differences between the elevated TBA group and the non-elevated TBA group in gestational age at birth, birth weight, proportion of small-for-gestational-age infants, proportion of infants undergoing ventilator-assisted ventilation, fasting time, parenteral nutrition time, and incidence of neonatal respiratory failure and sepsis (P<0.05). The unconditional multivariate logistic regression analysis showed that low birth weight (OR=3.84, 95%CI: 1.53-9.64) and neonatal sepsis (OR=2.56, 95%CI: 1.01-6.47) were independent risk factors for elevated TBA in preterm infants. Low birth weight and neonatal sepsis may lead to elevated TBA in preterm infants.
NASA Astrophysics Data System (ADS)
Shimada, Yutaka; Ikeguchi, Tohru; Shigehara, Takaomi
2012-10-01
In this Letter, we propose a framework to transform a complex network to a time series. The transformation from complex networks to time series is realized by the classical multidimensional scaling. Applying the transformation method to a model proposed by Watts and Strogatz [Nature (London) 393, 440 (1998)], we show that ring lattices are transformed to periodic time series, small-world networks to noisy periodic time series, and random networks to random time series. We also show that these relationships are analytically held by using the circulant-matrix theory and the perturbation theory of linear operators. The results are generalized to several high-dimensional lattices.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-15
... Options Series, adjusted option series and any options series until the time to expiration for such series... time to expiration for such series is less than nine months be treated differently. Specifically, under... until the time to expiration for such series is less than nine months. Accordingly, the requirement to...
Incremental fuzzy C medoids clustering of time series data using dynamic time warping distance
Chen, Jingli; Wu, Shuai; Liu, Zhizhong; Chao, Hao
2018-01-01
Clustering time series data is of great significance since it could extract meaningful statistics and other characteristics. Especially in biomedical engineering, outstanding clustering algorithms for time series may help improve the health level of people. Considering data scale and time shifts of time series, in this paper, we introduce two incremental fuzzy clustering algorithms based on a Dynamic Time Warping (DTW) distance. For recruiting Single-Pass and Online patterns, our algorithms could handle large-scale time series data by splitting it into a set of chunks which are processed sequentially. Besides, our algorithms select DTW to measure distance of pair-wise time series and encourage higher clustering accuracy because DTW could determine an optimal match between any two time series by stretching or compressing segments of temporal data. Our new algorithms are compared to some existing prominent incremental fuzzy clustering algorithms on 12 benchmark time series datasets. The experimental results show that the proposed approaches could yield high quality clusters and were better than all the competitors in terms of clustering accuracy. PMID:29795600
Incremental fuzzy C medoids clustering of time series data using dynamic time warping distance.
Liu, Yongli; Chen, Jingli; Wu, Shuai; Liu, Zhizhong; Chao, Hao
2018-01-01
Clustering time series data is of great significance since it could extract meaningful statistics and other characteristics. Especially in biomedical engineering, outstanding clustering algorithms for time series may help improve the health level of people. Considering data scale and time shifts of time series, in this paper, we introduce two incremental fuzzy clustering algorithms based on a Dynamic Time Warping (DTW) distance. For recruiting Single-Pass and Online patterns, our algorithms could handle large-scale time series data by splitting it into a set of chunks which are processed sequentially. Besides, our algorithms select DTW to measure distance of pair-wise time series and encourage higher clustering accuracy because DTW could determine an optimal match between any two time series by stretching or compressing segments of temporal data. Our new algorithms are compared to some existing prominent incremental fuzzy clustering algorithms on 12 benchmark time series datasets. The experimental results show that the proposed approaches could yield high quality clusters and were better than all the competitors in terms of clustering accuracy.
Visibility Graph Based Time Series Analysis.
Stephen, Mutua; Gu, Changgui; Yang, Huijie
2015-01-01
Network based time series analysis has made considerable achievements in the recent years. By mapping mono/multivariate time series into networks, one can investigate both it's microscopic and macroscopic behaviors. However, most proposed approaches lead to the construction of static networks consequently providing limited information on evolutionary behaviors. In the present paper we propose a method called visibility graph based time series analysis, in which series segments are mapped to visibility graphs as being descriptions of the corresponding states and the successively occurring states are linked. This procedure converts a time series to a temporal network and at the same time a network of networks. Findings from empirical records for stock markets in USA (S&P500 and Nasdaq) and artificial series generated by means of fractional Gaussian motions show that the method can provide us rich information benefiting short-term and long-term predictions. Theoretically, we propose a method to investigate time series from the viewpoint of network of networks.
Wang, Jin; Sun, Xiangping; Nahavandi, Saeid; Kouzani, Abbas; Wu, Yuchuan; She, Mary
2014-11-01
Biomedical time series clustering that automatically groups a collection of time series according to their internal similarity is of importance for medical record management and inspection such as bio-signals archiving and retrieval. In this paper, a novel framework that automatically groups a set of unlabelled multichannel biomedical time series according to their internal structural similarity is proposed. Specifically, we treat a multichannel biomedical time series as a document and extract local segments from the time series as words. We extend a topic model, i.e., the Hierarchical probabilistic Latent Semantic Analysis (H-pLSA), which was originally developed for visual motion analysis to cluster a set of unlabelled multichannel time series. The H-pLSA models each channel of the multichannel time series using a local pLSA in the first layer. The topics learned in the local pLSA are then fed to a global pLSA in the second layer to discover the categories of multichannel time series. Experiments on a dataset extracted from multichannel Electrocardiography (ECG) signals demonstrate that the proposed method performs better than previous state-of-the-art approaches and is relatively robust to the variations of parameters including length of local segments and dictionary size. Although the experimental evaluation used the multichannel ECG signals in a biometric scenario, the proposed algorithm is a universal framework for multichannel biomedical time series clustering according to their structural similarity, which has many applications in biomedical time series management. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
Invariance in the recurrence of large returns and the validation of models of price dynamics
NASA Astrophysics Data System (ADS)
Chang, Lo-Bin; Geman, Stuart; Hsieh, Fushing; Hwang, Chii-Ruey
2013-08-01
Starting from a robust, nonparametric definition of large returns (“excursions”), we study the statistics of their occurrences, focusing on the recurrence process. The empirical waiting-time distribution between excursions is remarkably invariant to year, stock, and scale (return interval). This invariance is related to self-similarity of the marginal distributions of returns, but the excursion waiting-time distribution is a function of the entire return process and not just its univariate probabilities. Generalized autoregressive conditional heteroskedasticity (GARCH) models, market-time transformations based on volume or trades, and generalized (Lévy) random-walk models all fail to fit the statistical structure of excursions.
Analysis of Nonstationary Time Series for Biological Rhythms Research.
Leise, Tanya L
2017-06-01
This article is part of a Journal of Biological Rhythms series exploring analysis and statistics topics relevant to researchers in biological rhythms and sleep research. The goal is to provide an overview of the most common issues that arise in the analysis and interpretation of data in these fields. In this article on time series analysis for biological rhythms, we describe some methods for assessing the rhythmic properties of time series, including tests of whether a time series is indeed rhythmic. Because biological rhythms can exhibit significant fluctuations in their period, phase, and amplitude, their analysis may require methods appropriate for nonstationary time series, such as wavelet transforms, which can measure how these rhythmic parameters change over time. We illustrate these methods using simulated and real time series.
Griffiths, Bethany E.; Grove White, Dai; Oikonomou, Georgios
2018-01-01
Lameness is one of the most pressing issues within the dairy industry; it has severe economic implications while causing a serious impact on animal welfare. A study conducted approximately 10 years ago found the within farm lameness prevalence in the UK to be 36.8%. Our objective here is to provide an update on within farm lameness prevalence in the UK, and to provide further evidence on farm level risk factors. A convenience sample of 61 dairy farms were recruited across England and Wales from September 2015 to December 2016. A single farm visit was made and the milking herd was mobility scored, as the cows exited the milking parlor after morning, afternoon, or evening milking. Information regarding the farm and management system was then collected using a short interview with the farmer followed by collection of various subjective and objective measurements of the environment. The same, trained researcher performed all animal and facility-based measures on all visits. A series of univariable analyses were conducted to evaluate the association between various risk factors and herd lameness prevalence (logit transformed). A multivariable linear regression model was then fitted. The median number of milking cows per herd was 193, ranging from 74 to 1,519 cows. The mean within farm lameness prevalence was 31.6%, ranging from 5.8 to 65.4%. In total, 14,700 cows were mobility scored with 4,145 cows found to be lame (28.2%). A number of risk factors were associated with lameness at the univariable analysis level. Categorical risk factors retained in the final model were: resting area type, collecting yard groove spacing width, whether farms were undertaking the 60- to 100-day post calving claw trimming and the frequency of footbathing in the winter. The amount of concentrates fed in the milking parlors or out of parlor feeders was also associated with lameness prevalence. The results of this study have provided an update on the UK herd lameness prevalence and have confirmed the importance of cow comfort and footbathing frequency. The association between early lactation claw trimming and reduced lameness prevalence is, to the best of our knowledge, reported for the first time. PMID:29675419
Impact of Fractionation and Dose in a Multivariate Model for Radiation-Induced Chest Wall Pain
DOE Office of Scientific and Technical Information (OSTI.GOV)
Din, Shaun U.; Williams, Eric L.; Jackson, Andrew
Purpose: To determine the role of patient/tumor characteristics, radiation dose, and fractionation using the linear-quadratic (LQ) model to predict stereotactic body radiation therapy–induced grade ≥2 chest wall pain (CWP2) in a larger series and develop clinically useful constraints for patients treated with different fraction numbers. Methods and Materials: A total of 316 lung tumors in 295 patients were treated with stereotactic body radiation therapy in 3 to 5 fractions to 39 to 60 Gy. Absolute dose–absolute volume chest wall (CW) histograms were acquired. The raw dose-volume histograms (α/β = ∞ Gy) were converted via the LQ model to equivalent doses in 2-Gy fractions (normalizedmore » total dose, NTD) with α/β from 0 to 25 Gy in 0.1-Gy steps. The Cox proportional hazards (CPH) model was used in univariate and multivariate models to identify and assess CWP2 exposed to a given physical and NTD. Results: The median follow-up was 15.4 months, and the median time to development of CWP2 was 7.4 months. On a univariate CPH model, prescription dose, prescription dose per fraction, number of fractions, D83cc, distance of tumor to CW, and body mass index were all statistically significant for the development of CWP2. Linear-quadratic correction improved the CPH model significance over the physical dose. The best-fit α/β was 2.1 Gy, and the physical dose (α/β = ∞ Gy) was outside the upper 95% confidence limit. With α/β = 2.1 Gy, V{sub NTD99Gy} was most significant, with median V{sub NTD99Gy} = 31.5 cm{sup 3} (hazard ratio 3.87, P<.001). Conclusion: There were several predictive factors for the development of CWP2. The LQ-adjusted doses using the best-fit α/β = 2.1 Gy is a better predictor of CWP2 than the physical dose. To aid dosimetrists, we have calculated the physical dose equivalent corresponding to V{sub NTD99Gy} = 31.5 cm{sup 3} for the 3- to 5-fraction groups.« less
NASA Astrophysics Data System (ADS)
Bargaoui, Zoubeida Kebaili; Bardossy, Andràs
2015-10-01
The paper aims to develop researches on the spatial variability of heavy rainfall events estimation using spatial copula analysis. To demonstrate the methodology, short time resolution rainfall time series from Stuttgart region are analyzed. They are constituted by rainfall observations on continuous 30 min time scale recorded over a network composed by 17 raingages for the period July 1989-July 2004. The analysis is performed aggregating the observations from 30 min up to 24 h. Two parametric bivariate extreme copula models, the Husler-Reiss model and the Gumbel model are investigated. Both involve a single parameter to be estimated. Thus, model fitting is operated for every pair of stations for a giving time resolution. A rainfall threshold value representing a fixed rainfall quantile is adopted for model inference. Generalized maximum pseudo-likelihood estimation is adopted with censoring by analogy with methods of univariate estimation combining historical and paleoflood information with systematic data. Only pairs of observations greater than the threshold are assumed as systematic data. Using the estimated copula parameter, a synthetic copula field is randomly generated and helps evaluating model adequacy which is achieved using Kolmogorov Smirnov distance test. In order to assess dependence or independence in the upper tail, the extremal coefficient which characterises the tail of the joint bivariate distribution is adopted. Hence, the extremal coefficient is reported as a function of the interdistance between stations. If it is less than 1.7, stations are interpreted as dependent in the extremes. The analysis of the fitted extremal coefficients with respect to stations inter distance highlights two regimes with different dependence structures: a short spatial extent regime linked to short duration intervals (from 30 min to 6 h) with an extent of about 8 km and a large spatial extent regime related to longer rainfall intervals (from 12 h to 24 h) with an extent of 34 to 38 km.
Clarke, Clare L; Sniehotta, Falko F; Vadiveloo, Thenmalar; Argo, Ishbel S; Donnan, Peter T; McMurdo, Marion E T; Witham, Miles D
2017-08-14
Cross-sectional relationships between physical activity and health have been explored extensively, but less is known about how physical activity changes with time in older people. The aim of this study was to assess baseline predictors of how objectively measured physical activity changes with time in older people. Longitudinal cohort study using data from the Physical Activity Cohort Scotland. A sample of community-dwelling older people aged 65 and over were recruited in 2009-2011, then followed up 2-3 years later. Physical activity was measured using Stayhealthy RT3 accelerometers over 7 days. Other data collected included baseline comorbidity, health-related quality of life (SF-36), extended Theory of Planned Behaviour Questionnaire and Social Capital Module of the General Household Survey. Associations between follow-up accelerometer counts and baseline predictors were analysed using a series of linear regression models, adjusting for baseline activity levels and follow-up time. Follow up data were available for 339 of the original 584 participants. The mean age was 77 years, 185 (55%) were female and mean follow up time was 26 months. Mean activity counts fell by between 2% per year (age < =80, deprivation decile 5-10) and 12% per year (age > 80, deprivation decile 5-10) from baseline values. In univariate analysis age, sex, deprivation decile, most SF-36 domains, most measures of social connectedness, most measures from the extended Theory of Planned Behaviour, hypertension, diabetes mellitus, chronic pain and depression score were significantly associated with adjusted activity counts at follow-up. In multivariate regression age, satisfactory friend network, SF-36 physical function score, and the presence of diabetes mellitus were independent predictors of activity counts at follow up after adjustment for baseline count and duration of follow up. Health status and social connectedness, but not extended Theory of Planned Behaviour measures, independently predicted changes in physical activity in community dwelling older people.
Maximum Likelihood and Minimum Distance Applied to Univariate Mixture Distributions.
ERIC Educational Resources Information Center
Wang, Yuh-Yin Wu; Schafer, William D.
This Monte-Carlo study compared modified Newton (NW), expectation-maximization algorithm (EM), and minimum Cramer-von Mises distance (MD), used to estimate parameters of univariate mixtures of two components. Data sets were fixed at size 160 and manipulated by mean separation, variance ratio, component proportion, and non-normality. Results…
Univariate and Bivariate Loglinear Models for Discrete Test Score Distributions.
ERIC Educational Resources Information Center
Holland, Paul W.; Thayer, Dorothy T.
2000-01-01
Applied the theory of exponential families of distributions to the problem of fitting the univariate histograms and discrete bivariate frequency distributions that often arise in the analysis of test scores. Considers efficient computation of the maximum likelihood estimates of the parameters using Newton's Method and computationally efficient…
Analysis of Developmental Data: Comparison Among Alternative Methods
ERIC Educational Resources Information Center
Wilson, Ronald S.
1975-01-01
To examine the ability of the correction factor epsilon to counteract statistical bias in univariate analysis, an analysis of variance (adjusted by epsilon) and a multivariate analysis of variance were performed on the same data. The results indicated that univariate analysis is a fully protected design when used with epsilon. (JMB)
Univariate Analysis of Multivariate Outcomes in Educational Psychology.
ERIC Educational Resources Information Center
Hubble, L. M.
1984-01-01
The author examined the prevalence of multiple operational definitions of outcome constructs and an estimate of the incidence of Type I error rates when univariate procedures were applied to multiple variables in educational psychology. Multiple operational definitions of constructs were advocated and wider use of multivariate analysis was…
EMC: Air Quality Forecast Home page
archive NAM Verification Meteorology Error Time Series EMC NAM Spatial Maps Real Time Mesoscale Analysis Precipitation verification NAQFC VERIFICATION CMAQ Ozone & PM Error Time Series AOD Error Time Series HYSPLIT Smoke forecasts vs GASP satellite Dust and Smoke Error Time Series HYSPLIT WCOSS Upgrade (July
Does speed matter? The impact of operative time on outcome in laparoscopic surgery
Jackson, Timothy D.; Wannares, Jeffrey J.; Lancaster, R. Todd; Rattner, David W.
2012-01-01
Introduction Controversy exists concerning the importance of operative time on patient outcomes. It is unclear whether faster is better or haste makes waste or similarly whether slower procedures represent a safe, meticulous approach or inexperienced dawdling. The objective of the present study was to determine the effect of operative time on 30-day outcomes in laparoscopic surgery. Methods Patients who underwent laparoscopic general surgery procedures (colectomy, cholecystectomy, Nissen fundoplication, inguinal hernia, and gastric bypass) from the ACS-NSQIP 2005–2008 participant use file were identified. Exclusion criteria were defined a priori to identify same-day admission, elective procedures. Operative time was divided into deciles and summary statistics were analyzed. Univariate analyses using a Cochran-Armitage test for trend were completed. The effect of operative time on 30-day morbidity was further analyzed for each procedure type using multivariate regression controlling for case complexity and additional patient factors. Patients within the highest deciles were excluded to reduce outlier effect. Results A total of 76,748 elective general surgical patients who underwent laparoscopic procedures were analyzed. Univariate analyses of deciles of operative time demonstrated a statistically significant trend (p \\ 0.0001) toward increasing odds of complications with increasing operative time for laparoscopic colectomy (n = 10,135), cholecystectomy (n = 37,407), Nissen fundoplication (n = 4,934), and gastric bypass (n = 17,842). The trend was not found to be significant for laparoscopic inguinal hernia repair (n = 6,430; p = 0.14). Multivariate modeling revealed the effect of operative time to remain significant after controlling for additional patient factors. Conclusion Increasing operative time was associated with increased odds of complications and, therefore, it appears that speed may matter in laparoscopic surgery. These analyses are limited in their inability to adjust for all patient factors, potential confounders, and case complexities. Additional hierarchical multivariate analyses at the surgeon level would be important to examine this relationship further. PMID:21298533
Does speed matter? The impact of operative time on outcome in laparoscopic surgery.
Jackson, Timothy D; Wannares, Jeffrey J; Lancaster, R Todd; Rattner, David W; Hutter, Matthew M
2011-07-01
Controversy exists concerning the importance of operative time on patient outcomes. It is unclear whether faster is better or haste makes waste or similarly whether slower procedures represent a safe, meticulous approach or inexperienced dawdling. The objective of the present study was to determine the effect of operative time on 30-day outcomes in laparoscopic surgery. Patients who underwent laparoscopic general surgery procedures (colectomy, cholecystectomy, Nissen fundoplication, inguinal hernia, and gastric bypass) from the ACS-NSQIP 2005-2008 participant use file were identified. Exclusion criteria were defined a priori to identify same-day admission, elective procedures. Operative time was divided into deciles and summary statistics were analyzed. Univariate analyses using a Cochran-Armitage test for trend were completed. The effect of operative time on 30-day morbidity was further analyzed for each procedure type using multivariate regression controlling for case complexity and additional patient factors. Patients within the highest deciles were excluded to reduce outlier effect. A total of 76,748 elective general surgical patients who underwent laparoscopic procedures were analyzed. Univariate analyses of deciles of operative time demonstrated a statistically significant trend (p<0.0001) toward increasing odds of complications with increasing operative time for laparoscopic colectomy (n=10,135), cholecystectomy (n=37,407), Nissen fundoplication (n=4,934), and gastric bypass (n=17,842). The trend was not found to be significant for laparoscopic inguinal hernia repair (n=6,430; p=0.14). Multivariate modeling revealed the effect of operative time to remain significant after controlling for additional patient factors. Increasing operative time was associated with increased odds of complications and, therefore, it appears that speed may matter in laparoscopic surgery. These analyses are limited in their inability to adjust for all patient factors, potential confounders, and case complexities. Additional hierarchical multivariate analyses at the surgeon level would be important to examine this relationship further.
Camargo, Affonso H; Rubenstein, Jonathan N; Ershoff, Brent D; Meng, Maxwell V; Kane, Christopher J; Stoller, Marshall L
2006-01-01
Compare the outcomes between kidney morcellation and two types of open specimen extraction incisions, several covariates need to be taken into consideration that have not yet been studied. We retrospectively reviewed 153 consecutive patients who underwent laparoscopic nephrectomy at our institution, 107 who underwent specimen morcellation and 46 with intact specimen removal, either those with connected port sites with a muscle-cutting incision and those with a remote, muscle-splitting incision. Operative time, postoperative analgesia requirements, and incisional complications were evaluated using univariate and multivariate analysis, comparing variables such as patient age, gender, body mass index (BMI), laterality, benign versus cancerous renal conditions, estimated blood loss, specimen weight, overall complications, and length of stay. There was no significant difference for operative time between the 2 treatment groups (p = 0.65). Incision related complications occurred in 2 patients (4.4%) from the intact specimen group but none in the morcellation group (p = 0.03). Overall narcotic requirement was lower in patients with morcellated (41 mg) compared to intact specimen retrieval (66 mg) on univariate (p = 0.03) and multivariate analysis (p = 0.049). Upon further stratification, however, there was no significant difference in mean narcotic requirement between the morcellation and muscle-splitting incision subgroup (p = 0.14). Morcellation does not extend operative time, and is associated with significantly less postoperative pain compared to intact specimen retrieval overall, although this is not statistically significant if a remote, muscle-splitting incision is made. Morcellation markedly reduces the risk of incisional-related complications.
NASA Astrophysics Data System (ADS)
Diao, Chunyuan
In today's big data era, the increasing availability of satellite and airborne platforms at various spatial and temporal scales creates unprecedented opportunities to understand the complex and dynamic systems (e.g., plant invasion). Time series remote sensing is becoming more and more important to monitor the earth system dynamics and interactions. To date, most of the time series remote sensing studies have been conducted with the images acquired at coarse spatial scale, due to their relatively high temporal resolution. The construction of time series at fine spatial scale, however, is limited to few or discrete images acquired within or across years. The objective of this research is to advance the time series remote sensing at fine spatial scale, particularly to shift from discrete time series remote sensing to continuous time series remote sensing. The objective will be achieved through the following aims: 1) Advance intra-annual time series remote sensing under the pure-pixel assumption; 2) Advance intra-annual time series remote sensing under the mixed-pixel assumption; 3) Advance inter-annual time series remote sensing in monitoring the land surface dynamics; and 4) Advance the species distribution model with time series remote sensing. Taking invasive saltcedar as an example, four methods (i.e., phenological time series remote sensing model, temporal partial unmixing method, multiyear spectral angle clustering model, and time series remote sensing-based spatially explicit species distribution model) were developed to achieve the objectives. Results indicated that the phenological time series remote sensing model could effectively map saltcedar distributions through characterizing the seasonal phenological dynamics of plant species throughout the year. The proposed temporal partial unmixing method, compared to conventional unmixing methods, could more accurately estimate saltcedar abundance within a pixel by exploiting the adequate temporal signatures of saltcedar. The multiyear spectral angle clustering model could guide the selection of the most representative remotely sensed image for repetitive saltcedar mapping over space and time. Through incorporating spatial autocorrelation, the species distribution model developed in the study could identify the suitable habitats of saltcedar at a fine spatial scale and locate appropriate areas at high risk of saltcedar infestation. Among 10 environmental variables, the distance to the river and the phenological attributes summarized by the time series remote sensing were regarded as the most important. These methods developed in the study provide new perspectives on how the continuous time series can be leveraged under various conditions to investigate the plant invasion dynamics.
A hybrid approach EMD-HW for short-term forecasting of daily stock market time series data
NASA Astrophysics Data System (ADS)
Awajan, Ahmad Mohd; Ismail, Mohd Tahir
2017-08-01
Recently, forecasting time series has attracted considerable attention in the field of analyzing financial time series data, specifically within the stock market index. Moreover, stock market forecasting is a challenging area of financial time-series forecasting. In this study, a hybrid methodology between Empirical Mode Decomposition with the Holt-Winter method (EMD-HW) is used to improve forecasting performances in financial time series. The strength of this EMD-HW lies in its ability to forecast non-stationary and non-linear time series without a need to use any transformation method. Moreover, EMD-HW has a relatively high accuracy and offers a new forecasting method in time series. The daily stock market time series data of 11 countries is applied to show the forecasting performance of the proposed EMD-HW. Based on the three forecast accuracy measures, the results indicate that EMD-HW forecasting performance is superior to traditional Holt-Winter forecasting method.
Fulcher, Ben D; Jones, Nick S
2017-11-22
Phenotype measurements frequently take the form of time series, but we currently lack a systematic method for relating these complex data streams to scientifically meaningful outcomes, such as relating the movement dynamics of organisms to their genotype or measurements of brain dynamics of a patient to their disease diagnosis. Previous work addressed this problem by comparing implementations of thousands of diverse scientific time-series analysis methods in an approach termed highly comparative time-series analysis. Here, we introduce hctsa, a software tool for applying this methodological approach to data. hctsa includes an architecture for computing over 7,700 time-series features and a suite of analysis and visualization algorithms to automatically select useful and interpretable time-series features for a given application. Using exemplar applications to high-throughput phenotyping experiments, we show how hctsa allows researchers to leverage decades of time-series research to quantify and understand informative structure in time-series data. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.
Time series momentum and contrarian effects in the Chinese stock market
NASA Astrophysics Data System (ADS)
Shi, Huai-Long; Zhou, Wei-Xing
2017-10-01
This paper concentrates on the time series momentum or contrarian effects in the Chinese stock market. We evaluate the performance of the time series momentum strategy applied to major stock indices in mainland China and explore the relation between the performance of time series momentum strategies and some firm-specific characteristics. Our findings indicate that there is a time series momentum effect in the short run and a contrarian effect in the long run in the Chinese stock market. The performances of the time series momentum and contrarian strategies are highly dependent on the look-back and holding periods and firm-specific characteristics.
Sornborger, Andrew T; Lauderdale, James D
2016-11-01
Neural data analysis has increasingly incorporated causal information to study circuit connectivity. Dimensional reduction forms the basis of most analyses of large multivariate time series. Here, we present a new, multitaper-based decomposition for stochastic, multivariate time series that acts on the covariance of the time series at all lags, C ( τ ), as opposed to standard methods that decompose the time series, X ( t ), using only information at zero-lag. In both simulated and neural imaging examples, we demonstrate that methods that neglect the full causal structure may be discarding important dynamical information in a time series.
Rodgers, Joseph Lee; Beasley, William Howard; Schuelke, Matthew
2014-01-01
Many data structures, particularly time series data, are naturally seasonal, cyclical, or otherwise circular. Past graphical methods for time series have focused on linear plots. In this article, we move graphical analysis onto the circle. We focus on 2 particular methods, one old and one new. Rose diagrams are circular histograms and can be produced in several different forms using the RRose software system. In addition, we propose, develop, illustrate, and provide software support for a new circular graphical method, called Wrap-Around Time Series Plots (WATS Plots), which is a graphical method useful to support time series analyses in general but in particular in relation to interrupted time series designs. We illustrate the use of WATS Plots with an interrupted time series design evaluating the effect of the Oklahoma City bombing on birthrates in Oklahoma County during the 10 years surrounding the bombing of the Murrah Building in Oklahoma City. We compare WATS Plots with linear time series representations and overlay them with smoothing and error bands. Each method is shown to have advantages in relation to the other; in our example, the WATS Plots more clearly show the existence and effect size of the fertility differential.
Nonlinear parametric model for Granger causality of time series
NASA Astrophysics Data System (ADS)
Marinazzo, Daniele; Pellicoro, Mario; Stramaglia, Sebastiano
2006-06-01
The notion of Granger causality between two time series examines if the prediction of one series could be improved by incorporating information of the other. In particular, if the prediction error of the first time series is reduced by including measurements from the second time series, then the second time series is said to have a causal influence on the first one. We propose a radial basis function approach to nonlinear Granger causality. The proposed model is not constrained to be additive in variables from the two time series and can approximate any function of these variables, still being suitable to evaluate causality. Usefulness of this measure of causality is shown in two applications. In the first application, a physiological one, we consider time series of heart rate and blood pressure in congestive heart failure patients and patients affected by sepsis: we find that sepsis patients, unlike congestive heart failure patients, show symmetric causal relationships between the two time series. In the second application, we consider the feedback loop in a model of excitatory and inhibitory neurons: we find that in this system causality measures the combined influence of couplings and membrane time constants.
Carretta, A; Canneto, B; Calori, G; Ceresoli, G L; Campagnoli, E; Arrigoni, G; Vagani, A; Zannini, P
2001-08-01
The incidence of adenocarcinoma and bronchoalveolar carcinoma has increased in recent years. The aim of this study was to retrospectively evaluate radiological and pathological factors affecting survival in patients with bronchoalveolar carcinoma (BAC) or BAC associated with adenocarcinoma who underwent surgical treatment. From May 1988 to September 1999, 49 patients with BAC or BAC and adenocarcinoma underwent surgical treatment. Complete resection was performed in 42 patients. In these patients the impact of the following factors on survival was evaluated: stage, TNM status, radiological and pathological findings (percentage of bronchoalveolar carcinoma in the tumour, presence or absence of sclerosing and mucinous patterns, vascular invasion and lymphocytic infiltration). Twenty-nine patients were male and 20 female. Mean age was 63 years. Five-year survival was 54%. Univariate analysis of the patients who underwent complete resection demonstrated a favourable impact on survival in stages Ia and Ib (P = 0.01) and in the absence of nodal involvement (P = 0.02) and mucinous patterns (P = 0.02). Mucinous pattern was also prognostically relevant at multivariate analysis (P = 0.02). In the 27 patients with stage Ia and Ib disease, univariate analysis demonstrated that the absence of mucinous pattern (P = 0.006) and a higher percentage of BAC (P = 0.01) favourably influenced survival. The latter data were also confirmed by multivariate analysis (P = 0.01). Surgical treatment of early-stage BAC and combined BAC and adenocarcinoma is associated with favourable results. However, the definition of prognostic factors is of utmost importance to improve the results of the treatment. In our series tumours of the mucinous subtype and with a lower percentage of BAC had a worse prognosis.
Matsumoto, Kazumasa; Novara, Giacomo; Gupta, Amit; Margulis, Vitaly; Walton, Thomas J; Roscigno, Marco; Ng, Casey; Kikuchi, Eiji; Zigeuner, Richard; Kassouf, Wassim; Fritsche, Hans-Martin; Ficarra, Vincenzo; Martignoni, Guido; Tritschler, Stefan; Rodriguez, Joaquin Carballido; Seitz, Christian; Weizer, Alon; Remzi, Mesut; Raman, Jay D; Bolenz, Christian; Bensalah, Karim; Koppie, Theresa M; Karakiewicz, Pierre I; Wood, Christopher G; Montorsi, Francesco; Iwamura, Masatsugu; Shariat, Shahrokh F
2011-10-01
•To assess the impact of differences in ethnicity on clinico-pathological characteristics and outcomes of patients with upper urinary tract urothelial carcinoma (UTUC) in a large multi-center series of patients treated with radical nephroureterectomy (RNU). •We retrospectively collected the data of 2163 patients treated with RNU at 20 academic centres in America, Asia, and Europe. •Univariable and multivariable Cox regression models addressed recurrence-free survival (RFS) and cancer-specific survival (CSS). •In all, 1794 (83%) patients were Caucasian and 369 (17%) were Japanese. All the main clinical and pathological features were significantly different between the two ethnicities. •The median follow-up of the whole cohort was 36 months. At last follow-up, 554 patients (26%) developed disease recurrence and 461 (21%) were dead from UTUC. •The 5-year RFS and CSS estimates were 71.5% and 74.2%, respectively, for Caucasian patients compared with 68.8% and 75.4%, respectively, for Japanese patients. •On univariable Cox regression analyses, ethnicity was not significantly associated with either RFS (P= 0.231) or CSS (P= 0.752). •On multivariable Cox regression analyses that adjusted for the effects of age, gender, surgical type, T stage, grade, tumour architecture, presence of concomitant carcinoma in situ, lymphovascular invasion, tumour necrosis, and lymph node status, ethnicity was not associated with either RFS (hazard ratio [HR] 1.1; P= 0.447) or CSS (HR 1.0; P= 0.908). •There were major differences in the clinico-pathological characteristics of Caucasian and Japanese patients. •However, RFS and CSS probabilities were not affected by ethnicity and race was not an independent predictor of either recurrence or cancer-related death. © 2011 THE AUTHORS; BJU INTERNATIONAL © 2011 BJU INTERNATIONAL.
Evaluation of clinical, laboratory and morphologic prognostic factors in colon cancer
Grande, Michele; Milito, Giovanni; Attinà, Grazia Maria; Cadeddu, Federica; Muzi, Marco Gallinella; Nigro, Casimiro; Rulli, Francesco; Farinon, Attilio Maria
2008-01-01
Background The long-term prognosis of patients with colon cancer is dependent on many factors. To investigate the influence of a series of clinical, laboratory and morphological variables on prognosis of colon carcinoma we conducted a retrospective analysis of our data. Methods Ninety-two patients with colon cancer, who underwent surgical resection between January 1999 and December 2001, were analyzed. On survival analysis, demographics, clinical, laboratory and pathomorphological parameters were tested for their potential prognostic value. Furthermore, univariate and multivariate analysis of the above mentioned data were performed considering the depth of tumour invasion into the bowel wall as independent variable. Results On survival analysis we found that depth of tumour invasion (P < 0.001; F-ratio 2.11), type of operation (P < 0.001; F-ratio 3.51) and CT scanning (P < 0.001; F-ratio 5.21) were predictors of survival. Considering the degree of mural invasion as independent variable, on univariate analysis, we observed that mucorrhea, anismus, hematocrit, WBC count, fibrinogen value and CT scanning were significantly related to the degree of mural invasion of the cancer. On the multivariate analysis, fibrinogen value was the most statistically significant variable (P < 0.001) with the highest F-ratio (F-ratio 5.86). Finally, in the present study, the tumour site was significantly related neither to the survival nor to the mural invasion of the tumour. Conclusion The various clinical, laboratory and patho-morphological parameters showed different prognostic value for colon carcinoma. In the future, preoperative prognostic markers will probably gain relevance in order to make a proper choice between surgery, chemotherapy and radiotherapy. Nevertheless, current data do not provide sufficient evidence for preoperative stratification of high and low risk patients. Further assessments in prospective large studies are warranted. PMID:18778464
Telangiectatic osteosarcoma: a review of 87 cases.
Angelini, Andrea; Mavrogenis, Andreas F; Trovarelli, Giulia; Ferrari, Stefano; Picci, Piero; Ruggieri, Pietro
2016-10-01
Telangiectatic osteosarcoma (TOS) is a rare subtype of osteosarcoma. We analyzed (1) oncologic outcome in a large homogeneous series and (2) the role of prognostic factors on prognosis, local recurrence and metastasis. Eighty-seven patients (47 males, 54 %) were retrospectively analyzed. All except 4 had extracompartmental disease, and ten patients had lung metastasis at diagnosis. Pathologic fracture was present in 27 cases (31 %). Seventy-eight patients were treated with neoadjuvant chemotherapy; nine had surgery as first treatment. Limb-salvage surgery was performed in 71 cases, amputation in 14, and rotationplasty in one. One patient died before surgery. Possible prognostic factors were statistically evaluated. Overall survival was 60.7 % at 10 years of follow-up. Fifty-one patients were disease-free (58.6 %), 2 were alive with disease (2.3 %), 31 died with disease (35.6 %), and 3 died of other causes (3.4 %). Ten local recurrences were observed (11 %). Twenty-five patients (29 %) developed lung (22) or bone (3) metastases. No statistical difference was found considering age, metastases at diagnosis, gender, pathologic fracture, tumor volume, compartmental status, number of neoadjuvant chemotherapy agents and treatment. Induced necrosis was significant at both univariate and multivariate analysis (p < 0.0001). TOS does not have a poor prognosis as previously reported in literature, with a survival of about 60 % at 10 years. Most of patients can be cured with neoadjuvant chemotherapy plus surgery (limb sparing surgery is possible and safe). Tumor response to chemotherapy as induced necrosis was the only significant prognostic factors on survival, even if small tumor volume at diagnosis correlates with better prognosis at univariate analysis. IV.
Janssen, Rob P A; du Mée, Arthur W F; van Valkenburg, Juliette; Sala, Harm A G M; Tseng, Carroll M
2013-09-01
Analysis of long-term clinical and radiological outcomes after anterior cruciate ligament (ACL) reconstruction with special attention to knee osteoarthritis and its predictors. A prospective, consecutive case series of 100 patients. Arthroscopic transtibial ACL reconstruction was performed using 4-strand hamstring tendon autografts with a standardized accelerated rehabilitation protocol. Analysis was performed preoperatively and 10 years postoperatively. Clinical examination included Lysholm and Tegner scores, IKDC, KT-1000 testing (MEDmetric Co., San Diego, CA, USA) and leg circumference measurements. Radiological evaluation included AP weight bearing, lateral knee, Rosenberg and sky view X-rays. Radiological classifications were according to Ahlbäck and Kellgren & Lawrence. Statistical analysis included univariate and multivariate logistic regressions. RESULTS CLINICAL OUTCOME: A significant improvement (p < 0.001) between preoperative and postoperative measurements could be demonstrated for the Lysholm and Tegner scores, IKDC patient subjective assessment, KT-1000 measurements, pivot shift test, IKDC score and one-leg hop test. A pivot shift phenomenon (glide) was still present in 43 (50%) patients and correlated with lower levels of activity (p < 0.022). Radiological outcome: At follow-up, 46 (53.5%) patients had signs of osteoarthritis (OA). In this group, 33 patients (72%) had chondral lesions (≥grade 2) at the time of ACL reconstruction. A history of medial meniscectomy before or at the time of ACL reconstruction increased the risk of knee OA 4 times (95% CI 1.41-11.5). An ICRS grade 3 at the time of ACL reconstruction increased the risk of knee OA by 5.2 times (95% CI 1.09-24.8). There was no correlation between OA and activity level (Tegner score ≥6) nor between OA and a positive pivot shift test. Transtibial ACL reconstruction with 4-strand hamstring autograft and accelerated rehabilitation restored anteroposterior knee stability. Clinical parameters and patient satisfaction improved significantly. At 10-year follow-up, radiological signs of OA were present in 53.5 % of the subjects. Risk factors for OA were meniscectomy prior to or at the time of ACL reconstruction and chondral lesions at the time of ACL reconstruction. II.
Brasme, Jean-Francois; Grill, Jacques; Doz, Francois; Lacour, Brigitte; Valteau-Couanet, Dominique; Gaillard, Stephan; Delalande, Olivier; Aghakhani, Nozar; Puget, Stéphanie; Chalumeau, Martin
2012-01-01
Background The long time to diagnosis of medulloblastoma, one of the most frequent brain tumors in children, is the source of painful remorse and sometimes lawsuits. We analyzed its consequences for tumor stage, survival, and sequelae. Patients and Methods This retrospective population-based cohort study included all cases of pediatric medulloblastoma from a region of France between 1990 and 2005. We collected the demographic, clinical, and tumor data and analyzed the relations between the interval from symptom onset until diagnosis, initial disease stage, survival, and neuropsychological and neurological outcome. Results The median interval from symptom onset until diagnosis for the 166 cases was 65 days (interquartile range 31–121, range 3–457). A long interval (defined as longer than the median) was associated with a lower frequency of metastasis in the univariate and multivariate analyses and with a larger tumor volume, desmoplastic histology, and longer survival in the univariate analysis, but not after adjustment for confounding factors. The time to diagnosis was significantly associated with IQ score among survivors. No significant relation was found between the time to diagnosis and neurological disability. In the 62 patients with metastases, a long prediagnosis interval was associated with a higher T stage, infiltration of the fourth ventricle floor, and incomplete surgical resection; it nonetheless did not influence survival significantly in this subgroup. Conclusions We found complex and often inverse relations between time to diagnosis of medulloblastoma in children and initial severity factors, survival, and neuropsychological and neurological outcome. This interval appears due more to the nature of the tumor and its progression than to parental or medical factors. These conclusions should be taken into account in the information provided to parents and in expert assessments produced for malpractice claims. PMID:22485143
Roychowdhury, D F; Hayden, A; Liepa, A M
2003-02-15
This retrospective analysis examined prognostic significance of health-related quality-of-life (HRQoL) parameters combined with baseline clinical factors on outcomes (overall survival, time to progressive disease, and time to treatment failure) in bladder cancer. Outcome and HRQoL (European Organization for Research and Treatment of Cancer Quality of Life Questionnaire C30) data were collected prospectively in a phase III study assessing gemcitabine and cisplatin versus methotrexate, vinblastine, doxorubicin, and cisplatin in locally advanced or metastatic bladder cancer. Prespecified baseline clinical factors (performance status, tumor-node-metastasis staging, visceral metastases [VM], alkaline phosphatase [AP] level, number of metastatic sites, prior radiotherapy, disease measurability, sex, time from diagnosis, and sites of disease) and selected HRQoL parameters (global QoL; all functional scales; symptoms: pain, fatigue, insomnia, dyspnea, anorexia) were evaluated using Cox's proportional hazards model. Factors with individual prognostic value (P <.05) on outcomes in univariate models were assessed for joint prognostic value in a multivariate model. A final model was developed using a backward selection strategy. Patients with baseline HRQoL were included (364 of 405, 90%). The final model predicted longer survival with low/normal AP levels, no VM, high physical functioning, low role functioning, and no anorexia. Positive prognostic factors for time to progressive disease were good performance status, low/normal AP levels, no VM, and minimal fatigue; for time to treatment failure, they were low/normal AP levels, minimal fatigue, and no anorexia. Global QoL was a significant predictor of outcome in univariate analyses but was not retained in the multivariate model. HRQoL parameters are independent prognostic factors for outcome in advanced bladder cancer; their prognostic importance needs further evaluation.
Factors associated with seasonal influenza vaccination in pregnant women.
Henninger, Michelle L; Irving, Stephanie A; Thompson, Mark; Avalos, Lyndsay Ammon; Ball, Sarah W; Shifflett, Pat; Naleway, Allison L
2015-05-01
This observational study followed a cohort of pregnant women during the 2010-2011 influenza season to determine factors associated with vaccination. Participants were 1105 pregnant women who completed a survey assessing health beliefs related to vaccination upon enrollment and were then followed to determine vaccination status by the end of the 2010-2011 influenza season. We conducted univariate and multivariate analyses to explore factors associated with vaccination status and a factor analysis of survey items to identify health beliefs associated with vaccination. Sixty-three percent (n=701) of the participants were vaccinated. In the univariate analyses, multiple factors were associated with vaccination status, including maternal age, race, marital status, educational level, and gravidity. Factor analysis identified two health belief factors associated with vaccination: participant's positive views (factor 1) and negative views (factor 2) of influenza vaccination. In a multivariate logistic regression model, factor 1 was associated with increased likelihood of vaccination (adjusted odds ratio [aOR]=2.18; 95% confidence interval [CI]=1.72-2.78), whereas factor 2 was associated with decreased likelihood of vaccination (aOR=0.36; 95% CI=0.28-0.46). After controlling for the two health belief factors in multivariate analyses, demographic factors significant in univariate analyses were no longer significant. Women who received a provider recommendation were about three times more likely to be vaccinated (aOR=3.14; 95% CI=1.99-4.96). Pregnant women's health beliefs about vaccination appear to be more important than demographic and maternal factors previously associated with vaccination status. Provider recommendation remains one of the most critical factors influencing vaccination during pregnancy.
Yoon, Jong H.; Tamir, Diana; Minzenberg, Michael J.; Ragland, J. Daniel; Ursu, Stefan; Carter, Cameron S.
2009-01-01
Background Multivariate pattern analysis is an alternative method of analyzing fMRI data, which is capable of decoding distributed neural representations. We applied this method to test the hypothesis of the impairment in distributed representations in schizophrenia. We also compared the results of this method with traditional GLM-based univariate analysis. Methods 19 schizophrenia and 15 control subjects viewed two runs of stimuli--exemplars of faces, scenes, objects, and scrambled images. To verify engagement with stimuli, subjects completed a 1-back matching task. A multi-voxel pattern classifier was trained to identify category-specific activity patterns on one run of fMRI data. Classification testing was conducted on the remaining run. Correlation of voxel-wise activity across runs evaluated variance over time in activity patterns. Results Patients performed the task less accurately. This group difference was reflected in the pattern analysis results with diminished classification accuracy in patients compared to controls, 59% and 72% respectively. In contrast, there was no group difference in GLM-based univariate measures. In both groups, classification accuracy was significantly correlated with behavioral measures. Both groups showed highly significant correlation between inter-run correlations and classification accuracy. Conclusions Distributed representations of visual objects are impaired in schizophrenia. This impairment is correlated with diminished task performance, suggesting that decreased integrity of cortical activity patterns is reflected in impaired behavior. Comparisons with univariate results suggest greater sensitivity of pattern analysis in detecting group differences in neural activity and reduced likelihood of non-specific factors driving these results. PMID:18822407
An Energy-Based Similarity Measure for Time Series
NASA Astrophysics Data System (ADS)
Boudraa, Abdel-Ouahab; Cexus, Jean-Christophe; Groussat, Mathieu; Brunagel, Pierre
2007-12-01
A new similarity measure, called SimilB, for time series analysis, based on the cross-[InlineEquation not available: see fulltext.]-energy operator (2004), is introduced. [InlineEquation not available: see fulltext.] is a nonlinear measure which quantifies the interaction between two time series. Compared to Euclidean distance (ED) or the Pearson correlation coefficient (CC), SimilB includes the temporal information and relative changes of the time series using the first and second derivatives of the time series. SimilB is well suited for both nonstationary and stationary time series and particularly those presenting discontinuities. Some new properties of [InlineEquation not available: see fulltext.] are presented. Particularly, we show that [InlineEquation not available: see fulltext.] as similarity measure is robust to both scale and time shift. SimilB is illustrated with synthetic time series and an artificial dataset and compared to the CC and the ED measures.
A hybrid algorithm for clustering of time series data based on affinity search technique.
Aghabozorgi, Saeed; Ying Wah, Teh; Herawan, Tutut; Jalab, Hamid A; Shaygan, Mohammad Amin; Jalali, Alireza
2014-01-01
Time series clustering is an important solution to various problems in numerous fields of research, including business, medical science, and finance. However, conventional clustering algorithms are not practical for time series data because they are essentially designed for static data. This impracticality results in poor clustering accuracy in several systems. In this paper, a new hybrid clustering algorithm is proposed based on the similarity in shape of time series data. Time series data are first grouped as subclusters based on similarity in time. The subclusters are then merged using the k-Medoids algorithm based on similarity in shape. This model has two contributions: (1) it is more accurate than other conventional and hybrid approaches and (2) it determines the similarity in shape among time series data with a low complexity. To evaluate the accuracy of the proposed model, the model is tested extensively using syntactic and real-world time series datasets.
A Hybrid Algorithm for Clustering of Time Series Data Based on Affinity Search Technique
Aghabozorgi, Saeed; Ying Wah, Teh; Herawan, Tutut; Jalab, Hamid A.; Shaygan, Mohammad Amin; Jalali, Alireza
2014-01-01
Time series clustering is an important solution to various problems in numerous fields of research, including business, medical science, and finance. However, conventional clustering algorithms are not practical for time series data because they are essentially designed for static data. This impracticality results in poor clustering accuracy in several systems. In this paper, a new hybrid clustering algorithm is proposed based on the similarity in shape of time series data. Time series data are first grouped as subclusters based on similarity in time. The subclusters are then merged using the k-Medoids algorithm based on similarity in shape. This model has two contributions: (1) it is more accurate than other conventional and hybrid approaches and (2) it determines the similarity in shape among time series data with a low complexity. To evaluate the accuracy of the proposed model, the model is tested extensively using syntactic and real-world time series datasets. PMID:24982966
Caregiving burden in foreign domestic workers caring for frail older adults in Singapore.
Ha, Ngoc Huong Lien; Chong, Mei Sian; Choo, Robin Wai Munn; Tam, Wai Jia; Yap, Philip Lin Kiat
2018-03-21
Although foreign domestic workers (FDWs) play a significant role in caring for frail seniors in Singapore and are vulnerable to caregiving burden, there has been little research conducted hitherto. We explored caregiver burden and its determinants in this study. FDWs (N = 221, M age = 32.3, SD = 6.23) recruited from a hospital geriatric unit completed the Zarit Burden Interview (ZBI) administered in English, Bahasa Melayu, or Burmese. Univariate and multivariate regression were employed to investigate factors influencing caregiving burden in FDWs. Majority were Indonesians (60.0%), married (57.5%) with children (62.4%), with secondary-level education (59.7%), and providing care for >1 year (79.9%). Importantly, 25.1% reported physical health problems and 23.1% encountered language difficulties with employers. Univariate analysis revealed three significant factors associated with caregiving burden: nationality (p < 0.001), lack of privacy (p = 0.029), and caring for persons with dementia (PWD) (p = 0.001). On multivariate regression, FDWs who cared for PWD were 5.47 times (p = 0.013) more likely to experience burden, while FDWs who encountered language difficulties were 5.46 times (p = 0.030) more likely to experience burden. Filipinos FDWs were 9.73 times more likely to express burden (p < 0.001) compared to their Indonesian and Burmese counterparts. The study highlights caregiver burden in FDWs and potential ways to alleviate it by empowering FDWs with dementia-specific caregiving skills, providing language training opportunities, and supporting particular FDW ethnic groups with more emotional and practical help.
Hebert, Jeffrey J; Fritz, Julie M; Koppenhaver, Shane L; Thackeray, Anne; Kjaer, Per
2016-01-01
Explore the relationships between preoperative findings and clinical outcome following lumbar disc surgery, and investigate the prognostic value of physical examination findings after accounting for information acquired from the clinical history. We recruited 55 adult patients scheduled for first time, single-level lumbar discectomy. Participants underwent a standardized preoperative evaluation including real-time ultrasound imaging assessment of lumbar multifidus function, and an 8-week postoperative rehabilitation programme. Clinical outcome was defined by change in disability, and leg and low back pain (LBP) intensity at 10 weeks. Linear regression models were used to identify univariate and multivariate predictors of outcome. Univariate predictors of better outcome varied depending on the outcome measure. Clinical history predictors included a greater proportion of leg pain to LBP, pain medication use, greater time to surgery, and no history of previous physical or injection therapy. Physical examination predictors were a positive straight or cross straight leg raise test, diminished lower extremity strength, sensation or reflexes, and the presence of postural abnormality or pain peripheralization. Preoperative pain peripheralization remained a significant predictor of improved disability (p = 0.04) and LBP (p = 0.02) after accounting for information from the clinical history. Preoperative lumbar multifidus function was not associated with clinical outcome. Information gleaned from the clinical history and physical examination helps to identify patients more likely to succeed with lumbar disc surgery. While this study helps to inform clinical practice, additional research confirming these results is required prior to confident clinical implementation.
Clustering Financial Time Series by Network Community Analysis
NASA Astrophysics Data System (ADS)
Piccardi, Carlo; Calatroni, Lisa; Bertoni, Fabio
In this paper, we describe a method for clustering financial time series which is based on community analysis, a recently developed approach for partitioning the nodes of a network (graph). A network with N nodes is associated to the set of N time series. The weight of the link (i, j), which quantifies the similarity between the two corresponding time series, is defined according to a metric based on symbolic time series analysis, which has recently proved effective in the context of financial time series. Then, searching for network communities allows one to identify groups of nodes (and then time series) with strong similarity. A quantitative assessment of the significance of the obtained partition is also provided. The method is applied to two distinct case-studies concerning the US and Italy Stock Exchange, respectively. In the US case, the stability of the partitions over time is also thoroughly investigated. The results favorably compare with those obtained with the standard tools typically used for clustering financial time series, such as the minimal spanning tree and the hierarchical tree.
Statistical methods for astronomical data with upper limits. I - Univariate distributions
NASA Technical Reports Server (NTRS)
Feigelson, E. D.; Nelson, P. I.
1985-01-01
The statistical treatment of univariate censored data is discussed. A heuristic derivation of the Kaplan-Meier maximum-likelihood estimator from first principles is presented which results in an expression amenable to analytic error analysis. Methods for comparing two or more censored samples are given along with simple computational examples, stressing the fact that most astronomical problems involve upper limits while the standard mathematical methods require lower limits. The application of univariate survival analysis to six data sets in the recent astrophysical literature is described, and various aspects of the use of survival analysis in astronomy, such as the limitations of various two-sample tests and the role of parametric modelling, are discussed.
A perturbative approach for enhancing the performance of time series forecasting.
de Mattos Neto, Paulo S G; Ferreira, Tiago A E; Lima, Aranildo R; Vasconcelos, Germano C; Cavalcanti, George D C
2017-04-01
This paper proposes a method to perform time series prediction based on perturbation theory. The approach is based on continuously adjusting an initial forecasting model to asymptotically approximate a desired time series model. First, a predictive model generates an initial forecasting for a time series. Second, a residual time series is calculated as the difference between the original time series and the initial forecasting. If that residual series is not white noise, then it can be used to improve the accuracy of the initial model and a new predictive model is adjusted using residual series. The whole process is repeated until convergence or the residual series becomes white noise. The output of the method is then given by summing up the outputs of all trained predictive models in a perturbative sense. To test the method, an experimental investigation was conducted on six real world time series. A comparison was made with six other methods experimented and ten other results found in the literature. Results show that not only the performance of the initial model is significantly improved but also the proposed method outperforms the other results previously published. Copyright © 2017 Elsevier Ltd. All rights reserved.
Multifractal analysis of the Korean agricultural market
NASA Astrophysics Data System (ADS)
Kim, Hongseok; Oh, Gabjin; Kim, Seunghwan
2011-11-01
We have studied the long-term memory effects of the Korean agricultural market using the detrended fluctuation analysis (DFA) method. In general, the return time series of various financial data, including stock indices, foreign exchange rates, and commodity prices, are uncorrelated in time, while the volatility time series are strongly correlated. However, we found that the return time series of Korean agricultural commodity prices are anti-correlated in time, while the volatility time series are correlated. The n-point correlations of time series were also examined, and it was found that a multifractal structure exists in Korean agricultural market prices.
Visibility Graph Based Time Series Analysis
Stephen, Mutua; Gu, Changgui; Yang, Huijie
2015-01-01
Network based time series analysis has made considerable achievements in the recent years. By mapping mono/multivariate time series into networks, one can investigate both it’s microscopic and macroscopic behaviors. However, most proposed approaches lead to the construction of static networks consequently providing limited information on evolutionary behaviors. In the present paper we propose a method called visibility graph based time series analysis, in which series segments are mapped to visibility graphs as being descriptions of the corresponding states and the successively occurring states are linked. This procedure converts a time series to a temporal network and at the same time a network of networks. Findings from empirical records for stock markets in USA (S&P500 and Nasdaq) and artificial series generated by means of fractional Gaussian motions show that the method can provide us rich information benefiting short-term and long-term predictions. Theoretically, we propose a method to investigate time series from the viewpoint of network of networks. PMID:26571115
Teaching Univariate Measures of Location-Using Loss Functions
ERIC Educational Resources Information Center
Paolino, Jon-Paul
2018-01-01
This article presents a new method for introductory teaching of the sample mean, median and mode(s) from a univariate dataset. These basic statistical concepts are taught at various levels of education from elementary school curriculums to courses at the tertiary level. These descriptive measures of location can be taught as optimized solutions to…
Meta-Analytic Structural Equation Modeling: A Two-Stage Approach
ERIC Educational Resources Information Center
Cheung, Mike W. L.; Chan, Wai
2005-01-01
To synthesize studies that use structural equation modeling (SEM), researchers usually use Pearson correlations (univariate r), Fisher z scores (univariate z), or generalized least squares (GLS) to combine the correlation matrices. The pooled correlation matrix is then analyzed by the use of SEM. Questionable inferences may occur for these ad hoc…
Chemical Contaminant and Decontaminant Test Methodology Source Document. Second Edition
2012-07-01
performance as described in “A Statistical Overview on Univariate Calibration, Inverse Regression, and Detection Limits: Application to Gas Chromatography...Overview on Univariate Calibration, Inverse Regression, and Detection Limits: Application to Gas Chromatography/Mass Spectrometry Technique. Mass... APPLICATIONS INTERNATIONAL CORPORATION Gunpowder, MD 21010-0068 July 2012 Approved for public release; distribution is unlimited
Quantifying memory in complex physiological time-series.
Shirazi, Amir H; Raoufy, Mohammad R; Ebadi, Haleh; De Rui, Michele; Schiff, Sami; Mazloom, Roham; Hajizadeh, Sohrab; Gharibzadeh, Shahriar; Dehpour, Ahmad R; Amodio, Piero; Jafari, G Reza; Montagnese, Sara; Mani, Ali R
2013-01-01
In a time-series, memory is a statistical feature that lasts for a period of time and distinguishes the time-series from a random, or memory-less, process. In the present study, the concept of "memory length" was used to define the time period, or scale over which rare events within a physiological time-series do not appear randomly. The method is based on inverse statistical analysis and provides empiric evidence that rare fluctuations in cardio-respiratory time-series are 'forgotten' quickly in healthy subjects while the memory for such events is significantly prolonged in pathological conditions such as asthma (respiratory time-series) and liver cirrhosis (heart-beat time-series). The memory length was significantly higher in patients with uncontrolled asthma compared to healthy volunteers. Likewise, it was significantly higher in patients with decompensated cirrhosis compared to those with compensated cirrhosis and healthy volunteers. We also observed that the cardio-respiratory system has simple low order dynamics and short memory around its average, and high order dynamics around rare fluctuations.
Quantifying Memory in Complex Physiological Time-Series
Shirazi, Amir H.; Raoufy, Mohammad R.; Ebadi, Haleh; De Rui, Michele; Schiff, Sami; Mazloom, Roham; Hajizadeh, Sohrab; Gharibzadeh, Shahriar; Dehpour, Ahmad R.; Amodio, Piero; Jafari, G. Reza; Montagnese, Sara; Mani, Ali R.
2013-01-01
In a time-series, memory is a statistical feature that lasts for a period of time and distinguishes the time-series from a random, or memory-less, process. In the present study, the concept of “memory length” was used to define the time period, or scale over which rare events within a physiological time-series do not appear randomly. The method is based on inverse statistical analysis and provides empiric evidence that rare fluctuations in cardio-respiratory time-series are ‘forgotten’ quickly in healthy subjects while the memory for such events is significantly prolonged in pathological conditions such as asthma (respiratory time-series) and liver cirrhosis (heart-beat time-series). The memory length was significantly higher in patients with uncontrolled asthma compared to healthy volunteers. Likewise, it was significantly higher in patients with decompensated cirrhosis compared to those with compensated cirrhosis and healthy volunteers. We also observed that the cardio-respiratory system has simple low order dynamics and short memory around its average, and high order dynamics around rare fluctuations. PMID:24039811
Scale-dependent intrinsic entropies of complex time series.
Yeh, Jia-Rong; Peng, Chung-Kang; Huang, Norden E
2016-04-13
Multi-scale entropy (MSE) was developed as a measure of complexity for complex time series, and it has been applied widely in recent years. The MSE algorithm is based on the assumption that biological systems possess the ability to adapt and function in an ever-changing environment, and these systems need to operate across multiple temporal and spatial scales, such that their complexity is also multi-scale and hierarchical. Here, we present a systematic approach to apply the empirical mode decomposition algorithm, which can detrend time series on various time scales, prior to analysing a signal's complexity by measuring the irregularity of its dynamics on multiple time scales. Simulated time series of fractal Gaussian noise and human heartbeat time series were used to study the performance of this new approach. We show that our method can successfully quantify the fractal properties of the simulated time series and can accurately distinguish modulations in human heartbeat time series in health and disease. © 2016 The Author(s).
NASA Astrophysics Data System (ADS)
Muñoz-Diosdado, A.
2005-01-01
We analyzed databases with gait time series of adults and persons with Parkinson, Huntington and amyotrophic lateral sclerosis (ALS) diseases. We obtained the staircase graphs of accumulated events that can be bounded by a straight line whose slope can be used to distinguish between gait time series from healthy and ill persons. The global Hurst exponent of these series do not show tendencies, we intend that this is because some gait time series have monofractal behavior and others have multifractal behavior so they cannot be characterized with a single Hurst exponent. We calculated the multifractal spectra, obtained the spectra width and found that the spectra of the healthy young persons are almost monofractal. The spectra of ill persons are wider than the spectra of healthy persons. In opposition to the interbeat time series where the pathology implies loss of multifractality, in the gait time series the multifractal behavior emerges with the pathology. Data were collected from healthy and ill subjects as they walked in a roughly circular path and they have sensors in both feet, so we have one time series for the left foot and other for the right foot. First, we analyzed these time series separately, and then we compared both results, with direct comparison and with a cross correlation analysis. We tried to find differences in both time series that can be used as indicators of equilibrium problems.
The examination of headache activity using time-series research designs.
Houle, Timothy T; Remble, Thomas A; Houle, Thomas A
2005-05-01
The majority of research conducted on headache has utilized cross-sectional designs which preclude the examination of dynamic factors and principally rely on group-level effects. The present article describes the application of an individual-oriented process model using time-series analytical techniques. The blending of a time-series approach with an interactive process model allows consideration of the relationships of intra-individual dynamic processes, while not precluding the researcher to examine inter-individual differences. The authors explore the nature of time-series data and present two necessary assumptions underlying the time-series approach. The concept of shock and its contribution to headache activity is also presented. The time-series approach is not without its problems and two such problems are specifically reported: autocorrelation and the distribution of daily observations. The article concludes with the presentation of several analytical techniques suited to examine the time-series interactive process model.
Long-range correlations in time series generated by time-fractional diffusion: A numerical study
NASA Astrophysics Data System (ADS)
Barbieri, Davide; Vivoli, Alessandro
2005-09-01
Time series models showing power law tails in autocorrelation functions are common in econometrics. A special non-Markovian model for such kind of time series is provided by the random walk introduced by Gorenflo et al. as a discretization of time fractional diffusion. The time series so obtained are analyzed here from a numerical point of view in terms of autocorrelations and covariance matrices.
In aquatic systems, time series of dissolved oxygen (DO) have been used to compute estimates of ecosystem metabolism. Central to this open-water method is the assumption that the DO time series is a Lagrangian specification of the flow field. However, most DO time series are coll...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-19
... time series became closer (while depletion at the end of the time series became more divergent); (4) the agreement in the recruitment time series was much improved; (5) recruitment deviations in log space showed much closer agreement; and (6) the fishing intensity time series showed much closer...
75 FR 4570 - Government-Owned Inventions; Availability for Licensing
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-28
... applications. Signal-to-Noise Enhancement in Imaging Applications Using a Time-Series of Images Description of... applications that use a time-series of images. In one embodiment of the invention, a time-series of images is... Imaging Applications Using a Time-Series of Images'' (HHS Reference No. E-292- 2009/0-US-01). Related...
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-11
... Series, any adjusted option series, and any option series until the time to expiration for such series is... existing requirement may at times discourage liquidity in particular options series because a market maker... the option is subject to the Price/Time execution algorithm, the Directed Market Maker shall receive...
A novel water quality data analysis framework based on time-series data mining.
Deng, Weihui; Wang, Guoyin
2017-07-01
The rapid development of time-series data mining provides an emerging method for water resource management research. In this paper, based on the time-series data mining methodology, we propose a novel and general analysis framework for water quality time-series data. It consists of two parts: implementation components and common tasks of time-series data mining in water quality data. In the first part, we propose to granulate the time series into several two-dimensional normal clouds and calculate the similarities in the granulated level. On the basis of the similarity matrix, the similarity search, anomaly detection, and pattern discovery tasks in the water quality time-series instance dataset can be easily implemented in the second part. We present a case study of this analysis framework on weekly Dissolve Oxygen time-series data collected from five monitoring stations on the upper reaches of Yangtze River, China. It discovered the relationship of water quality in the mainstream and tributary as well as the main changing patterns of DO. The experimental results show that the proposed analysis framework is a feasible and efficient method to mine the hidden and valuable knowledge from water quality historical time-series data. Copyright © 2017 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Dyar, M. Darby; Fassett, Caleb I.; Giguere, Stephen; Lepore, Kate; Byrne, Sarah; Boucher, Thomas; Carey, CJ; Mahadevan, Sridhar
2016-09-01
This study uses 1356 spectra from 452 geologically-diverse samples, the largest suite of LIBS rock spectra ever assembled, to compare the accuracy of elemental predictions in models that use only spectral regions thought to contain peaks arising from the element of interest versus those that use information in the entire spectrum. Results show that for the elements Si, Al, Ti, Fe, Mg, Ca, Na, K, Ni, Mn, Cr, Co, and Zn, univariate predictions based on single emission lines are by far the least accurate, no matter how carefully the region of channels/wavelengths is chosen and despite the prominence of the selected emission lines. An automated iterative algorithm was developed to sweep through all 5485 channels of data and select the single region that produces the optimal prediction accuracy for each element using univariate analysis. For the eight major elements, use of this technique results in a 35% improvement in prediction accuracy; for minors, the improvement is 13%. The best wavelength region choice for any given univariate analysis is likely to be an inherent property of the specific training set that cannot be generalized. In comparison, multivariate analysis using partial least-squares (PLS) almost universally outperforms univariate analysis. PLS using all the same wavelength regions from the univariate analysis produces results that improve in accuracy by 63% for major elements and 3% for minor element. This difference is likely a reflection of signal to noise ratios, which are far better for major elements than for minor elements, and likely limit their prediction accuracy by any technique. We also compare predictions using specific wavelength ranges for each element against those employing all channels. Masking out channels to focus on emission lines from a specific element that occurs decreases prediction accuracy for major elements but is useful for minor elements with low signals and proportionally much higher noise; use of PLS rather than univariate analysis is still recommended. Finally, we tested the generalizability of our results by analyzing a second data set from a different instrument. Overall prediction accuracies for the mixed data sets are higher than for either set alone for all major and minor elements except Ni, Cr, and Co, where results are roughly comparable.
Waki, Fusako; Ando, Masashi; Takashima, Atsuo; Yonemori, Kan; Nokihara, Hiroshi; Miyake, Mototaka; Tateishi, Ukihide; Tsuta, Koji; Shimada, Yasuhiro; Fujiwara, Yasuhiro; Tamura, Tomohide
2009-06-01
Leptomeningeal metastasis (LM) occurs in 4-15% of patients with solid tumors. Although the clinical outcomes in cancer patients have been improving recently, no standard treatment for LM has been established as yet. The purpose of this study was to identify the prognostic factors in patients with solid tumors with cytologically proven LM. We retrospectively analyzed a series of 85 consecutive patients with cytologically proven LM who were treated between 1997 and 2005. The primary diseases were as follows; lung cancer (n = 36), breast cancer (n = 33), gastric cancer (n = 8), and others (n = 8). Forty-nine patients had brain metastasis at the time of diagnosis of the LM, and in 51 patients, MRI revealed meningeal dissemination in the brain or spine. The performance status (PS) was 0-1 in 26 patients and 2-4 in 59 patients. Thirty-one patients, including 19 with breast cancer, four with lung cancer, five with gastric cancer and three with other cancers, were treated by intrathecal (IT) chemotherapy. The response rate to the IT was 52% (95% confidence interval (CI): 41.4-62.6%). The median survival was 51 days (range, 3-759 days). A univariate analysis identified breast cancer, good PS (0-1), time to development of the LM (>1 year), and treatment by IT chemotherapy as being associated with a good prognosis, and multivariate analysis identified poor PS (HR: 1.72 (95% CI, 1.04-2.86) P = 0.04) and MRI-proven LM (HR: 1.82 (95% CI, 1.11-2.98) P = 0.02) as being associated with a poor prognosis. In patients with poor prognostic factors, such as poor PS or MRI-proven LM, palliative therapy might be the most suitable treatment strategy.
Kim, Judy E.; Weber, Paul; Szabo, Aniko
2012-01-01
Purpose: To review malpractice claims associated with retained lens fragments during cataract surgery to identify ways to improve patient outcomes. Methods: Retrospective, noncomparative, consecutive case series. Closed claims data related to cataract surgeries complicated by retained lens fragments (1989 through 2009) from an ophthalmic insurance carrier were reviewed. Factors associated with these claims and claims outcomes were analyzed. Results: During the 21-year period, 117 (12.5%) of 937 closed claims associated with cataract surgery were related to retained lens fragments with 108 unique cataract surgeries, 97% against cataract surgeon and 3% against retinal surgeon. Twelve (11%) of 108 claims were resolved by a trial, 30 (28%) were settled, and 66 (61%) were dismissed. The defendant prevailed in 83% of trials. Indemnity payments totaling more than $3,586,000 were made in 32 (30%) of the claims (median payment, $90,000). The difference between the preoperative visual acuity and the final visual acuity was predictive of an indemnity payment (odds ratio [OR], 2.28; P=.001) and going to a trial (OR, 2.93; P=.000). Development of corneal edema was associated with an indemnity payment (OR, 3.50; P=.037). Timing of referral and elevated intraocular pressure (IOP) were statistically significant in univariate analyses but not in multivariate analyses for a trial. Conclusions: Whereas the majority of claims were dismissed, claims associated with greater visual acuity decline, corneal edema, or elevated IOP were more likely to result in a trial or payment. Ways to reduce significant vision loss, including improved management of corneal edema and IOP, and timely referral to a subspecialist should be considered. PMID:23818737
Zhu, Yong; Yuan, Dongxing; Huang, Yongming; Ma, Jian; Feng, Sichao
2013-09-10
Combining fluorescence detection with flow analysis and solid phase extraction (SPE), a highly sensitive and automatic flow system for measurement of ultra-trace ammonium in open ocean water was established. Determination was based on fluorescence detection of a typical product of o-phthaldialdehyde and ammonium. In this study, the fluorescence reaction product could be efficiently extracted onto an SPE cartridge (HLB, hydrophilic-lipophilic balance). The extracted fluorescence compounds were rapidly eluted with ethanol and directed into a flow cell for fluorescence detection. Compared with the common used fluorescence method, the proposed one offered the benefits of improved sensitivity, reduced reagent consumption, negligible salinity effect and lower cost. Experimental parameters were optimized using a univariate experimental design. Calibration curves, ranging from 1.67 to 300nM, were obtained with different reaction times. The recoveries were between 89.5 and 96.5%, and the detection limits in land-based and shipboard laboratories were 0.7 and 1.2nM, respectively. The relative standard deviation was 3.5% (n=5) for an aged seawater sample spiked with 20nM ammonium. Compared with the analytical results obtained using the indophenol blue method coupled to a long-path liquid waveguide capillary cell, the proposed method showed good agreement. The method had been applied on board during a South China Sea cruise in August 2012. A vertical profile of ammonium in the South East Asia Time-Series (SEATS, 18° N, 116° E) station was produced. The distribution of ammonium in the surface seawater of the Qiongdong upwelling in South China Sea is also presented. Copyright © 2013 Elsevier B.V. All rights reserved.
Climate Prediction Center - Stratosphere: Polar Stratosphere and Ozone
depletion processes can occur. In addition, the latitudinal-time cross sections shows the thermal evolution UV Daily Dosage Estimate South Polar Vertical Ozone Profile Time Series of Size of S.H. Polar Vortex Time Series of Size of S.H. PSC Temperature Time Series of Size of N.H. Polar Vortex Time Series of
Using forbidden ordinal patterns to detect determinism in irregularly sampled time series.
Kulp, C W; Chobot, J M; Niskala, B J; Needhammer, C J
2016-02-01
It is known that when symbolizing a time series into ordinal patterns using the Bandt-Pompe (BP) methodology, there will be ordinal patterns called forbidden patterns that do not occur in a deterministic series. The existence of forbidden patterns can be used to identify deterministic dynamics. In this paper, the ability to use forbidden patterns to detect determinism in irregularly sampled time series is tested on data generated from a continuous model system. The study is done in three parts. First, the effects of sampling time on the number of forbidden patterns are studied on regularly sampled time series. The next two parts focus on two types of irregular-sampling, missing data and timing jitter. It is shown that forbidden patterns can be used to detect determinism in irregularly sampled time series for low degrees of sampling irregularity (as defined in the paper). In addition, comments are made about the appropriateness of using the BP methodology to symbolize irregularly sampled time series.
Sensor-Generated Time Series Events: A Definition Language
Anguera, Aurea; Lara, Juan A.; Lizcano, David; Martínez, Maria Aurora; Pazos, Juan
2012-01-01
There are now a great many domains where information is recorded by sensors over a limited time period or on a permanent basis. This data flow leads to sequences of data known as time series. In many domains, like seismography or medicine, time series analysis focuses on particular regions of interest, known as events, whereas the remainder of the time series contains hardly any useful information. In these domains, there is a need for mechanisms to identify and locate such events. In this paper, we propose an events definition language that is general enough to be used to easily and naturally define events in time series recorded by sensors in any domain. The proposed language has been applied to the definition of time series events generated within the branch of medicine dealing with balance-related functions in human beings. A device, called posturograph, is used to study balance-related functions. The platform has four sensors that record the pressure intensity being exerted on the platform, generating four interrelated time series. As opposed to the existing ad hoc proposals, the results confirm that the proposed language is valid, that is generally applicable and accurate, for identifying the events contained in the time series.
Homogenising time series: Beliefs, dogmas and facts
NASA Astrophysics Data System (ADS)
Domonkos, P.
2010-09-01
For obtaining reliable information about climate change and climate variability the use of high quality data series is essentially important, and one basic tool of quality improvements is the statistical homogenisation of observed time series. In the recent decades large number of homogenisation methods has been developed, but the real effects of their application on time series are still not known entirely. The ongoing COST HOME project (COST ES0601) is devoted to reveal the real impacts of homogenisation methods more detailed and with higher confidence than earlier. As part of the COST activity, a benchmark dataset was built whose characteristics approach well the characteristics of real networks of observed time series. This dataset offers much better opportunity than ever to test the wide variety of homogenisation methods, and analyse the real effects of selected theoretical recommendations. The author believes that several old theoretical rules have to be re-evaluated. Some examples of the hot questions, a) Statistically detected change-points can be accepted only with the confirmation of metadata information? b) Do semi-hierarchic algorithms for detecting multiple change-points in time series function effectively in practise? c) Is it good to limit the spatial comparison of candidate series with up to five other series in the neighbourhood? Empirical results - those from the COST benchmark, and other experiments too - show that real observed time series usually include several inhomogeneities of different sizes. Small inhomogeneities seem like part of the climatic variability, thus the pure application of classic theory that change-points of observed time series can be found and corrected one-by-one is impossible. However, after homogenisation the linear trends, seasonal changes and long-term fluctuations of time series are usually much closer to the reality, than in raw time series. The developers and users of homogenisation methods have to bear in mind that the eventual purpose of homogenisation is not to find change-points, but to have the observed time series with statistical properties those characterise well the climate change and climate variability.
Predicting the difficulty of a lead extraction procedure: the LED index.
Bontempi, Luca; Vassanelli, Francesca; Cerini, Manuel; D'Aloia, Antonio; Vizzardi, Enrico; Gargaro, Alessio; Chiusso, Francesco; Mamedouv, Rashad; Lipari, Alessandro; Curnis, Antonio
2014-08-01
According to recent surveys, many sites performing permanent lead extractions do not meet the minimum prerequisites concerning personnel training, procedures' volume, or facility requirements. The current Heart Rhythm Society consensus on lead extractions suggests that patients should be referred to more experienced sites when a better outcome could be achieved. The purpose of this study was to develop a score aimed at predicting the difficulty of a lead extraction procedure through the analysis of a high-volume center database. This score could help to discriminate patients who should be sent to a referral site. A total of 889 permanent leads were extracted from 469 patients. All procedures were performed from January 2009 to May 2012 by two expert electrophysiologists, at the University Hospital of Brescia. Factors influencing the difficulty of a procedure were assessed using a univariate and a multivariate logistic regression model. The fluoroscopy time of the procedure was taken as an index of difficulty. A Lead Extraction Difficulty (LED) score was defined, considering the strongest predictors. Overall, 873 of 889 (98.2%) leads were completely removed. Major complications were reported in one patient (0.2%) who manifested cardiac tamponade. Minor complications occurred in six (1.3%) patients. No deaths occurred. Median fluoroscopic time was 8.7 min (3.3-17.3). A procedure was classified as difficult when fluoroscopy time was more than 31.2 min [90th percentile (PCTL)].At a univariate analysis, the number of extracted leads and years from implant were significantly associated with an increased risk of fluoroscopy time above 90th PCTL [odds ratio (OR) 1.51, 95% confidence interval (CI) 1.08-2.11, P = 0.01; and OR 1.19, 95% CI 1.12-1.25, P < 0.001, respectively). After adjusting for patient age and sex, and combining with other covariates potentially influencing the extraction procedure, a multivariate analysis confirmed a 71% increased risk of fluoroscopy time above 90th PCTL for each additional lead extracted (OR 1.71, 95% CI 1.06-2.77, P = 0.028) and a 23% increased risk for each year of lead age (OR 1.23, 95% CI 1.15-1.31, P < 0.001). Further nonindependent factors increasing the risk were the presence of active fixation leads and dual-coil implantable cardiac defibrillator leads. Conversely, vegetations significantly favored lead extraction.The LED score was defined as: number of extracted leads within a procedure + lead age (years from implant) + 1 if dual-coil - 1 if vegetation. The LED score independently predicted complex procedure (with fluoroscopic time >90th PCTL) both at univariate and multivariate analysis. A receiver-operating characteristic analysis showed an area under the curve of 0.81. A LED score greater than 10 could predict fluoroscopy time above 90th PCTL with a sensitivity of 78.3% and a specificity of 76.7%. The LED score is easy to compute and potentially predicts fluoroscopy time above 90th PCTL with a relatively high accuracy.
ERIC Educational Resources Information Center
St. Clair, Travis; Hallberg, Kelly; Cook, Thomas D.
2014-01-01
Researchers are increasingly using comparative interrupted time series (CITS) designs to estimate the effects of programs and policies when randomized controlled trials are not feasible. In a simple interrupted time series design, researchers compare the pre-treatment values of a treatment group time series to post-treatment values in order to…
Time Series Model Identification and Prediction Variance Horizon.
1980-06-01
stationary time series Y(t). -6- In terms of p(v), the definition of the three time series memory types is: No Memory Short Memory Long Memory X IP (v)I 0 0...X lp(v)l < - I IP (v) = v=1 v=l v=l Within short memory time series there are three types whose classification in terms of correlation functions is...1974) "Some Recent Advances in Time Series Modeling", TEEE Transactions on Automatic ControZ, VoZ . AC-19, No. 6, December, 723-730. Parzen, E. (1976) "An
Extending nonlinear analysis to short ecological time series.
Hsieh, Chih-hao; Anderson, Christian; Sugihara, George
2008-01-01
Nonlinearity is important and ubiquitous in ecology. Though detectable in principle, nonlinear behavior is often difficult to characterize, analyze, and incorporate mechanistically into models of ecosystem function. One obvious reason is that quantitative nonlinear analysis tools are data intensive (require long time series), and time series in ecology are generally short. Here we demonstrate a useful method that circumvents data limitation and reduces sampling error by combining ecologically similar multispecies time series into one long time series. With this technique, individual ecological time series containing as few as 20 data points can be mined for such important information as (1) significantly improved forecast ability, (2) the presence and location of nonlinearity, and (3) the effective dimensionality (the number of relevant variables) of an ecological system.
Hohn, M. Ed; Nuhfer, E.B.; Vinopal, R.J.; Klanderman, D.S.
1980-01-01
Classifying very fine-grained rocks through fabric elements provides information about depositional environments, but is subject to the biases of visual taxonomy. To evaluate the statistical significance of an empirical classification of very fine-grained rocks, samples from Devonian shales in four cored wells in West Virginia and Virginia were measured for 15 variables: quartz, illite, pyrite and expandable clays determined by X-ray diffraction; total sulfur, organic content, inorganic carbon, matrix density, bulk density, porosity, silt, as well as density, sonic travel time, resistivity, and ??-ray response measured from well logs. The four lithologic types comprised: (1) sharply banded shale, (2) thinly laminated shale, (3) lenticularly laminated shale, and (4) nonbanded shale. Univariate and multivariate analyses of variance showed that the lithologic classification reflects significant differences for the variables measured, difference that can be detected independently of stratigraphic effects. Little-known statistical methods found useful in this work included: the multivariate analysis of variance with more than one effect, simultaneous plotting of samples and variables on canonical variates, and the use of parametric ANOVA and MANOVA on ranked data. ?? 1980 Plenum Publishing Corporation.
Barimani, Shirin; Kleinebudde, Peter
2017-10-01
A multivariate analysis method, Science-Based Calibration (SBC), was used for the first time for endpoint determination of a tablet coating process using Raman data. Two types of tablet cores, placebo and caffeine cores, received a coating suspension comprising a polyvinyl alcohol-polyethylene glycol graft-copolymer and titanium dioxide to a maximum coating thickness of 80µm. Raman spectroscopy was used as in-line PAT tool. The spectra were acquired every minute and correlated to the amount of applied aqueous coating suspension. SBC was compared to another well-known multivariate analysis method, Partial Least Squares-regression (PLS) and a simpler approach, Univariate Data Analysis (UVDA). All developed calibration models had coefficient of determination values (R 2 ) higher than 0.99. The coating endpoints could be predicted with root mean square errors (RMSEP) less than 3.1% of the applied coating suspensions. Compared to PLS and UVDA, SBC proved to be an alternative multivariate calibration method with high predictive power. Copyright © 2017 Elsevier B.V. All rights reserved.
ERIC Educational Resources Information Center
Hess, Brian; Olejnik, Stephen; Huberty, Carl J.
2001-01-01
Studied the efficacy of two improvement-over-chance or "I" effect sizes derived from predictive discriminant analysis and logistic regression analysis for two-group univariate mean comparisons through simulation. Discusses the ways in which the usefulness of each of the indices depends on the population characteristics. (SLD)
Vaporize, anatomically vaporize or enucleate the prostate? The flexible use of the GreenLight laser.
Cindolo, Luca; Ruggera, Lorenzo; Destefanis, Paolo; Dadone, Claudio; Ferrari, Giovanni
2017-03-01
GreenLight laser has gained increasing acceptance as a less invasive treatment for lower urinary tract symptoms due to benign prostatic hyperplasia (BPH/LUTS). Three surgical options were developed: standard photovaporization (PVP), anatomical PVP and GreenLight enucleation of prostate (GreenLEP); however, literature lacks a direct comparison among the procedures. Aim of the present study is to compare the three techniques in a multicentre series of patients. Data were collected from consecutive patients with indication to surgical management of BPH/LUTS in five institutions. Patients underwent standard PVP, anatomical PVP or GreenLEP according to surgeon preferences. Standard parameters associated with transurethral prostate surgery were documented prior surgery and during the follow-up. Patients' perception of improvement was measured using a single-item scale. Early (within first 30 post-operative days) and delayed post-operative complications were recorded. Descriptive statistics, univariate and multivariate analysis were used. We evaluate 367 consecutive patients (mean age 69.1 years). Median prostate size and PSA were 68 ml (IQR 50-90) and 2.8 ng/ml (IQR 1.7-4.3), respectively. The median operative time and applied energy were 60 min (IQR 45-75) and 250 kJ (IQR 160-364). Catheterization time and median post-operative stay were 1 and 2 days. No patient was transfused. The overall median Q max values increased for 8-19 ml/s (p < 0.05), median International Prostate Symptoms Score decreased from 24 to 7 (p < 0.05). A total of 7.4% urinary retention, 33.4% bothersome storage symptoms, 2.5% short-term stress incontinence were recorded. Three heart attacks, one pulmonary embolism and one death occurred. Prostate volume was a predictive factor for post-operative storage symptoms (p = 0.049). Nine percentage of patients experienced long-term complications (4, 0.9 and 0.9% of urethral stricture, bladder neck contracture and prostatic fossa sclerosis, respectively) with 2.5% of long-term stress urinary incontinence (conservatively managed). The reintervention rate was 6%. Late complications were associated at univariate analysis with pharmacological therapy (combination therapy vs. alpha blockers alone vs. none: p value = 0.042) and with the surgical approach (standard PVP vs. anatomical PVP vs. GreenLEP p value = 0.011). The patients' perception of satisfaction was 68% "greatly improved", 27% "improved", 4% "not changed" and 1% "worsened" with no differences between techniques. The availability of three different GreenLight laser techniques allows surgeons with different skills to safety use this technology that remains effective with high patient satisfaction. Anatomical vaporization seems to guarantee the best balance between functional outcomes, surgical procedures and complications.
Moran, John L; Solomon, Patricia J
2011-02-01
Time series analysis has seen limited application in the biomedical Literature. The utility of conventional and advanced time series estimators was explored for intensive care unit (ICU) outcome series. Monthly mean time series, 1993-2006, for hospital mortality, severity-of-illness score (APACHE III), ventilation fraction and patient type (medical and surgical), were generated from the Australia and New Zealand Intensive Care Society adult patient database. Analyses encompassed geographical seasonal mortality patterns, series structural time changes, mortality series volatility using autoregressive moving average and Generalized Autoregressive Conditional Heteroscedasticity models in which predicted variances are updated adaptively, and bivariate and multivariate (vector error correction models) cointegrating relationships between series. The mortality series exhibited marked seasonality, declining mortality trend and substantial autocorrelation beyond 24 lags. Mortality increased in winter months (July-August); the medical series featured annual cycling, whereas the surgical demonstrated long and short (3-4 months) cycling. Series structural breaks were apparent in January 1995 and December 2002. The covariance stationary first-differenced mortality series was consistent with a seasonal autoregressive moving average process; the observed conditional-variance volatility (1993-1995) and residual Autoregressive Conditional Heteroscedasticity effects entailed a Generalized Autoregressive Conditional Heteroscedasticity model, preferred by information criterion and mean model forecast performance. Bivariate cointegration, indicating long-term equilibrium relationships, was established between mortality and severity-of-illness scores at the database level and for categories of ICUs. Multivariate cointegration was demonstrated for {log APACHE III score, log ICU length of stay, ICU mortality and ventilation fraction}. A system approach to understanding series time-dependence may be established using conventional and advanced econometric time series estimators. © 2010 Blackwell Publishing Ltd.
Sea change: Charting the course for biogeochemical ocean time-series research in a new millennium
NASA Astrophysics Data System (ADS)
Church, Matthew J.; Lomas, Michael W.; Muller-Karger, Frank
2013-09-01
Ocean time-series provide vital information needed for assessing ecosystem change. This paper summarizes the historical context, major program objectives, and future research priorities for three contemporary ocean time-series programs: The Hawaii Ocean Time-series (HOT), the Bermuda Atlantic Time-series Study (BATS), and the CARIACO Ocean Time-Series. These three programs operate in physically and biogeochemically distinct regions of the world's oceans, with HOT and BATS located in the open-ocean waters of the subtropical North Pacific and North Atlantic, respectively, and CARIACO situated in the anoxic Cariaco Basin of the tropical Atlantic. All three programs sustain near-monthly shipboard occupations of their field sampling sites, with HOT and BATS beginning in 1988, and CARIACO initiated in 1996. The resulting data provide some of the only multi-disciplinary, decadal-scale determinations of time-varying ecosystem change in the global ocean. Facilitated by a scoping workshop (September 2010) sponsored by the Ocean Carbon Biogeochemistry (OCB) program, leaders of these time-series programs sought community input on existing program strengths and for future research directions. Themes that emerged from these discussions included: 1. Shipboard time-series programs are key to informing our understanding of the connectivity between changes in ocean-climate and biogeochemistry 2. The scientific and logistical support provided by shipboard time-series programs forms the backbone for numerous research and education programs. Future studies should be encouraged that seek mechanistic understanding of ecological interactions underlying the biogeochemical dynamics at these sites. 3. Detecting time-varying trends in ocean properties and processes requires consistent, high-quality measurements. Time-series must carefully document analytical procedures and, where possible, trace the accuracy of analyses to certified standards and internal reference materials. 4. Leveraged implementation, testing, and validation of autonomous and remote observing technologies at time-series sites provide new insights into spatiotemporal variability underlying ecosystem changes. 5. The value of existing time-series data for formulating and validating ecosystem models should be promoted. In summary, the scientific underpinnings of ocean time-series programs remain as strong and important today as when these programs were initiated. The emerging data inform our knowledge of the ocean's biogeochemistry and ecology, and improve our predictive capacity about planetary change.
Characterization of time series via Rényi complexity-entropy curves
NASA Astrophysics Data System (ADS)
Jauregui, M.; Zunino, L.; Lenzi, E. K.; Mendes, R. S.; Ribeiro, H. V.
2018-05-01
One of the most useful tools for distinguishing between chaotic and stochastic time series is the so-called complexity-entropy causality plane. This diagram involves two complexity measures: the Shannon entropy and the statistical complexity. Recently, this idea has been generalized by considering the Tsallis monoparametric generalization of the Shannon entropy, yielding complexity-entropy curves. These curves have proven to enhance the discrimination among different time series related to stochastic and chaotic processes of numerical and experimental nature. Here we further explore these complexity-entropy curves in the context of the Rényi entropy, which is another monoparametric generalization of the Shannon entropy. By combining the Rényi entropy with the proper generalization of the statistical complexity, we associate a parametric curve (the Rényi complexity-entropy curve) with a given time series. We explore this approach in a series of numerical and experimental applications, demonstrating the usefulness of this new technique for time series analysis. We show that the Rényi complexity-entropy curves enable the differentiation among time series of chaotic, stochastic, and periodic nature. In particular, time series of stochastic nature are associated with curves displaying positive curvature in a neighborhood of their initial points, whereas curves related to chaotic phenomena have a negative curvature; finally, periodic time series are represented by vertical straight lines.
Long-term memory and volatility clustering in high-frequency price changes
NASA Astrophysics Data System (ADS)
oh, Gabjin; Kim, Seunghwan; Eom, Cheoljun
2008-02-01
We studied the long-term memory in diverse stock market indices and foreign exchange rates using Detrended Fluctuation Analysis (DFA). For all high-frequency market data studied, no significant long-term memory property was detected in the return series, while a strong long-term memory property was found in the volatility time series. The possible causes of the long-term memory property were investigated using the return data filtered by the AR(1) model, reflecting the short-term memory property, the GARCH(1,1) model, reflecting the volatility clustering property, and the FIGARCH model, reflecting the long-term memory property of the volatility time series. The memory effect in the AR(1) filtered return and volatility time series remained unchanged, while the long-term memory property diminished significantly in the volatility series of the GARCH(1,1) filtered data. Notably, there is no long-term memory property, when we eliminate the long-term memory property of volatility by the FIGARCH model. For all data used, although the Hurst exponents of the volatility time series changed considerably over time, those of the time series with the volatility clustering effect removed diminish significantly. Our results imply that the long-term memory property of the volatility time series can be attributed to the volatility clustering observed in the financial time series.
Modeling seasonal variation of hip fracture in Montreal, Canada.
Modarres, Reza; Ouarda, Taha B M J; Vanasse, Alain; Orzanco, Maria Gabriela; Gosselin, Pierre
2012-04-01
The investigation of the association of the climate variables with hip fracture incidences is important in social health issues. This study examined and modeled the seasonal variation of monthly population based hip fracture rate (HFr) time series. The seasonal ARIMA time series modeling approach is used to model monthly HFr incidences time series of female and male patients of the ages 40-74 and 75+ of Montreal, Québec province, Canada, in the period of 1993-2004. The correlation coefficients between meteorological variables such as temperature, snow depth, rainfall depth and day length and HFr are significant. The nonparametric Mann-Kendall test for trend assessment and the nonparametric Levene's test and Wilcoxon's test for checking the difference of HFr before and after change point are also used. The seasonality in HFr indicated sharp difference between winter and summer time. The trend assessment showed decreasing trends in HFr of female and male groups. The nonparametric test also indicated a significant change of the mean HFr. A seasonal ARIMA model was applied for HFr time series without trend and a time trend ARIMA model (TT-ARIMA) was developed and fitted to HFr time series with a significant trend. The multi criteria evaluation showed the adequacy of SARIMA and TT-ARIMA models for modeling seasonal hip fracture time series with and without significant trend. In the time series analysis of HFr of the Montreal region, the effects of the seasonal variation of climate variables on hip fracture are clear. The Seasonal ARIMA model is useful for modeling HFr time series without trend. However, for time series with significant trend, the TT-ARIMA model should be applied for modeling HFr time series. Copyright © 2011 Elsevier Inc. All rights reserved.
Statistical analysis of temperature data sampled at Station-M in the Norwegian Sea
NASA Astrophysics Data System (ADS)
Lorentzen, Torbjørn
2014-02-01
The paper analyzes sea temperature data sampled at Station-M in the Norwegian Sea. The data cover the period 1948-2010. The following questions are addressed: What type of stochastic process characterizes the temperature series? Are there any changes or patterns which indicate climate change? Are there any characteristics in the data which can be linked to the shrinking sea-ice in the Arctic area? Can the series be modeled consistently and applied in forecasting of the future sea temperature? The paper applies the following methods: Augmented Dickey-Fuller tests for testing of unit-root and stationarity, ARIMA-models in univariate modeling, cointegration and error-correcting models are applied in estimating short- and long-term dynamics of non-stationary series, Granger-causality tests in analyzing the interaction pattern between the deep and upper layer temperatures, and simultaneous equation systems are applied in forecasting future temperature. The paper shows that temperature at 2000 m Granger-causes temperature at 150 m, and that the 2000 m series can represent an important information carrier of the long-term development of the sea temperature in the geographical area. Descriptive statistics shows that the temperature level has been on a positive trend since the beginning of the 1980s which is also measured in most of the oceans in the North Atlantic. The analysis shows that the temperature series are cointegrated which means they share the same long-term stochastic trend and they do not diverge too far from each other. The measured long-term temperature increase is one of the factors that can explain the shrinking summer sea-ice in the Arctic region. The analysis shows that there is a significant negative correlation between the shrinking sea ice and the sea temperature at Station-M. The paper shows that the temperature forecasts are conditioned on the properties of the stochastic processes, causality pattern between the variables and specification of model, respectively. The estimated models forecast that temperature at 150 m is expected to increase by 0.018 °C per year, while deep water temperature at 2000 m is expected to increase between 0.0022 and 0.0024 °C per year.
Smoothing of climate time series revisited
NASA Astrophysics Data System (ADS)
Mann, Michael E.
2008-08-01
We present an easily implemented method for smoothing climate time series, generalizing upon an approach previously described by Mann (2004). The method adaptively weights the three lowest order time series boundary constraints to optimize the fit with the raw time series. We apply the method to the instrumental global mean temperature series from 1850-2007 and to various surrogate global mean temperature series from 1850-2100 derived from the CMIP3 multimodel intercomparison project. These applications demonstrate that the adaptive method systematically out-performs certain widely used default smoothing methods, and is more likely to yield accurate assessments of long-term warming trends.
NASA Astrophysics Data System (ADS)
Hamid, Nor Zila Abd; Adenan, Nur Hamiza; Noorani, Mohd Salmi Md
2017-08-01
Forecasting and analyzing the ozone (O3) concentration time series is important because the pollutant is harmful to health. This study is a pilot study for forecasting and analyzing the O3 time series in one of Malaysian educational area namely Shah Alam using chaotic approach. Through this approach, the observed hourly scalar time series is reconstructed into a multi-dimensional phase space, which is then used to forecast the future time series through the local linear approximation method. The main purpose is to forecast the high O3 concentrations. The original method performed poorly but the improved method addressed the weakness thereby enabling the high concentrations to be successfully forecast. The correlation coefficient between the observed and forecasted time series through the improved method is 0.9159 and both the mean absolute error and root mean squared error are low. Thus, the improved method is advantageous. The time series analysis by means of the phase space plot and Cao method identified the presence of low-dimensional chaotic dynamics in the observed O3 time series. Results showed that at least seven factors affect the studied O3 time series, which is consistent with the listed factors from the diurnal variations investigation and the sensitivity analysis from past studies. In conclusion, chaotic approach has been successfully forecast and analyzes the O3 time series in educational area of Shah Alam. These findings are expected to help stakeholders such as Ministry of Education and Department of Environment in having a better air pollution management.
Transformation-cost time-series method for analyzing irregularly sampled data
NASA Astrophysics Data System (ADS)
Ozken, Ibrahim; Eroglu, Deniz; Stemler, Thomas; Marwan, Norbert; Bagci, G. Baris; Kurths, Jürgen
2015-06-01
Irregular sampling of data sets is one of the challenges often encountered in time-series analysis, since traditional methods cannot be applied and the frequently used interpolation approach can corrupt the data and bias the subsequence analysis. Here we present the TrAnsformation-Cost Time-Series (TACTS) method, which allows us to analyze irregularly sampled data sets without degenerating the quality of the data set. Instead of using interpolation we consider time-series segments and determine how close they are to each other by determining the cost needed to transform one segment into the following one. Using a limited set of operations—with associated costs—to transform the time series segments, we determine a new time series, that is our transformation-cost time series. This cost time series is regularly sampled and can be analyzed using standard methods. While our main interest is the analysis of paleoclimate data, we develop our method using numerical examples like the logistic map and the Rössler oscillator. The numerical data allows us to test the stability of our method against noise and for different irregular samplings. In addition we provide guidance on how to choose the associated costs based on the time series at hand. The usefulness of the TACTS method is demonstrated using speleothem data from the Secret Cave in Borneo that is a good proxy for paleoclimatic variability in the monsoon activity around the maritime continent.
Transformation-cost time-series method for analyzing irregularly sampled data.
Ozken, Ibrahim; Eroglu, Deniz; Stemler, Thomas; Marwan, Norbert; Bagci, G Baris; Kurths, Jürgen
2015-06-01
Irregular sampling of data sets is one of the challenges often encountered in time-series analysis, since traditional methods cannot be applied and the frequently used interpolation approach can corrupt the data and bias the subsequence analysis. Here we present the TrAnsformation-Cost Time-Series (TACTS) method, which allows us to analyze irregularly sampled data sets without degenerating the quality of the data set. Instead of using interpolation we consider time-series segments and determine how close they are to each other by determining the cost needed to transform one segment into the following one. Using a limited set of operations-with associated costs-to transform the time series segments, we determine a new time series, that is our transformation-cost time series. This cost time series is regularly sampled and can be analyzed using standard methods. While our main interest is the analysis of paleoclimate data, we develop our method using numerical examples like the logistic map and the Rössler oscillator. The numerical data allows us to test the stability of our method against noise and for different irregular samplings. In addition we provide guidance on how to choose the associated costs based on the time series at hand. The usefulness of the TACTS method is demonstrated using speleothem data from the Secret Cave in Borneo that is a good proxy for paleoclimatic variability in the monsoon activity around the maritime continent.
a Method of Time-Series Change Detection Using Full Polsar Images from Different Sensors
NASA Astrophysics Data System (ADS)
Liu, W.; Yang, J.; Zhao, J.; Shi, H.; Yang, L.
2018-04-01
Most of the existing change detection methods using full polarimetric synthetic aperture radar (PolSAR) are limited to detecting change between two points in time. In this paper, a novel method was proposed to detect the change based on time-series data from different sensors. Firstly, the overall difference image of a time-series PolSAR was calculated by ominous statistic test. Secondly, difference images between any two images in different times ware acquired by Rj statistic test. Generalized Gaussian mixture model (GGMM) was used to obtain time-series change detection maps in the last step for the proposed method. To verify the effectiveness of the proposed method, we carried out the experiment of change detection by using the time-series PolSAR images acquired by Radarsat-2 and Gaofen-3 over the city of Wuhan, in China. Results show that the proposed method can detect the time-series change from different sensors.
The method of trend analysis of parameters time series of gas-turbine engine state
NASA Astrophysics Data System (ADS)
Hvozdeva, I.; Myrhorod, V.; Derenh, Y.
2017-10-01
This research substantiates an approach to interval estimation of time series trend component. The well-known methods of spectral and trend analysis are used for multidimensional data arrays. The interval estimation of trend component is proposed for the time series whose autocorrelation matrix possesses a prevailing eigenvalue. The properties of time series autocorrelation matrix are identified.
Network structure of multivariate time series.
Lacasa, Lucas; Nicosia, Vincenzo; Latora, Vito
2015-10-21
Our understanding of a variety of phenomena in physics, biology and economics crucially depends on the analysis of multivariate time series. While a wide range tools and techniques for time series analysis already exist, the increasing availability of massive data structures calls for new approaches for multidimensional signal processing. We present here a non-parametric method to analyse multivariate time series, based on the mapping of a multidimensional time series into a multilayer network, which allows to extract information on a high dimensional dynamical system through the analysis of the structure of the associated multiplex network. The method is simple to implement, general, scalable, does not require ad hoc phase space partitioning, and is thus suitable for the analysis of large, heterogeneous and non-stationary time series. We show that simple structural descriptors of the associated multiplex networks allow to extract and quantify nontrivial properties of coupled chaotic maps, including the transition between different dynamical phases and the onset of various types of synchronization. As a concrete example we then study financial time series, showing that a multiplex network analysis can efficiently discriminate crises from periods of financial stability, where standard methods based on time-series symbolization often fail.
Homogenising time series: beliefs, dogmas and facts
NASA Astrophysics Data System (ADS)
Domonkos, P.
2011-06-01
In the recent decades various homogenisation methods have been developed, but the real effects of their application on time series are still not known sufficiently. The ongoing COST action HOME (COST ES0601) is devoted to reveal the real impacts of homogenisation methods more detailed and with higher confidence than earlier. As a part of the COST activity, a benchmark dataset was built whose characteristics approach well the characteristics of real networks of observed time series. This dataset offers much better opportunity than ever before to test the wide variety of homogenisation methods, and analyse the real effects of selected theoretical recommendations. Empirical results show that real observed time series usually include several inhomogeneities of different sizes. Small inhomogeneities often have similar statistical characteristics than natural changes caused by climatic variability, thus the pure application of the classic theory that change-points of observed time series can be found and corrected one-by-one is impossible. However, after homogenisation the linear trends, seasonal changes and long-term fluctuations of time series are usually much closer to the reality than in raw time series. Some problems around detecting multiple structures of inhomogeneities, as well as that of time series comparisons within homogenisation procedures are discussed briefly in the study.
Kamphuis, C; Frank, E; Burke, J K; Verkerk, G A; Jago, J G
2013-01-01
The hypothesis was that sensors currently available on farm that monitor behavioral and physiological characteristics have potential for the detection of lameness in dairy cows. This was tested by applying additive logistic regression to variables derived from sensor data. Data were collected between November 2010 and June 2012 on 5 commercial pasture-based dairy farms. Sensor data from weigh scales (liveweight), pedometers (activity), and milk meters (milking order, unadjusted and adjusted milk yield in the first 2 min of milking, total milk yield, and milking duration) were collected at every milking from 4,904 cows. Lameness events were recorded by farmers who were trained in detecting lameness before the study commenced. A total of 318 lameness events affecting 292 cows were available for statistical analyses. For each lameness event, the lame cow's sensor data for a time period of 14 d before observation date were randomly matched by farm and date to 10 healthy cows (i.e., cows that were not lame and had no other health event recorded for the matched time period). Sensor data relating to the 14-d time periods were used for developing univariable (using one source of sensor data) and multivariable (using multiple sources of sensor data) models. Model development involved the use of additive logistic regression by applying the LogitBoost algorithm with a regression tree as base learner. The model's output was a probability estimate for lameness, given the sensor data collected during the 14-d time period. Models were validated using leave-one-farm-out cross-validation and, as a result of this validation, each cow in the data set (318 lame and 3,180 nonlame cows) received a probability estimate for lameness. Based on the area under the curve (AUC), results indicated that univariable models had low predictive potential, with the highest AUC values found for liveweight (AUC=0.66), activity (AUC=0.60), and milking order (AUC=0.65). Combining these 3 sensors improved AUC to 0.74. Detection performance of this combined model varied between farms but it consistently and significantly outperformed univariable models across farms at a fixed specificity of 80%. Still, detection performance was not high enough to be implemented in practice on large, pasture-based dairy farms. Future research may improve performance by developing variables based on sensor data of liveweight, activity, and milking order, but that better describe changes in sensor data patterns when cows go lame. Copyright © 2013 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.
The Timeseries Toolbox - A Web Application to Enable Accessible, Reproducible Time Series Analysis
NASA Astrophysics Data System (ADS)
Veatch, W.; Friedman, D.; Baker, B.; Mueller, C.
2017-12-01
The vast majority of data analyzed by climate researchers are repeated observations of physical process or time series data. This data lends itself of a common set of statistical techniques and models designed to determine trends and variability (e.g., seasonality) of these repeated observations. Often, these same techniques and models can be applied to a wide variety of different time series data. The Timeseries Toolbox is a web application designed to standardize and streamline these common approaches to time series analysis and modeling with particular attention to hydrologic time series used in climate preparedness and resilience planning and design by the U. S. Army Corps of Engineers. The application performs much of the pre-processing of time series data necessary for more complex techniques (e.g. interpolation, aggregation). With this tool, users can upload any dataset that conforms to a standard template and immediately begin applying these techniques to analyze their time series data.
Fuzzy time-series based on Fibonacci sequence for stock price forecasting
NASA Astrophysics Data System (ADS)
Chen, Tai-Liang; Cheng, Ching-Hsue; Jong Teoh, Hia
2007-07-01
Time-series models have been utilized to make reasonably accurate predictions in the areas of stock price movements, academic enrollments, weather, etc. For promoting the forecasting performance of fuzzy time-series models, this paper proposes a new model, which incorporates the concept of the Fibonacci sequence, the framework of Song and Chissom's model and the weighted method of Yu's model. This paper employs a 5-year period TSMC (Taiwan Semiconductor Manufacturing Company) stock price data and a 13-year period of TAIEX (Taiwan Stock Exchange Capitalization Weighted Stock Index) stock index data as experimental datasets. By comparing our forecasting performances with Chen's (Forecasting enrollments based on fuzzy time-series. Fuzzy Sets Syst. 81 (1996) 311-319), Yu's (Weighted fuzzy time-series models for TAIEX forecasting. Physica A 349 (2004) 609-624) and Huarng's (The application of neural networks to forecast fuzzy time series. Physica A 336 (2006) 481-491) models, we conclude that the proposed model surpasses in accuracy these conventional fuzzy time-series models.
Time-series modeling of long-term weight self-monitoring data.
Helander, Elina; Pavel, Misha; Jimison, Holly; Korhonen, Ilkka
2015-08-01
Long-term self-monitoring of weight is beneficial for weight maintenance, especially after weight loss. Connected weight scales accumulate time series information over long term and hence enable time series analysis of the data. The analysis can reveal individual patterns, provide more sensitive detection of significant weight trends, and enable more accurate and timely prediction of weight outcomes. However, long term self-weighing data has several challenges which complicate the analysis. Especially, irregular sampling, missing data, and existence of periodic (e.g. diurnal and weekly) patterns are common. In this study, we apply time series modeling approach on daily weight time series from two individuals and describe information that can be extracted from this kind of data. We study the properties of weight time series data, missing data and its link to individuals behavior, periodic patterns and weight series segmentation. Being able to understand behavior through weight data and give relevant feedback is desired to lead to positive intervention on health behaviors.
FALSE DETERMINATIONS OF CHAOS IN SHORT NOISY TIME SERIES. (R828745)
A method (NEMG) proposed in 1992 for diagnosing chaos in noisy time series with 50 or fewer observations entails fitting the time series with an empirical function which predicts an observation in the series from previous observations, and then estimating the rate of divergenc...
NASA Astrophysics Data System (ADS)
Diosdado, A. Muñoz; Cruz, H. Reyes; Hernández, D. Bueno; Coyt, G. Gálvez; González, J. Arellanes
2008-08-01
Heartbeat fluctuations exhibit temporal structure with fractal and nonlinear features that reflect changes in the neuroautonomic control. In this work we have used the detrended fluctuation analysis (DFA) to analyze heartbeat (RR) intervals of 54 healthy subjects and 40 patients with congestive heart failure during 24 hours; we separate time series for sleep and wake phases. We observe long-range correlations in time series of healthy persons and CHF patients. However, the correlations for CHF patients are weaker than the correlations for healthy persons; this fact has been reported by Ashkenazy et al. [1] but with a smaller group of subjects. In time series of CHF patients there is a crossover, it means that the correlations for high and low frequencies are different, but in time series of healthy persons there are not crossovers even if they are sleeping. These crossovers are more pronounced for CHF patients in the sleep phase. We decompose the heartbeat interval time series into magnitude and sign series, we know that these kinds of signals can exhibit different time organization for the magnitude and sign and the magnitude series relates to nonlinear properties of the original time series, while the sign series relates to the linear properties. Magnitude series are long-range correlated, while the sign series are anticorrelated. Newly, the correlations for healthy persons are different that the correlations for CHF patients both for magnitude and sign time series. In the paper of Ashkenazy et al. they proposed the empirical relation: αsign≈1/2(αoriginal+αmagnitude) for the short-range regime (high frequencies), however, we have found a different relation that in our calculations is valid for short and long-range regime: αsign≈1/4(αoriginal+αmagnitude).
Koltun, G.F.
2015-01-01
Streamflow hydrographs were plotted for modeled/computed time series for the Ohio River near the USGS Sardis gage and the Ohio River at the Hannibal Lock and Dam. In general, the time series at these two locations compared well. Some notable differences include the exclusive presence of short periods of negative streamflows in the USGS 15-minute time-series data for the gage on the Ohio River above Sardis, Ohio, and the occurrence of several peak streamflows in the USACE gate/hydropower time series for the Hannibal Lock and Dam that were appreciably larger than corresponding peaks in the other time series, including those modeled/computed for the downstream Sardis gage
Dunea, Daniel; Pohoata, Alin; Iordache, Stefania
2015-07-01
The paper presents the screening of various feedforward neural networks (FANN) and wavelet-feedforward neural networks (WFANN) applied to time series of ground-level ozone (O3), nitrogen dioxide (NO2), and particulate matter (PM10 and PM2.5 fractions) recorded at four monitoring stations located in various urban areas of Romania, to identify common configurations with optimal generalization performance. Two distinct model runs were performed as follows: data processing using hourly-recorded time series of airborne pollutants during cold months (O3, NO2, and PM10), when residential heating increases the local emissions, and data processing using 24-h daily averaged concentrations (PM2.5) recorded between 2009 and 2012. Dataset variability was assessed using statistical analysis. Time series were passed through various FANNs. Each time series was decomposed in four time-scale components using three-level wavelets, which have been passed also through FANN, and recomposed into a single time series. The agreement between observed and modelled output was evaluated based on the statistical significance (r coefficient and correlation between errors and data). Daubechies db3 wavelet-Rprop FANN (6-4-1) utilization gave positive results for O3 time series optimizing the exclusive use of the FANN for hourly-recorded time series. NO2 was difficult to model due to time series specificity, but wavelet integration improved FANN performances. Daubechies db3 wavelet did not improve the FANN outputs for PM10 time series. Both models (FANN/WFANN) overestimated PM2.5 forecasted values in the last quarter of time series. A potential improvement of the forecasted values could be the integration of a smoothing algorithm to adjust the PM2.5 model outputs.
Zhang, Yatao; Wei, Shoushui; Liu, Hai; Zhao, Lina; Liu, Chengyu
2016-09-01
The Lempel-Ziv (LZ) complexity and its variants have been extensively used to analyze the irregularity of physiological time series. To date, these measures cannot explicitly discern between the irregularity and the chaotic characteristics of physiological time series. Our study compared the performance of an encoding LZ (ELZ) complexity algorithm, a novel variant of the LZ complexity algorithm, with those of the classic LZ (CLZ) and multistate LZ (MLZ) complexity algorithms. Simulation experiments on Gaussian noise, logistic chaotic, and periodic time series showed that only the ELZ algorithm monotonically declined with the reduction in irregularity in time series, whereas the CLZ and MLZ approaches yielded overlapped values for chaotic time series and time series mixed with Gaussian noise, demonstrating the accuracy of the proposed ELZ algorithm in capturing the irregularity, rather than the complexity, of physiological time series. In addition, the effect of sequence length on the ELZ algorithm was more stable compared with those on CLZ and MLZ, especially when the sequence length was longer than 300. A sensitivity analysis for all three LZ algorithms revealed that both the MLZ and the ELZ algorithms could respond to the change in time sequences, whereas the CLZ approach could not. Cardiac interbeat (RR) interval time series from the MIT-BIH database were also evaluated, and the results showed that the ELZ algorithm could accurately measure the inherent irregularity of the RR interval time series, as indicated by lower LZ values yielded from a congestive heart failure group versus those yielded from a normal sinus rhythm group (p < 0.01). Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Mesli, Smain Nabil; Regagba, Derbali; Tidjane, Anisse; Benkalfat, Mokhtar; Abi-Ayad, Chakib
2016-01-01
The aim of our study was to analyze histoprognostic factors in patients with non-metastatic rectal cancer operated at the division of surgery "A" in Tlemcen, west Algeria, over a period of six years. Retrospective study of 58 patients with rectal adenocarcinoma. Evaluation criterion was survival. Parameters studied were sex, age, tumor stage, tumor recurrence. The average age was 58 years, 52% of men and 48% of women, with sex-ratio (1,08). Tumor seat was: middle rectum 41.37%, lower rectum 34.48% and upper rectum 24.13%. Concerning TNM clinical staging, patients were classified as stage I (17.65%), stage II (18.61%), stage III (53.44%) and stage IV (7.84%). Median overall survival was 40 months ±2,937 months. Survival based on tumor staging: stage III and IV had a lower 3 years survival rate (19%) versus stage I, II which had a survival rate of 75% (P = 0.000) (95%). Patients with tumor recurrences had a lower 3 years survival rate compared to those who had no tumoral recurrences (30.85% vs 64.30% P = 0.043). In this series, univariate analysis of prognostic factors affecting survival allowed to retain only three factors influencing survival: tumor size, stage and tumor recurrences. In multivariate analysis using Cox's model only one factor was retained: tumor recurrence.
2016-01-01
Multivariate calibration (MVC) and near-infrared (NIR) spectroscopy have demonstrated potential for rapid analysis of melamine in various dairy products. However, the practical application of ordinary MVC can be largely restricted because the prediction of a new sample from an uncalibrated batch would be subject to a significant bias due to matrix effect. In this study, the feasibility of using NIR spectroscopy and the standard addition (SA) net analyte signal (NAS) method (SANAS) for rapid quantification of melamine in different brands/types of milk powders was investigated. In SANAS, the NAS vector of melamine in an unknown sample as well as in a series of samples added with melamine standards was calculated and then the Euclidean norms of series standards were used to build a straightforward univariate regression model. The analysis results of 10 different brands/types of milk powders with melamine levels 0~0.12% (w/w) indicate that SANAS obtained accurate results with the root mean squared error of prediction (RMSEP) values ranging from 0.0012 to 0.0029. An additional advantage of NAS is to visualize and control the possible unwanted variations during standard addition. The proposed method will provide a practically useful tool for rapid and nondestructive quantification of melamine in different brands/types of milk powders. PMID:27525154
Advanced statistics: linear regression, part II: multiple linear regression.
Marill, Keith A
2004-01-01
The applications of simple linear regression in medical research are limited, because in most situations, there are multiple relevant predictor variables. Univariate statistical techniques such as simple linear regression use a single predictor variable, and they often may be mathematically correct but clinically misleading. Multiple linear regression is a mathematical technique used to model the relationship between multiple independent predictor variables and a single dependent outcome variable. It is used in medical research to model observational data, as well as in diagnostic and therapeutic studies in which the outcome is dependent on more than one factor. Although the technique generally is limited to data that can be expressed with a linear function, it benefits from a well-developed mathematical framework that yields unique solutions and exact confidence intervals for regression coefficients. Building on Part I of this series, this article acquaints the reader with some of the important concepts in multiple regression analysis. These include multicollinearity, interaction effects, and an expansion of the discussion of inference testing, leverage, and variable transformations to multivariate models. Examples from the first article in this series are expanded on using a primarily graphic, rather than mathematical, approach. The importance of the relationships among the predictor variables and the dependence of the multivariate model coefficients on the choice of these variables are stressed. Finally, concepts in regression model building are discussed.
New approach in bivariate drought duration and severity analysis
NASA Astrophysics Data System (ADS)
Montaseri, Majid; Amirataee, Babak; Rezaie, Hossein
2018-04-01
The copula functions have been widely applied as an advance technique to create joint probability distribution of drought duration and severity. The approach of data collection as well as the amount of data and dispersion of data series can last a significant impact on creating such joint probability distribution using copulas. Usually, such traditional analyses have shed an Unconnected Drought Runs (UDR) approach towards droughts. In other word, droughts with different durations would be independent of each other. Emphasis on such data collection method causes the omission of actual potentials of short-term extreme droughts located within a long-term UDR. Meanwhile, traditional method is often faced with significant gap in drought data series. However, a long-term UDR can be approached as a combination of short-term Connected Drought Runs (CDR). Therefore this study aims to evaluate systematically two UDR and CDR procedures in joint probability of drought duration and severity investigations. For this purpose, rainfall data (1971-2013) from 24 rain gauges in Lake Urmia basin, Iran were applied. Also, seven common univariate marginal distributions and seven types of bivariate copulas were examined. Compared to traditional approach, the results demonstrated a significant comparative advantage of the new approach. Such comparative advantages led to determine the correct copula function, more accurate estimation of copula parameter, more realistic estimation of joint/conditional probabilities of drought duration and severity and significant reduction in uncertainty for modeling.
Xu, Bo; Hu, Jinghai; Chen, Anxiang; Hao, Yuanyuan; Liu, GuoHui; Wang, Chunxi; Wang, Xiaoqing
2017-06-01
The present study was designed to investigate the risk factors affecting the conversion to open surgery in retroperitoneal laparoscopic nephrectomy of nonfunctioning renal tuberculosis (TB). The records of 144 patients who underwent a retroperitoneal laparoscopic nephrectomy procedure by a single surgeon were retrospectively reviewed. The following factors, including age, sex, body mass index (BMI), diabetes status, hypertension status, side of kidney, size of kidney, degree of calcification, mild perirenal extravasation, contralateral hydronephrosis, the time of anti-TB, and surgeon experience were analyzed. Univariate and multivariate logistic regression analyses were used for statistical assessment. Twenty-three patients were converted to open surgery and the conversion rate was 15.97%. In univariate analysis, BMI ≥35 kg/m 2 (p = 0.023), hypertension (p = 0.011), diabetes (p = 0.003), and kidney size (p = 0.032) were the main factors of conversion to open surgery. Sex, age, side, anti-TB time, calcification, mild extravasation, and surgeon experience were not significantly related. In multivariate regression analysis, BMI ≥35 kg/m 2 , hypertension, diabetes, and enlargement of kidney were the most important factors for conversion to open surgery. Depending on the results achieved by a single surgeon, BMI ≥30 kg/m 2 , diabetes, hypertension, and enlargement of kidney significantly increased the conversion risk in retroperitoneal laparoscopic nephrectomy for nonfunctioning renal TB.
Zhang, H-L; Li, L; Cheng, C-J; Sun, X-C
2018-02-01
The study aims to detect the association of miR-146a-5p with intracranial aneurysms (IAs). The expression of miR-146a-5p was compared from plasma samples between 72 patients with intracranial aneurysms (IAs) and 40 healthy volunteers by quantitative Real-time polymerase chain reaction (qRT-PCR). Statistical analysis was performed to analyze the relationship between miR-146a-5p expression and clinical data and overall survival (OS) time of IAs patients. Univariate and multivariate Cox proportional hazards have also been performed. Notably, higher miR-146a-5p expression was found in plasma samples from 72 patients with intracranial aneurysms (IAs) compared with 40 healthy controls. Higher miR-146a-5p expression was significantly associated with rupture and Hunt-Hess level in IAs patients. Kaplan-Meier survival analysis verified that higher miR-146a-5p expression predicted a shorter overall survival (OS) compared with lower miR-146a-5p expression in IAs patients. Univariate and multivariate Cox proportional hazards demonstrated that higher miR-146a-5p expression, rupture, and Hunt-Hess were independent risk factors of OS in patients with intracranial aneurysms (IAs). MiR-146a-5p expression may serve as a biomarker for predicting prognosis in patients with IAs.
Time averaging, ageing and delay analysis of financial time series
NASA Astrophysics Data System (ADS)
Cherstvy, Andrey G.; Vinod, Deepak; Aghion, Erez; Chechkin, Aleksei V.; Metzler, Ralf
2017-06-01
We introduce three strategies for the analysis of financial time series based on time averaged observables. These comprise the time averaged mean squared displacement (MSD) as well as the ageing and delay time methods for varying fractions of the financial time series. We explore these concepts via statistical analysis of historic time series for several Dow Jones Industrial indices for the period from the 1960s to 2015. Remarkably, we discover a simple universal law for the delay time averaged MSD. The observed features of the financial time series dynamics agree well with our analytical results for the time averaged measurables for geometric Brownian motion, underlying the famed Black-Scholes-Merton model. The concepts we promote here are shown to be useful for financial data analysis and enable one to unveil new universal features of stock market dynamics.
Ziv, Etay; Bergen, Michael; Yarmohammadi, Hooman; Boas, F Ed; Petre, E Nadia; Sofocleous, Constantinos T; Yaeger, Rona; Solit, David B; Solomon, Stephen B; Erinjeri, Joseph P
2017-04-04
To establish the relationship between common mutations in the MAPK and PI3K signaling pathways and local progression after radioembolization. Retrospective review of a HIPAA-compliant institutional review-board approved database identified 40 patients with chemo-refractory colorectal liver metastases treated with radioembolization who underwent tumor genotyping for hotspot mutations in 6 key genes in the MAPK/PI3K pathways (KRAS, NRAS, BRAF, MEK1, PIK3CA, and AKT1). Mutation status as well as clinical, tumor, and treatment variables were recorded. These factors were evaluated in relation to time to local progression (TTLP), which was calculated from time of radioembolization to first radiographic evidence of local progression. Predictors of outcome were identified using a proportional hazards model for both univariate and multivariate analysis with death as a competing risk. Sixteen patients (40%) had no mutations in either pathway, eighteen patients (45%) had mutations in the MAPK pathway, ten patients (25%) had mutations in the PI3K pathway and four patients (10%) had mutations in both pathways. The cumulative incidence of progression at 6 and 12 months was 33% and 55% for the PI3K mutated group compared with 76% and 92% in the PI3K wild type group. Mutation in the PI3K pathway was a significant predictor of longer TTLP in both univariate (p=0.031, sHR 0.31, 95% CI: 0.11-0.90) and multivariate (p=0.015, sHR=0.27, 95% CI: 0.096-0.77) analysis. MAPK pathway alterations were not associated with TTLP. PI3K pathway mutation predicts longer time to local progression after radioembolization of colorectal liver metastases.
Neuwirth, Alexander L; Stitzlein, Russell N; Neuwirth, Madalyn G; Kelz, Rachel K; Mehta, Samir
2018-01-17
Future generations of orthopaedic surgeons must continue to be trained in the surgical management of hip fractures. This study assesses the effect of resident participation on outcomes for the treatment of intertrochanteric hip fractures. The National Surgical Quality Improvement Program (NSQIP) database (2010 to 2013) was queried for intertrochanteric hip fractures (International Classification of Diseases, 9th Revision, Clinical Modification [ICD-9-CM] code 820.21) treated with either extramedullary (Current Procedural Terminology [CPT] code 27244) or intramedullary (CPT code 27245) fixation. Demographic variables, including resident participation, as well as primary (death and serious morbidity) and secondary outcome variables were extracted for analysis. Univariate, propensity score-matched, and multivariate logistic regression analyses were performed to evaluate outcome variables. Data on resident participation were available for 1,764 cases (21.0%). Univariate analyses for all intertrochanteric hip fractures demonstrated no significant difference in 30-day mortality (6.3% versus 7.8%; p = 0.264) or serious morbidity (44.9% versus 43.2%; p = 0.506) between the groups with and without resident participation. Multivariate and propensity score-matched analyses gave similar results. Resident involvement was associated with prolonged operating-room time, length of stay, and time to discharge when a prolonged case was defined as one above the 90th percentile for time parameters. Resident participation was not associated with an increase in morbidity or mortality but was associated with an increase in time-related secondary outcome measures. While attending surgeon supervision is necessary, residents can and should be involved in the care of these patients without concern that resident involvement negatively impacts perioperative morbidity and mortality. Therapeutic Level III. See Instructions for Authors for a complete description of levels of evidence.
Konski, Andre; Desilvio, Michelle; Hartsell, William; Watkins-Bruner, Deborah; Coyne, James; Scarantino, Charles; Janjan, Nora
2006-09-01
The specific aim of this study was to evaluate outcome differences by gender and partner status for patients treated on Radiation Therapy Oncology Group (RTOG) protocol 97-14. RTOG 97-14 randomized patients with metastatic breast or prostate cancer to bone to receive 8 Gy in 1 fraction or 30 Gy in 10 fractions. Retreatment rates and overall survival were made based upon gender, marital status, and Karnofsky Performance Status (KPS). The cumulative incidence method was used to estimate retreatment time at 36 months from enrollment, and Gray's test was used to test for treatment differences within the same groupings. Marital status, gender, KPS, and treatment were variables tested in a univariate Cox model evaluating the time to retreatment. Married men and women and single women receiving 30 Gy had significantly longer time to retreatment, p = 0.0067, p = 0.0052, and p = 0.0009 respectively. We failed to show a difference in retreatment rates over time in single men receiving either 30 Gy or 8 Gy. Univariate analysis of the entire group determined patients receiving 30 Gy in 10 fractions significantly less likely to receive retreatment, p < 0.0001, with a trend toward single patients less likely to be re-treated, p = 0.07. Non-disease-related variables, such as social support, might influence the results of clinical trials with subjective endpoints such as retreatment rates. The statistically nonsignificant difference in the 36-month retreatment rates observed in single male patients receiving 8 Gy may be a result of inadequate social support systems in place to facilitate additional care. Patients receiving 8 Gy in a single fraction had significantly higher retreatment rates compared with patients receiving 30 Gy in 10 fractions.
Polanco, Patricio M; Zenati, Mazen S; Hogg, Melissa E; Shakir, Murtaza; Boone, Brian A; Bartlett, David L; Zeh, Herbert J; Zureikat, Amer H
2016-04-01
Proponents of the robotic platform site its potential advantages in complex reconstructions such as the pancreaticojejunal anastomosis; however, the incidence and risk factors for postoperative pancreatic fistula (POPF) after robotic pancreaticoduodenectomy (RPD) have not been characterized. To identify independent risk factors for POPF after RPD. A prospectively maintained database of patients that underwent RPD (2008-2013) with a standardized pancreaticojejunostomy was analyzed. Univariate and multivariate analyses (UVA/MVA) were used to identify independent predictors for POPF. The POPF prognostic scores developed by Braga and Callery for open pancreaticoduodenectomy were then applied with logistic regression analysis on this RPD cohort. One hundred and fifty consecutive RPDs were analyzed. POPF occurred in 26 (17.3%); 13 (8.6%) of which were ISGPF category B and C. On UVA, patients with POPF had larger body mass index (BMI), smaller pancreatic duct diameter, smaller tumor size, longer OR time, larger estimated blood loss (EBL) and RBC transfusion (all p < 0.05). Higher EBL, duct size <4 mm, larger BMI and small tumor size remained the best independent risk factors for POPF on MVA. Increased Callery (OR 1.46, 95% CI, p = 0.001) and Braga (OR 1.2, 95% CI, p = 0.005) scores predicted an increased risk of POPF in this RPD cohort. Larger BMI, higher EBL, smaller tumor size and smaller duct diameter are the main predictors of POPF in RPD.
Maltese, Giuseppa; Fontanella, Caterina; Lepori, Stefano; Scaffa, Cono; Fucà, Giovanni; Bogani, Giorgio; Provenzano, Salvatore; Carcangiu, Maria Luisa; Raspagliesi, Francesco; Lorusso, Domenica
2018-01-01
Clinical characteristics combined with new biomarkers help discriminate between atypical uterine smooth muscle tumors (AUSMT) and leiomyosarcomas (LMS). We retrospectively collected a series of leiomyomas (LM), AUSMT, and LMS. Estrogen receptors (ER), progesterone receptors (PR), p16, Ki-67, and p53 expression were assessed by immunohistochemistry. For AUSMT patients, immunohistochemistry evaluations were performed at the time of diagnosis and at recurrences. A total of 27 cases of AUSMT, 22 LM, and 31 LMS were identified. The expression of ER and PR decreased from LM to LMS (ER+: LM 95.5%, AUSMT 88.9%, LMS 41.9%, p < 0.001; PR+: LM 100%, AUSMT 88.9%, LMS 38.2%, p = 0.002). By contrast, p16 and p53 expression increased (p16+: LM 4.5%, AUSMT 40.7%, LMS 45.2%, p = 0.004; p53: LM 9.1%, AUSMT 33.3%, LMS 58.1%, p = 0.001). At a median follow-up of 33.47 months, 40.7% of patients with AUSMT experienced recurrent disease, 6 patients relapsed as AUSMT and 5 as LMS. In univariate analysis was observed that ER status (p = 0.027) and p53 expression (p = 0.015) predicted risk of relapse. Treatment of AUSMT should be centralized in dedicated centers. International collaborations are needed to optimize research strategy, which may lead to the identification of new useful biomarkers and to improvement in the clinical management of this rare disease. © 2017 S. Karger AG, Basel.
Hepatic Resection for Colorectal Liver Metastases: A Cost-Effectiveness Analysis
Beard, Stephen M.; Holmes, Michael; Price, Charles; Majeed, Ali W.
2000-01-01
Objective To analyze the cost-effectiveness of resection for liver metastases compared with standard nonsurgical cytotoxic treatment. Summary Background Data The efficacy of hepatic resection for metastases from colorectal cancer has been debated, despite reported 5-year survival rates of 20% to 40%. Resection is confined to specialized centers and is not widely available, perhaps because of lack of appropriate expertise, resources, or awareness of its efficacy. The cost-effectiveness of resection is important from the perspective of managed care in the United States and for the commissioning of health services in the United Kingdom. Methods A simple decision-based model was developed to evaluate the marginal costs and health benefits of hepatic resection. Estimates of resectability for liver metastases were taken from UK-reported case series data. The results of 100 hepatic resections conducted in Sheffield from 1997 to 1999 were used for the cost calculation of liver resection. Survival data from published series of resections were compiled to estimate the incremental cost per life-year gained (LYG) because of the short period of follow-up in the Sheffield series. Results Hepatic resection for colorectal liver metastases provides an estimated marginal benefit of 1.6 life-years (undiscounted) at a marginal cost of £6,742. If 17% of patients have only palliative resections, the overall cost per LYG is approximately £5,236 (£5,985 with discounted benefits). If potential benefits are extended to include 20-year survival rates, these figures fall to approximately £1,821 (£2,793 with discounted benefits). Further univariate sensitivity analysis of key model parameters showed the cost per LYG to be consistently less than £15,000. Conclusion In this model, hepatic resection appears highly cost-effective compared with nonsurgical treatments for colorectal-related liver metastases. PMID:11088071
Kolakowski, Stephen; Kirkland, Matt L; Schuricht, Alan L
2007-10-01
To evaluate the clinical utility of the routine use of postoperative barium swallow to diagnose postoperative complications in patients undergoing open or laparoscopic Roux-en-Y gastric bypass. A total of 417 consecutive patients undergoing Roux-en-Y gastric bypass at our institution between January 1, 2001, and December 31, 2002, were included. We performed 341 open procedures and 76 laparoscopic gastric bypasses. All patients received a limited postoperative fluoroscopic upper gastrointestinal series, except for the patients who exceeded the weight limitation of the radiologic equipment. Radiologic findings of anastomotic complications were anastomotic leak, delayed gastric emptying, gastric outlet obstruction, and gastrogastric fistula. We evaluated clinical signs and symptoms to obtain a list of criteria suggesting these complications. Patients were stratified into 2 groups: those with and those without radiographic anastomotic complications. Clinical and radiologic criteria were compared using univariate and multivariate logistic regression analysis. We noted 42 radiologic abnormalities during a routine postoperative barium swallow evaluation. Among our 417 patients, we documented 12 leaks (2.9%), 19 cases of delayed gastric emptying (4.6%), 4 gastric outlet obstructions (1.0%), and 7 gastrogastric fistulas (1.7%). The combination of fever, tachycardia, and tachypnea was the most specific indicator of a leak, at 0.99 (95% confidence limit, 0.99, 1.01). Nausea with vomiting was the most predictive indicator of delayed gastric emptying and gastric outlet obstruction, with a specificity of 0.99 (95% confidence limit, 0.98, 0.99) and 0.97 (95% confidence limit, 0.96, 0.99), respectively. Postoperative complications after Roux-en-Y gastric bypass surgery are predictable based on the patient's symptoms. The use of routine postoperative fluoroscopic upper gastrointestinal series is unnecessary in asymptomatic patients.
[Predictive value of pre-treatment hypoalbuminemia in prognosis of resected colorectal cancer].
Borda, Fernando; Borda, Ana; Jiménez, Javier; Zozaya, José Manuel; Prieto, Carlos; Gómez, Marta; Urman, Jesús; Ibáñez, Berta
2014-05-01
Albuminemia is part of the antitumoral systemic inflammatory response. We therefore analyzed its possible value in establishing the preoperative prognosis of colorectal carcinoma (CRC). We conducted a retrospective, observational study of a series of consecutive patients who underwent CRC resection. Univariate and multivariate analyses of survival curves were performed in patients with and without pre-treatment hypoalbuminemia (<3.5g/dl), both in the overall group of patients and in the subgroup of those with pTNM stage ii tumors. In addition, we compared the 5-year tumor-related mortality in patients with and without hypoalbuminemia. A total of 207 patients were reviewed (median follow-up: 81 months). In the overall multivariate analysis, survival curves were better in patients with normal albumin levels than in those with hypoalbuminemia (HR=2.82; CI 95%=[1.54-5.19]; P=.001). This better prognostic value of normal albumin levels was also significant in pTNM stage ii tumors: (HR=3.76; CI 95%=[1.40-10.08]; P=.009). The 5-year mortality index was lower in patients with normal albumin levels: overall series=18.8% vs 42.9% (OR=3.24; CI 95%=[1.48-7.12]; p=0.001); pTNM stage ii=13.3% vs 44.4% (OR=5.2; CI 95%=[1.36-20.34]; P=0.004). Pre-treatment hypoalbuminemia (<3.5g/dl) was independently related to shorter survival after tumor resection, both in the overall series of patients and in pTNM stage ii CRC. If these results are confirmed, hypoalbuminemia would be a simple and significant marker of poor prognosis, available at the initial diagnosis. Copyright © 2013 Elsevier España, S.L. and AEEH y AEG. All rights reserved.
Barton, Mitch; Yeatts, Paul E; Henson, Robin K; Martin, Scott B
2016-12-01
There has been a recent call to improve data reporting in kinesiology journals, including the appropriate use of univariate and multivariate analysis techniques. For example, a multivariate analysis of variance (MANOVA) with univariate post hocs and a Bonferroni correction is frequently used to investigate group differences on multiple dependent variables. However, this univariate approach decreases power, increases the risk for Type 1 error, and contradicts the rationale for conducting multivariate tests in the first place. The purpose of this study was to provide a user-friendly primer on conducting descriptive discriminant analysis (DDA), which is a post-hoc strategy to MANOVA that takes into account the complex relationships among multiple dependent variables. A real-world example using the Statistical Package for the Social Sciences syntax and data from 1,095 middle school students on their body composition and body image are provided to explain and interpret the results from DDA. While univariate post hocs increased the risk for Type 1 error to 76%, the DDA identified which dependent variables contributed to group differences and which groups were different from each other. For example, students in the very lean and Healthy Fitness Zone categories for body mass index experienced less pressure to lose weight, more satisfaction with their body, and higher physical self-concept than the Needs Improvement Zone groups. However, perceived pressure to gain weight did not contribute to group differences because it was a suppressor variable. Researchers are encouraged to use DDA when investigating group differences on multiple correlated dependent variables to determine which variables contributed to group differences.
Duffy, Sonia A; Ronis, David L; McLean, Scott; Fowler, Karen E; Gruber, Stephen B; Wolf, Gregory T; Terrell, Jeffrey E
2009-04-20
Our prior work has shown that the health behaviors of head and neck cancer patients are interrelated and are associated with quality of life; however, other than smoking, the relationship between health behaviors and survival is unclear. A prospective cohort study was conducted to determine the relationship between five pretreatment health behaviors (smoking, alcohol, diet, physical activity, and sleep) and all-cause survival among 504 head and neck cancer patients. Smoking status was the strongest predictor of survival, with both current smokers (hazard ratio [HR] = 2.4; 95% CI, 1.3 to 4.4) and former smokers (HR = 2.0; 95% CI, 1.2 to 3.5) showing significant associations with poor survival. Problem drinking was associated with survival in the univariate analysis (HR = 1.4; 95% CI, 1.0 to 2.0) but lost significance when controlling for other factors. Low fruit intake was negatively associated with survival in the univariate analysis only (HR = 1.6; 95% CI, 1.1 to 2.1), whereas vegetable intake was not significant in either univariate or multivariate analyses. Although physical activity was associated with survival in the univariate analysis (HR = 0.95; 95% CI, 0.93 to 0.97), it was not significant in the multivariate model. Sleep was not significantly associated with survival in either univariate or multivariate analysis. Control variables that were also independently associated with survival in the multivariate analysis were age, education, tumor site, cancer stage, and surgical treatment. Variation in selected pretreatment health behaviors (eg, smoking, fruit intake, and physical activity) in this population is associated with variation in survival.
Modified DTW for a quantitative estimation of the similarity between rainfall time series
NASA Astrophysics Data System (ADS)
Djallel Dilmi, Mohamed; Barthès, Laurent; Mallet, Cécile; Chazottes, Aymeric
2017-04-01
The Precipitations are due to complex meteorological phenomenon and can be described as intermittent process. The spatial and temporal variability of this phenomenon is significant and covers large scales. To analyze and model this variability and / or structure, several studies use a network of rain gauges providing several time series of precipitation measurements. To compare these different time series, the authors compute for each time series some parameters (PDF, rain peak intensity, occurrence, amount, duration, intensity …). However, and despite the calculation of these parameters, the comparison of the parameters between two series of measurements remains qualitative. Due to the advection processes, when different sensors of an observation network measure precipitation time series identical in terms of intermitency or intensities, there is a time lag between the different measured series. Analyzing and extracting relevant information on physical phenomena from these precipitation time series implies the development of automatic analytical methods capable of comparing two time series of precipitation measured by different sensors or at two different locations and thus quantifying the difference / similarity. The limits of the Euclidean distance to measure the similarity between the time series of precipitation have been well demonstrated and explained (eg the Euclidian distance is indeed very sensitive to the effects of phase shift : between two identical but slightly shifted time series, this distance is not negligible). To quantify and analysis these time lag, the correlation functions are well established, normalized and commonly used to measure the spatial dependences that are required by many applications. However, authors generally observed that there is always a considerable scatter of the inter-rain gauge correlation coefficients obtained from the individual pairs of rain gauges. Because of a substantial dispersion of estimated time lag, the interpretation of this inter-correlation is not straightforward. We propose here to use an improvement of the Euclidian distance which integrates the global complexity of the rainfall series. The Dynamic Time Wrapping (DTW) used in speech recognition allows matching two time series instantly different and provide the most probable time lag. However, the original formulation of the DTW suffers from some limitations. In particular, it is not adequate to the rain intermittency. In this study we present an adaptation of the DTW for the analysis of rainfall time series : we used time series from the "Météo France" rain gauge network observed between January 1st, 2007 and December 31st, 2015 on 25 stations located in the Île de France area. Then we analyze the results (eg. The distance, the relationship between the time lag detected by our methods and others measured parameters like speed and direction of the wind…) to show the ability of the proposed similarity to provide usefull information on the rain structure. The possibility of using this measure of similarity to define a quality indicator of a sensor integrated into an observation network is also envisaged.
Daily rainfall forecasting for one year in a single run using Singular Spectrum Analysis
NASA Astrophysics Data System (ADS)
Unnikrishnan, Poornima; Jothiprakash, V.
2018-06-01
Effective modelling and prediction of smaller time step rainfall is reported to be very difficult owing to its highly erratic nature. Accurate forecast of daily rainfall for longer duration (multi time step) may be exceptionally helpful in the efficient planning and management of water resources systems. Identification of inherent patterns in a rainfall time series is also important for an effective water resources planning and management system. In the present study, Singular Spectrum Analysis (SSA) is utilized to forecast the daily rainfall time series pertaining to Koyna watershed in Maharashtra, India, for 365 days after extracting various components of the rainfall time series such as trend, periodic component, noise and cyclic component. In order to forecast the time series for longer time step (365 days-one window length), the signal and noise components of the time series are forecasted separately and then added together. The results of the study show that the method of SSA could extract the various components of the time series effectively and could also forecast the daily rainfall time series for longer duration such as one year in a single run with reasonable accuracy.
Cross-correlation of point series using a new method
NASA Technical Reports Server (NTRS)
Strothers, Richard B.
1994-01-01
Traditional methods of cross-correlation of two time series do not apply to point time series. Here, a new method, devised specifically for point series, utilizes a correlation measure that is based in the rms difference (or, alternatively, the median absolute difference) between nearest neightbors in overlapped segments of the two series. Error estimates for the observed locations of the points, as well as a systematic shift of one series with respect to the other to accommodate a constant, but unknown, lead or lag, are easily incorporated into the analysis using Monte Carlo techniques. A methodological restriction adopted here is that one series be treated as a template series against which the other, called the target series, is cross-correlated. To estimate a significance level for the correlation measure, the adopted alternative (null) hypothesis is that the target series arises from a homogeneous Poisson process. The new method is applied to cross-correlating the times of the greatest geomagnetic storms with the times of maximum in the undecennial solar activity cycle.
Self-organising mixture autoregressive model for non-stationary time series modelling.
Ni, He; Yin, Hujun
2008-12-01
Modelling non-stationary time series has been a difficult task for both parametric and nonparametric methods. One promising solution is to combine the flexibility of nonparametric models with the simplicity of parametric models. In this paper, the self-organising mixture autoregressive (SOMAR) network is adopted as a such mixture model. It breaks time series into underlying segments and at the same time fits local linear regressive models to the clusters of segments. In such a way, a global non-stationary time series is represented by a dynamic set of local linear regressive models. Neural gas is used for a more flexible structure of the mixture model. Furthermore, a new similarity measure has been introduced in the self-organising network to better quantify the similarity of time series segments. The network can be used naturally in modelling and forecasting non-stationary time series. Experiments on artificial, benchmark time series (e.g. Mackey-Glass) and real-world data (e.g. numbers of sunspots and Forex rates) are presented and the results show that the proposed SOMAR network is effective and superior to other similar approaches.
Testing the causality of Hawkes processes with time reversal
NASA Astrophysics Data System (ADS)
Cordi, Marcus; Challet, Damien; Muni Toke, Ioane
2018-03-01
We show that univariate and symmetric multivariate Hawkes processes are only weakly causal: the true log-likelihoods of real and reversed event time vectors are almost equal, thus parameter estimation via maximum likelihood only weakly depends on the direction of the arrow of time. In ideal (synthetic) conditions, tests of goodness of parametric fit unambiguously reject backward event times, which implies that inferring kernels from time-symmetric quantities, such as the autocovariance of the event rate, only rarely produce statistically significant fits. Finally, we find that fitting financial data with many-parameter kernels may yield significant fits for both arrows of time for the same event time vector, sometimes favouring the backward time direction. This goes to show that a significant fit of Hawkes processes to real data with flexible kernels does not imply a definite arrow of time unless one tests it.
Qiao, Mingzhou; Zhang, Haifang; Zhou, Chenlong
2015-11-24
To explore the factors affecting the residual stones after percutaneous nephrolithotomy (PCNL) in patients with renal calculus. A retrospective analysis was performed for 1 200 patients who were affected by renal calculus and treated with PCNL between Jan 2008 and May 2014 in People's Hospital of Anyang City. Among those patients, 16 were diagnosed as bilateral renal stone and had two successive operations. The size, location and number of stones, previous history of surgery, the degree of hydronephrosis, urinary infection were included in the univariate analysis. Significant factors in univariate analysis were included in the multivariate analysis to determine factors affecting stone residual. A total of 385 cases developed stone residual after surgery. The overall residual rate was 31.7%. In univariate analysis, renal pelvis combined with caliceal calculus (P=0.006), stone size larger than 4 cm (P=0.005), stone number more than 4 (P=0.002), the amount of bleeding more than 200 ml (P=0.025), operation time longer than 120 minutes (P=0.028) were associated with an increased rate of stone residual. When subjected to the Cox multivariate analysis, the independent risk factors for residual stones were renal pelvis combined with caliceal calculus (P=0.049), stone size larger than 4 cm (P=0.038) and stone number more than 4 (P=0.018). Factors affecting the incidence of residual stones after PCNL are the size, location and number of stones. Larger size stone and the presence of renal pelvis combined with caliceal calculus are significantly associated with residual stones. Nevertheless, stone number less than 4 indicates an increased stone clearance rate.
Wang, Yinqing; Cai, Ranze; Wang, Rui; Wang, Chunhua; Chen, Chunmei
2018-06-01
This is a retrospective study.The aim of this study was to illustrate the survival outcomes of patients with classic ependymoma (CE) and identify potential prognostic factors.CE is the most common category of spinal ependymomas, but few published studies have discussed predictors of the survival outcome.A Boolean search of the PubMed, Embase, and OVID databases was conducted by 2 investigators independently. The objects were intramedullary grade II ependymoma according to 2007 WHO classification. Univariate Kaplan-Meier analysis and Log-Rank tests were performed to identify variables associated with progression-free survival (PFS) or overall survival (OS). Multivariate Cox regression was performed to assess hazard ratios (HRs) with 95% confidence intervals (95% CIs). Statistical analysis was performed by SPSS version 23.0 (IBM Corp.) with statistical significance defined as P < .05.A total of 35 studies were identified, including 169 cases of CE. The mean follow-up time across cases was 64.2 ± 51.5 months. Univariate analysis showed that patients who had undergone total resection (TR) had better PFS and OS than those with subtotal resection (STR) and biopsy (P = .002, P = .004, respectively). Within either univariate or multivariate analysis (P = .000, P = .07, respectively), histological type was an independent prognostic factor for PFS of CE [papillary type: HR 0.002, 95% CI (0.000-0.073), P = .001, tanycytic type: HR 0.010, 95% CI (0.000-0.218), P = .003].It was the first integrative analysis of CE to elucidate the correlation between kinds of factors and prognostic outcomes. Definite histological type and safely TR were foundation of CE's management. 4.
Huynh-Thu, Vân Anh; Saeys, Yvan; Wehenkel, Louis; Geurts, Pierre
2012-07-01
Univariate statistical tests are widely used for biomarker discovery in bioinformatics. These procedures are simple, fast and their output is easily interpretable by biologists but they can only identify variables that provide a significant amount of information in isolation from the other variables. As biological processes are expected to involve complex interactions between variables, univariate methods thus potentially miss some informative biomarkers. Variable relevance scores provided by machine learning techniques, however, are potentially able to highlight multivariate interacting effects, but unlike the p-values returned by univariate tests, these relevance scores are usually not statistically interpretable. This lack of interpretability hampers the determination of a relevance threshold for extracting a feature subset from the rankings and also prevents the wide adoption of these methods by practicians. We evaluated several, existing and novel, procedures that extract relevant features from rankings derived from machine learning approaches. These procedures replace the relevance scores with measures that can be interpreted in a statistical way, such as p-values, false discovery rates, or family wise error rates, for which it is easier to determine a significance level. Experiments were performed on several artificial problems as well as on real microarray datasets. Although the methods differ in terms of computing times and the tradeoff, they achieve in terms of false positives and false negatives, some of them greatly help in the extraction of truly relevant biomarkers and should thus be of great practical interest for biologists and physicians. As a side conclusion, our experiments also clearly highlight that using model performance as a criterion for feature selection is often counter-productive. Python source codes of all tested methods, as well as the MATLAB scripts used for data simulation, can be found in the Supplementary Material.
Time Series Model Identification by Estimating Information.
1982-11-01
principle, Applications of Statistics, P. R. Krishnaiah , ed., North-Holland: Amsterdam, 27-41. Anderson, T. W. (1971). The Statistical Analysis of Time Series...E. (1969). Multiple Time Series Modeling, Multivariate Analysis II, edited by P. Krishnaiah , Academic Press: New York, 389-409. Parzen, E. (1981...Newton, H. J. (1980). Multiple Time Series Modeling, II Multivariate Analysis - V, edited by P. Krishnaiah , North Holland: Amsterdam, 181-197. Shibata, R
NASA Astrophysics Data System (ADS)
Baldysz, Zofia; Nykiel, Grzegorz; Figurski, Mariusz; Szafranek, Karolina; Kroszczynski, Krzysztof; Araszkiewicz, Andrzej
2015-04-01
In recent years, the GNSS system began to play an increasingly important role in the research related to the climate monitoring. Based on the GPS system, which has the longest operational capability in comparison with other systems, and a common computational strategy applied to all observations, long and homogeneous ZTD (Zenith Tropospheric Delay) time series were derived. This paper presents results of analysis of 16-year ZTD time series obtained from the EPN (EUREF Permanent Network) reprocessing performed by the Military University of Technology. To maintain the uniformity of data, analyzed period of time (1998-2013) is exactly the same for all stations - observations carried out before 1998 were removed from time series and observations processed using different strategy were recalculated according to the MUT LAC approach. For all 16-year time series (59 stations) Lomb-Scargle periodograms were created to obtain information about the oscillations in ZTD time series. Due to strong annual oscillations which disturb the character of oscillations with smaller amplitude and thus hinder their investigation, Lomb-Scargle periodograms for time series with the deleted annual oscillations were created in order to verify presence of semi-annual, ter-annual and quarto-annual oscillations. Linear trend and seasonal components were estimated using LSE (Least Square Estimation) and Mann-Kendall trend test were used to confirm the presence of linear trend designated by LSE method. In order to verify the effect of the length of time series on the estimated size of the linear trend, comparison between two different length of ZTD time series was performed. To carry out a comparative analysis, 30 stations which have been operating since 1996 were selected. For these stations two periods of time were analyzed: shortened 16-year (1998-2013) and full 18-year (1996-2013). For some stations an additional two years of observations have significant impact on changing the size of linear trend - only for 4 stations the size of linear trend was exactly the same for two periods of time. In one case, the nature of the trend has changed from negative (16-year time series) for positive (18-year time series). The average value of a linear trends for 16-year time series is 1,5 mm/decade, but their spatial distribution is not uniform. The average value of linear trends for all 18-year time series is 2,0 mm/decade, with better spatial distribution and smaller discrepancies.
Umesh P. Agarwal; Richard S. Reiner; Sally A. Ralph
2010-01-01
Two new methods based on FTâRaman spectroscopy, one simple, based on band intensity ratio, and the other using a partial least squares (PLS) regression model, are proposed to determine cellulose I crystallinity. In the simple method, crystallinity in cellulose I samples was determined based on univariate regression that was first developed using the Raman band...
Regression Is a Univariate General Linear Model Subsuming Other Parametric Methods as Special Cases.
ERIC Educational Resources Information Center
Vidal, Sherry
Although the concept of the general linear model (GLM) has existed since the 1960s, other univariate analyses such as the t-test and the analysis of variance models have remained popular. The GLM produces an equation that minimizes the mean differences of independent variables as they are related to a dependent variable. From a computer printout…
Using First Differences to Reduce Inhomogeneity in Radiosonde Temperature Datasets.
NASA Astrophysics Data System (ADS)
Free, Melissa; Angell, James K.; Durre, Imke; Lanzante, John; Peterson, Thomas C.; Seidel, Dian J.
2004-11-01
The utility of a “first difference” method for producing temporally homogeneous large-scale mean time series is assessed. Starting with monthly averages, the method involves dropping data around the time of suspected discontinuities and then calculating differences in temperature from one year to the next, resulting in a time series of year-to-year differences for each month at each station. These first difference time series are then combined to form large-scale means, and mean temperature time series are constructed from the first difference series. When applied to radiosonde temperature data, the method introduces random errors that decrease with the number of station time series used to create the large-scale time series and increase with the number of temporal gaps in the station time series. Root-mean-square errors for annual means of datasets produced with this method using over 500 stations are estimated at no more than 0.03 K, with errors in trends less than 0.02 K decade-1 for 1960 97 at 500 mb. For a 50-station dataset, errors in trends in annual global means introduced by the first differencing procedure may be as large as 0.06 K decade-1 (for six breaks per series), which is greater than the standard error of the trend. Although the first difference method offers significant resource and labor advantages over methods that attempt to adjust the data, it introduces an error in large-scale mean time series that may be unacceptable in some cases.
NASA Astrophysics Data System (ADS)
Riad, Safaa M.; Salem, Hesham; Elbalkiny, Heba T.; Khattab, Fatma I.
2015-04-01
Five, accurate, precise, and sensitive univariate and multivariate spectrophotometric methods were developed for the simultaneous determination of a ternary mixture containing Trimethoprim (TMP), Sulphamethoxazole (SMZ) and Oxytetracycline (OTC) in waste water samples collected from different cites either production wastewater or livestock wastewater after their solid phase extraction using OASIS HLB cartridges. In univariate methods OTC was determined at its λmax 355.7 nm (0D), while (TMP) and (SMZ) were determined by three different univariate methods. Method (A) is based on successive spectrophotometric resolution technique (SSRT). The technique starts with the ratio subtraction method followed by ratio difference method for determination of TMP and SMZ. Method (B) is successive derivative ratio technique (SDR). Method (C) is mean centering of the ratio spectra (MCR). The developed multivariate methods are principle component regression (PCR) and partial least squares (PLS). The specificity of the developed methods is investigated by analyzing laboratory prepared mixtures containing different ratios of the three drugs. The obtained results are statistically compared with those obtained by the official methods, showing no significant difference with respect to accuracy and precision at p = 0.05.
Riad, Safaa M; Salem, Hesham; Elbalkiny, Heba T; Khattab, Fatma I
2015-04-05
Five, accurate, precise, and sensitive univariate and multivariate spectrophotometric methods were developed for the simultaneous determination of a ternary mixture containing Trimethoprim (TMP), Sulphamethoxazole (SMZ) and Oxytetracycline (OTC) in waste water samples collected from different cites either production wastewater or livestock wastewater after their solid phase extraction using OASIS HLB cartridges. In univariate methods OTC was determined at its λmax 355.7 nm (0D), while (TMP) and (SMZ) were determined by three different univariate methods. Method (A) is based on successive spectrophotometric resolution technique (SSRT). The technique starts with the ratio subtraction method followed by ratio difference method for determination of TMP and SMZ. Method (B) is successive derivative ratio technique (SDR). Method (C) is mean centering of the ratio spectra (MCR). The developed multivariate methods are principle component regression (PCR) and partial least squares (PLS). The specificity of the developed methods is investigated by analyzing laboratory prepared mixtures containing different ratios of the three drugs. The obtained results are statistically compared with those obtained by the official methods, showing no significant difference with respect to accuracy and precision at p=0.05. Copyright © 2015 Elsevier B.V. All rights reserved.
Carleton, W. Christopher; Campbell, David
2018-01-01
Statistical time-series analysis has the potential to improve our understanding of human-environment interaction in deep time. However, radiocarbon dating—the most common chronometric technique in archaeological and palaeoenvironmental research—creates challenges for established statistical methods. The methods assume that observations in a time-series are precisely dated, but this assumption is often violated when calibrated radiocarbon dates are used because they usually have highly irregular uncertainties. As a result, it is unclear whether the methods can be reliably used on radiocarbon-dated time-series. With this in mind, we conducted a large simulation study to investigate the impact of chronological uncertainty on a potentially useful time-series method. The method is a type of regression involving a prediction algorithm called the Poisson Exponentially Weighted Moving Average (PEMWA). It is designed for use with count time-series data, which makes it applicable to a wide range of questions about human-environment interaction in deep time. Our simulations suggest that the PEWMA method can often correctly identify relationships between time-series despite chronological uncertainty. When two time-series are correlated with a coefficient of 0.25, the method is able to identify that relationship correctly 20–30% of the time, providing the time-series contain low noise levels. With correlations of around 0.5, it is capable of correctly identifying correlations despite chronological uncertainty more than 90% of the time. While further testing is desirable, these findings indicate that the method can be used to test hypotheses about long-term human-environment interaction with a reasonable degree of confidence. PMID:29351329
Carleton, W Christopher; Campbell, David; Collard, Mark
2018-01-01
Statistical time-series analysis has the potential to improve our understanding of human-environment interaction in deep time. However, radiocarbon dating-the most common chronometric technique in archaeological and palaeoenvironmental research-creates challenges for established statistical methods. The methods assume that observations in a time-series are precisely dated, but this assumption is often violated when calibrated radiocarbon dates are used because they usually have highly irregular uncertainties. As a result, it is unclear whether the methods can be reliably used on radiocarbon-dated time-series. With this in mind, we conducted a large simulation study to investigate the impact of chronological uncertainty on a potentially useful time-series method. The method is a type of regression involving a prediction algorithm called the Poisson Exponentially Weighted Moving Average (PEMWA). It is designed for use with count time-series data, which makes it applicable to a wide range of questions about human-environment interaction in deep time. Our simulations suggest that the PEWMA method can often correctly identify relationships between time-series despite chronological uncertainty. When two time-series are correlated with a coefficient of 0.25, the method is able to identify that relationship correctly 20-30% of the time, providing the time-series contain low noise levels. With correlations of around 0.5, it is capable of correctly identifying correlations despite chronological uncertainty more than 90% of the time. While further testing is desirable, these findings indicate that the method can be used to test hypotheses about long-term human-environment interaction with a reasonable degree of confidence.
Short Term Rain Prediction For Sustainability of Tanks in the Tropic Influenced by Shadow Rains
NASA Astrophysics Data System (ADS)
Suresh, S.
2007-07-01
Rainfall and flow prediction, adapting the Venkataraman single time series approach and Wiener multiple time series approach were conducted for Aralikottai tank system, and Kothamangalam tank system, Tamilnadu, India. The results indicated that the raw prediction of daily values is closer to actual values than trend identified predictions. The sister seasonal time series were more amenable for prediction than whole parent time series. Venkataraman single time approach was more suited for rainfall prediction. Wiener approach proved better for daily prediction of flow based on rainfall. The major conclusion is that the sister seasonal time series of rain and flow have their own identities even though they form part of the whole parent time series. Further studies with other tropical small watersheds are necessary to establish this unique characteristic of independent but not exclusive behavior of seasonal stationary stochastic processes as compared to parent non stationary stochastic processes.
The Value of Interrupted Time-Series Experiments for Community Intervention Research
Biglan, Anthony; Ary, Dennis; Wagenaar, Alexander C.
2015-01-01
Greater use of interrupted time-series experiments is advocated for community intervention research. Time-series designs enable the development of knowledge about the effects of community interventions and policies in circumstances in which randomized controlled trials are too expensive, premature, or simply impractical. The multiple baseline time-series design typically involves two or more communities that are repeatedly assessed, with the intervention introduced into one community at a time. It is particularly well suited to initial evaluations of community interventions and the refinement of those interventions. This paper describes the main features of multiple baseline designs and related repeated-measures time-series experiments, discusses the threats to internal validity in multiple baseline designs, and outlines techniques for statistical analyses of time-series data. Examples are given of the use of multiple baseline designs in evaluating community interventions and policy changes. PMID:11507793
Rainfall disaggregation for urban hydrology: Effects of spatial consistence
NASA Astrophysics Data System (ADS)
Müller, Hannes; Haberlandt, Uwe
2015-04-01
For urban hydrology rainfall time series with a high temporal resolution are crucial. Observed time series of this kind are very short in most cases, so they cannot be used. On the contrary, time series with lower temporal resolution (daily measurements) exist for much longer periods. The objective is to derive time series with a long duration and a high resolution by disaggregating time series of the non-recording stations with information of time series of the recording stations. The multiplicative random cascade model is a well-known disaggregation model for daily time series. For urban hydrology it is often assumed, that a day consists of only 1280 minutes in total as starting point for the disaggregation process. We introduce a new variant for the cascade model, which is functional without this assumption and also outperforms the existing approach regarding time series characteristics like wet and dry spell duration, average intensity, fraction of dry intervals and extreme value representation. However, in both approaches rainfall time series of different stations are disaggregated without consideration of surrounding stations. This yields in unrealistic spatial patterns of rainfall. We apply a simulated annealing algorithm that has been used successfully for hourly values before. Relative diurnal cycles of the disaggregated time series are resampled to reproduce the spatial dependence of rainfall. To describe spatial dependence we use bivariate characteristics like probability of occurrence, continuity ratio and coefficient of correlation. Investigation area is a sewage system in Northern Germany. We show that the algorithm has the capability to improve spatial dependence. The influence of the chosen disaggregation routine and the spatial dependence on overflow occurrences and volumes of the sewage system will be analyzed.
Clinical time series prediction: Toward a hierarchical dynamical system framework.
Liu, Zitao; Hauskrecht, Milos
2015-09-01
Developing machine learning and data mining algorithms for building temporal models of clinical time series is important for understanding of the patient condition, the dynamics of a disease, effect of various patient management interventions and clinical decision making. In this work, we propose and develop a novel hierarchical framework for modeling clinical time series data of varied length and with irregularly sampled observations. Our hierarchical dynamical system framework for modeling clinical time series combines advantages of the two temporal modeling approaches: the linear dynamical system and the Gaussian process. We model the irregularly sampled clinical time series by using multiple Gaussian process sequences in the lower level of our hierarchical framework and capture the transitions between Gaussian processes by utilizing the linear dynamical system. The experiments are conducted on the complete blood count (CBC) panel data of 1000 post-surgical cardiac patients during their hospitalization. Our framework is evaluated and compared to multiple baseline approaches in terms of the mean absolute prediction error and the absolute percentage error. We tested our framework by first learning the time series model from data for the patients in the training set, and then using it to predict future time series values for the patients in the test set. We show that our model outperforms multiple existing models in terms of its predictive accuracy. Our method achieved a 3.13% average prediction accuracy improvement on ten CBC lab time series when it was compared against the best performing baseline. A 5.25% average accuracy improvement was observed when only short-term predictions were considered. A new hierarchical dynamical system framework that lets us model irregularly sampled time series data is a promising new direction for modeling clinical time series and for improving their predictive performance. Copyright © 2014 Elsevier B.V. All rights reserved.
FT-IR-cPAS—New Photoacoustic Measurement Technique for Analysis of Hot Gases: A Case Study on VOCs
Hirschmann, Christian Bernd; Koivikko, Niina Susanna; Raittila, Jussi; Tenhunen, Jussi; Ojala, Satu; Rahkamaa-Tolonen, Katariina; Marbach, Ralf; Hirschmann, Sarah; Keiski, Riitta Liisa
2011-01-01
This article describes a new photoacoustic FT-IR system capable of operating at elevated temperatures. The key hardware component is an optical-readout cantilever microphone that can work up to 200 °C. All parts in contact with the sample gas were put into a heated oven, incl. the photoacoustic cell. The sensitivity of the built photoacoustic system was tested by measuring 18 different VOCs. At 100 ppm gas concentration, the univariate signal to noise ratios (1σ, measurement time 25.5 min, at highest peak, optical resolution 8 cm−1) of the spectra varied from minimally 19 for o-xylene up to 329 for butyl acetate. The sensitivity can be improved by multivariate analyses over broad wavelength ranges, which effectively co-adds the univariate sensitivities achievable at individual wavelengths. The multivariate limit of detection (3σ, 8.5 min, full useful wavelength range), i.e., the best possible inverse analytical sensitivity achievable at optimum calibration, was calculated using the SBC method and varied from 2.60 ppm for dichloromethane to 0.33 ppm for butyl acetate. Depending on the shape of the spectra, which often only contain a few sharp peaks, the multivariate analysis improved the analytical sensitivity by 2.2 to 9.2 times compared to the univariate case. Selectivity and multi component ability were tested by a SBC calibration including 5 VOCs and water. The average cross selectivities turned out to be less than 2% and the resulting inverse analytical sensitivities of the 5 interfering VOCs was increased by maximum factor of 2.2 compared to the single component sensitivities. Water subtraction using SBC gave the true analyte concentration with a variation coefficient of 3%, although the sample spectra (methyl ethyl ketone, 200 ppm) contained water from 1,400 to 100k ppm and for subtraction only one water spectra (10k ppm) was used. The developed device shows significant improvement to the current state-of-the-art measurement methods used in industrial VOC measurements. PMID:22163900
Vidić, Igor; Egnell, Liv; Jerome, Neil P; Teruel, Jose R; Sjøbakk, Torill E; Østlie, Agnes; Fjøsne, Hans E; Bathen, Tone F; Goa, Pål Erik
2018-05-01
Diffusion-weighted MRI (DWI) is currently one of the fastest developing MRI-based techniques in oncology. Histogram properties from model fitting of DWI are useful features for differentiation of lesions, and classification can potentially be improved by machine learning. To evaluate classification of malignant and benign tumors and breast cancer subtypes using support vector machine (SVM). Prospective. Fifty-one patients with benign (n = 23) and malignant (n = 28) breast tumors (26 ER+, whereof six were HER2+). Patients were imaged with DW-MRI (3T) using twice refocused spin-echo echo-planar imaging with echo time / repetition time (TR/TE) = 9000/86 msec, 90 × 90 matrix size, 2 × 2 mm in-plane resolution, 2.5 mm slice thickness, and 13 b-values. Apparent diffusion coefficient (ADC), relative enhanced diffusivity (RED), and the intravoxel incoherent motion (IVIM) parameters diffusivity (D), pseudo-diffusivity (D*), and perfusion fraction (f) were calculated. The histogram properties (median, mean, standard deviation, skewness, kurtosis) were used as features in SVM (10-fold cross-validation) for differentiation of lesions and subtyping. Accuracies of the SVM classifications were calculated to find the combination of features with highest prediction accuracy. Mann-Whitney tests were performed for univariate comparisons. For benign versus malignant tumors, univariate analysis found 11 histogram properties to be significant differentiators. Using SVM, the highest accuracy (0.96) was achieved from a single feature (mean of RED), or from three feature combinations of IVIM or ADC. Combining features from all models gave perfect classification. No single feature predicted HER2 status of ER + tumors (univariate or SVM), although high accuracy (0.90) was achieved with SVM combining several features. Importantly, these features had to include higher-order statistics (kurtosis and skewness), indicating the importance to account for heterogeneity. Our findings suggest that SVM, using features from a combination of diffusion models, improves prediction accuracy for differentiation of benign versus malignant breast tumors, and may further assist in subtyping of breast cancer. 3 Technical Efficacy: Stage 3 J. Magn. Reson. Imaging 2018;47:1205-1216. © 2017 International Society for Magnetic Resonance in Medicine.
Update on clinically isolated syndrome.
Thouvenot, Éric
2015-04-01
Optic neuritis, myelitis and brainstem syndrome accompanied by a symptomatic MRI T2 or FLAIR hyperintensity and T1 hypointensity are highly suggestive of multiple sclerosis (MS) in young adults. They are called "clinically isolated syndrome" (CIS) and correspond to the typical first multiple sclerosis (MS) episode, especially when associated with other asymptomatic demyelinating lesions, without clinical, radiological and immunological sign of differential diagnosis. After a CIS, the delay of apparition of a relapse, which corresponds to the conversion to clinically definite MS (CDMS), varies from several months to more than 10 years (10-15% of cases, generally called benign RRMS). This delay is generally associated with the number and location of demyelinating lesions of the brain and spinal cord and the results of CSF analysis. Several studies comparing different MRI criteria for dissemination in space and dissemination in time of demyelinating lesions, two hallmarks of MS, provided enough substantial data to update diagnostic criteria for MS after a CIS. In the last revision of the McDonald's criteria in 2010, diagnostic criteria were simplified and now the diagnosis can be made by a single initial scan that proves the presence of active asymptomatic lesions (with gadolinium enhancement) and of unenhanced lesions. However, time to conversion remains highly unpredictable for a given patient and CIS can remain isolated, especially for idiopathic unilateral optic neuritis or myelitis. Univariate analyses of clinical, radiological, biological or electrophysiological characteristics of CIS patients in small series identified numerous risk factors of rapid conversion to MS. However, large series of CIS patients analyzing several characteristics of CIS patients and the influence of disease modifying therapies brought important information about the risk of CDMS or RRMS over up to 20 years of follow-up. They confirmed the importance of the initial MRI pattern of demyelinating lesions and of CSF oligoclonal bands. Available treatments of MS (immunomodulators or immunosuppressants) have also shown unequivocal efficacy to slow the conversion to RRMS after a CIS, but they could be unnecessary for patients with benign RRMS. Beyond diagnostic criteria, knowledge of established and potential risk factors of conversion to MS and of disability progression is essential for CIS patients' follow-up and initiation of disease modifying therapies. Copyright © 2015 Elsevier Masson SAS. All rights reserved.
NASA Astrophysics Data System (ADS)
Nakada, Tomohiro; Takadama, Keiki; Watanabe, Shigeyoshi
This paper proposes the classification method using Bayesian analytical method to classify the time series data in the international emissions trading market depend on the agent-based simulation and compares the case with Discrete Fourier transform analytical method. The purpose demonstrates the analytical methods mapping time series data such as market price. These analytical methods have revealed the following results: (1) the classification methods indicate the distance of mapping from the time series data, it is easier the understanding and inference than time series data; (2) these methods can analyze the uncertain time series data using the distance via agent-based simulation including stationary process and non-stationary process; and (3) Bayesian analytical method can show the 1% difference description of the emission reduction targets of agent.
How long will the traffic flow time series keep efficacious to forecast the future?
NASA Astrophysics Data System (ADS)
Yuan, PengCheng; Lin, XuXun
2017-02-01
This paper investigate how long will the historical traffic flow time series keep efficacious to forecast the future. In this frame, we collect the traffic flow time series data with different granularity at first. Then, using the modified rescaled range analysis method, we analyze the long memory property of the traffic flow time series by computing the Hurst exponent. We calculate the long-term memory cycle and test its significance. We also compare it with the maximum Lyapunov exponent method result. Our results show that both of the freeway traffic flow time series and the ground way traffic flow time series demonstrate positively correlated trend (have long-term memory property), both of their memory cycle are about 30 h. We think this study is useful for the short-term or long-term traffic flow prediction and management.
Measurements of spatial population synchrony: influence of time series transformations.
Chevalier, Mathieu; Laffaille, Pascal; Ferdy, Jean-Baptiste; Grenouillet, Gaël
2015-09-01
Two mechanisms have been proposed to explain spatial population synchrony: dispersal among populations, and the spatial correlation of density-independent factors (the "Moran effect"). To identify which of these two mechanisms is driving spatial population synchrony, time series transformations (TSTs) of abundance data have been used to remove the signature of one mechanism, and highlight the effect of the other. However, several issues with TSTs remain, and to date no consensus has emerged about how population time series should be handled in synchrony studies. Here, by using 3131 time series involving 34 fish species found in French rivers, we computed several metrics commonly used in synchrony studies to determine whether a large-scale climatic factor (temperature) influenced fish population dynamics at the regional scale, and to test the effect of three commonly used TSTs (detrending, prewhitening and a combination of both) on these metrics. We also tested whether the influence of TSTs on time series and population synchrony levels was related to the features of the time series using both empirical and simulated time series. For several species, and regardless of the TST used, we evidenced a Moran effect on freshwater fish populations. However, these results were globally biased downward by TSTs which reduced our ability to detect significant signals. Depending on the species and the features of the time series, we found that TSTs could lead to contradictory results, regardless of the metric considered. Finally, we suggest guidelines on how population time series should be processed in synchrony studies.
Transition Icons for Time-Series Visualization and Exploratory Analysis.
Nickerson, Paul V; Baharloo, Raheleh; Wanigatunga, Amal A; Manini, Todd M; Tighe, Patrick J; Rashidi, Parisa
2018-03-01
The modern healthcare landscape has seen the rapid emergence of techniques and devices that temporally monitor and record physiological signals. The prevalence of time-series data within the healthcare field necessitates the development of methods that can analyze the data in order to draw meaningful conclusions. Time-series behavior is notoriously difficult to intuitively understand due to its intrinsic high-dimensionality, which is compounded in the case of analyzing groups of time series collected from different patients. Our framework, which we call transition icons, renders common patterns in a visual format useful for understanding the shared behavior within groups of time series. Transition icons are adept at detecting and displaying subtle differences and similarities, e.g., between measurements taken from patients receiving different treatment strategies or stratified by demographics. We introduce various methods that collectively allow for exploratory analysis of groups of time series, while being free of distribution assumptions and including simple heuristics for parameter determination. Our technique extracts discrete transition patterns from symbolic aggregate approXimation representations, and compiles transition frequencies into a bag of patterns constructed for each group. These transition frequencies are normalized and aligned in icon form to intuitively display the underlying patterns. We demonstrate the transition icon technique for two time-series datasets-postoperative pain scores, and hip-worn accelerometer activity counts. We believe transition icons can be an important tool for researchers approaching time-series data, as they give rich and intuitive information about collective time-series behaviors.
Relating the large-scale structure of time series and visibility networks.
Rodríguez, Miguel A
2017-06-01
The structure of time series is usually characterized by means of correlations. A new proposal based on visibility networks has been considered recently. Visibility networks are complex networks mapped from surfaces or time series using visibility properties. The structures of time series and visibility networks are closely related, as shown by means of fractional time series in recent works. In these works, a simple relationship between the Hurst exponent H of fractional time series and the exponent of the distribution of edges γ of the corresponding visibility network, which exhibits a power law, is shown. To check and generalize these results, in this paper we delve into this idea of connected structures by defining both structures more properly. In addition to the exponents used before, H and γ, which take into account local properties, we consider two more exponents that, as we will show, characterize global properties. These are the exponent α for time series, which gives the scaling of the variance with the size as var∼T^{2α}, and the exponent κ of their corresponding network, which gives the scaling of the averaged maximum of the number of edges, 〈k_{M}〉∼N^{κ}. With this representation, a more precise connection between the structures of general time series and their associated visibility network is achieved. Similarities and differences are more clearly established, and new scaling forms of complex networks appear in agreement with their respective classes of time series.
Short-term vs. long-term heart rate variability in ischemic cardiomyopathy risk stratification.
Voss, Andreas; Schroeder, Rico; Vallverdú, Montserrat; Schulz, Steffen; Cygankiewicz, Iwona; Vázquez, Rafael; Bayés de Luna, Antoni; Caminal, Pere
2013-01-01
In industrialized countries with aging populations, heart failure affects 0.3-2% of the general population. The investigation of 24 h-ECG recordings revealed the potential of nonlinear indices of heart rate variability (HRV) for enhanced risk stratification in patients with ischemic heart failure (IHF). However, long-term analyses are time-consuming, expensive, and delay the initial diagnosis. The objective of this study was to investigate whether 30 min short-term HRV analysis is sufficient for comparable risk stratification in IHF in comparison to 24 h-HRV analysis. From 256 IHF patients [221 at low risk (IHFLR) and 35 at high risk (IHFHR)] (a) 24 h beat-to-beat time series (b) the first 30 min segment (c) the 30 min most stationary day segment and (d) the 30 min most stationary night segment were investigated. We calculated linear (time and frequency domain) and nonlinear HRV analysis indices. Optimal parameter sets for risk stratification in IHF were determined for 24 h and for each 30 min segment by applying discriminant analysis on significant clinical and non-clinical indices. Long- and short-term HRV indices from frequency domain and particularly from nonlinear dynamics revealed high univariate significances (p < 0.01) discriminating between IHFLR and IHFHR. For multivariate risk stratification, optimal mixed parameter sets consisting of 5 indices (clinical and nonlinear) achieved 80.4% AUC (area under the curve of receiver operating characteristics) from 24 h HRV analysis, 84.3% AUC from first 30 min, 82.2 % AUC from daytime 30 min and 81.7% AUC from nighttime 30 min. The optimal parameter set obtained from the first 30 min showed nearly the same classification power when compared to the optimal 24 h-parameter set. As results from stationary daytime and nighttime, 30 min segments indicate that short-term analyses of 30 min may provide at least a comparable risk stratification power in IHF in comparison to a 24 h analysis period.
Clinical peripherality: development of a peripherality index for rural health services.
Swan, Gillian M; Selvaraj, Sivasubramaniam; Godden, David J
2008-01-25
The configuration of rural health services is influenced by geography. Rural health practitioners provide a broader range of services to smaller populations scattered over wider areas or more difficult terrain than their urban counterparts. This has implications for training and quality assurance of outcomes. This exploratory study describes the development of a "clinical peripherality" indicator that has potential application to remote and rural general practice communities for planning and research purposes. Profiles of general practice communities in Scotland were created from a variety of public data sources. Four candidate variables were chosen that described demographic and geographic characteristics of each practice: population density, number of patients on the practice list, travel time to nearest specialist led hospital and travel time to Health Board administrative headquarters. A clinical peripherality index, based on these variables, was derived using factor analysis. Relationships between the clinical peripherality index and services offered by the practices and the staff profile of the practices were explored in a series of univariate analyses. Factor analysis on the four candidate variables yielded a robust one-factor solution explaining 75% variance with factor loadings ranging from 0.83 to 0.89. Rural and remote areas had higher median values and a greater scatter of clinical peripherality indices among their practices than an urban comparison area. The range of services offered and the profile of staffing of practices was associated with the peripherality index. Clinical peripherality is determined by the nature of the practice and its location relative to secondary care and administrative and educational facilities. It has features of both gravity model-based and travel time/accessibility indicators and has the potential to be applied to training of staff for rural and remote locations and to other aspects of health policy and planning. It may assist planners in conceptualising the effects on general practices of centralising specialist clinical services or administrative and educational facilities.
Neurological Change after Gamma Knife Radiosurgery for Brain Metastases Involving the Motor Cortex
Park, Chang-Yong; Choi, Hyun-Yong; Lee, Sang-Ryul; Roh, Tae Hoon; Seo, Mi-Ra
2016-01-01
Background Although Gamma Knife radiosurgery (GKRS) can provide beneficial therapeutic effects for patients with brain metastases, lesions involving the eloquent areas carry a higher risk of neurologic deterioration after treatment, compared to those located in the non-eloquent areas. We aimed to investigate neurological change of the patients with brain metastases involving the motor cortex (MC) and the relevant factors related to neurological deterioration after GKRS. Methods We retrospectively reviewed clinical, radiological and dosimetry data of 51 patients who underwent GKRS for 60 brain metastases involving the MC. Prior to GKRS, motor deficits existed in 26 patients (50.9%). The mean target volume was 3.2 cc (range 0.001–14.1) at the time of GKRS, and the mean prescription dose was 18.6 Gy (range 12–24 Gy). Results The actuarial median survival time from GKRS was 19.2±5.0 months. The calculated local tumor control rates at 6 and 12 months after GKRS were 89.7% and 77.4%, respectively. During the median clinical follow-up duration of 12.3±2.6 months (range 1–54 months), 18 patients (35.3%) experienced new or worsened neurologic deficits with a median onset time of 2.5±0.5 months (range 0.3–9.7 months) after GKRS. Among various factors, prescription dose (>20 Gy) was a significant factor for the new or worsened neurologic deficits in univariate (p=0.027) and multivariate (p=0.034) analysis. The managements of 18 patients were steroid medication (n=10), boost radiation therapy (n=5), and surgery (n=3), and neurological improvement was achieved in 9 (50.0%). Conclusion In our series, prescription dose (>20 Gy) was significantly related to neurological deterioration after GKRS for brain metastases involving the MC. Therefore, we suggest that careful dose adjustment would be required for lesions involving the MC to avoid neurological deterioration requiring additional treatment in the patients with limited life expectancy. PMID:27867921
Clinical peripherality: development of a peripherality index for rural health services
Swan, Gillian M; Selvaraj, Sivasubramaniam; Godden, David J
2008-01-01
Background The configuration of rural health services is influenced by geography. Rural health practitioners provide a broader range of services to smaller populations scattered over wider areas or more difficult terrain than their urban counterparts. This has implications for training and quality assurance of outcomes. This exploratory study describes the development of a "clinical peripherality" indicator that has potential application to remote and rural general practice communities for planning and research purposes. Methods Profiles of general practice communities in Scotland were created from a variety of public data sources. Four candidate variables were chosen that described demographic and geographic characteristics of each practice: population density, number of patients on the practice list, travel time to nearest specialist led hospital and travel time to Health Board administrative headquarters. A clinical peripherality index, based on these variables, was derived using factor analysis. Relationships between the clinical peripherality index and services offered by the practices and the staff profile of the practices were explored in a series of univariate analyses. Results Factor analysis on the four candidate variables yielded a robust one-factor solution explaining 75% variance with factor loadings ranging from 0.83 to 0.89. Rural and remote areas had higher median values and a greater scatter of clinical peripherality indices among their practices than an urban comparison area. The range of services offered and the profile of staffing of practices was associated with the peripherality index. Conclusion Clinical peripherality is determined by the nature of the practice and its location relative to secondary care and administrative and educational facilities. It has features of both gravity model-based and travel time/accessibility indicators and has the potential to be applied to training of staff for rural and remote locations and to other aspects of health policy and planning. It may assist planners in conceptualising the effects on general practices of centralising specialist clinical services or administrative and educational facilities. PMID:18221533
Culp, William T N; Olea-Popelka, Francisco; Sefton, Jennifer; Aldridge, Charles F; Withrow, Stephen J; Lafferty, Mary H; Rebhun, Robert B; Kent, Michael S; Ehrhart, Nicole
2014-11-15
To evaluate clinical characteristics, outcome, and prognostic variables in a cohort of dogs surviving > 1 year after an initial diagnosis of osteosarcoma. Retrospective case series. 90 client-owned dogs. Medical records for an 11-year period from 1997 through 2008 were reviewed, and patients with appendicular osteosarcoma that lived > 1 year after initial histopathologic diagnosis were studied. Variables including signalment, weight, serum alkaline phosphatase activity, tumor location, surgery, and adjuvant therapies were recorded. Median survival times were calculated by means of a Kaplan-Meier survival function. Univariate analysis was conducted to compare the survival function for categorical variables, and the Cox proportional hazard model was used to evaluate the likelihood of death > 1 year after diagnosis on the basis of the selected risk factors. 90 dogs met the inclusion criteria; clinical laboratory information was not available in all cases. Median age was 8.2 years (range, 2.7 to 13.3 years), and median weight was 38 kg (83.6 lb; range, 21 to 80 kg [46.2 to 176 lb]). Serum alkaline phosphatase activity was high in 29 of 60 (48%) dogs. The most common tumor location was the distal portion of the radius (54/90 [60%]). Eighty-nine of 90 (99%) dogs underwent surgery, and 78 (87%) received chemotherapy. Overall, 49 of 90 (54%) dogs developed metastatic disease. The median survival time beyond 1 year was 243 days (range, 1 to 1,899 days). Dogs that developed a surgical-site infection after limb-sparing surgery had a significantly improved prognosis > 1 year after osteosarcoma diagnosis, compared with dogs that did not develop infections. Results of the present study indicated that dogs with an initial diagnosis of osteosarcoma that lived > 1 year had a median survival time beyond the initial year of approximately 8 months. As reported previously, the development of a surgical-site infection in dogs undergoing a limb-sparing surgery significantly affected prognosis and warrants further study.
Mineo, Tommaso Claudio; Tamburrini, Alessandro; Schillaci, Orazio; Ambrogi, Vincenzo
2018-03-06
Patients with thymoma and without clinical or electromyographical myasthenic signs may occasionally develop myasthenia several years after thymectomy. Hereby, we investigated the predictors and the evolution of this peculiar disease. We performed a retrospective analysis in 104 consecutive patients who underwent thymectomy between 1987 and 2013 for thymoma without clinical or electromyographic signs of myasthenia gravis. Predictors of post-thymectomy onset of myasthenia gravis were investigated with univariate time-to-disease analysis. Evolution of myasthenia was analyzed with time-to-regression analysis. Eight patients developed late myasthenia gravis after a median period of 33 months from thymectomy. No significant correlation was found for age, gender, Masaoka's stage, and World Health Organization histology. Only high preoperative serum acetylcholine-receptor antibodies titer (>0.3 nmol/L) was significantly associated with post-thymectomy myasthenia gravis at univariate time-to-disease (P = 0.003) analysis. Positron emission tomography was always performed in high-titer patients, and increased metabolic activity was detected in 4 of these patients. Surgical treatment through redo-sternotomy or video-assisted thoracoscopy was performed in these last cases with a remission in all patients after 12, 24, 32 and 48 months, respectively. No patient under medical treatment has yet developed a complete remission. In our study the presence of preoperative high-level serum acetylcholine receptor antibodies was the only factor significantly associated with the development of post-thymectomy myasthenia gravis. The persistence of residual islet of ectopic thymic tissue was one of the causes of the onset of myasthenia and its surgical removal was successful. Copyright © 2018 Elsevier Inc. All rights reserved.
Hydronephrosis in patients with cervical cancer: an assessment of morbidity and survival.
Patel, Krishna; Foster, Nathan R; Kumar, Amanika; Grudem, Megan; Longenbach, Sherri; Bakkum-Gamez, Jamie; Haddock, Michael; Dowdy, Sean; Jatoi, Aminah
2015-05-01
Hydronephrosis is a frequently observed but understudied complication in patients with cervical cancer. To better characterize hydronephrosis in cervical cancer patients, the current study sought (1) to describe hydronephrosis-associated morbidity and (2) to analyze the prognostic effect of hydronephrosis in patients with a broad range of cancer stages over time. The Mayo Clinic Tumor Registry was interrogated for all invasive cervical cancer patients seen at the Mayo Clinic from 2008 through 2013 in Rochester, Minnesota; these patients' medical records were then reviewed in detail. Two hundred seventy-nine cervical cancer patients with a median age of 49 years and a range of cancer stages were included. Sixty-five patients (23 %) were diagnosed with hydronephrosis at some point during their disease course. In univariate analyses, hydronephrosis was associated with advanced cancer stage (p < 0.0001), squamous histology (p = 0.0079), and nonsurgical cancer treatment (p = 0.0039). In multivariate analyses, stage and tumor histology were associated with hydronephrosis. All but one patient underwent stent placement or urinary diversion; hydronephrosis-related morbidity included pain, urinary tract infections, nausea and vomiting, renal failure, and urinary tract bleeding. In landmark univariate survival analyses, hydronephrosis was associated with worse survival at all time points. In landmark multivariate analyses (adjusted for patient age, stage, cancer treatment, and tumor histology), hydronephrosis was associated with a trend toward worse survival over time (hazard ratios ranged from 1.47 to 4.69). Hydronephrosis in cervical cancer patients is associated with notable morbidity. It is also associated with trends toward worse survival-even if it occurs after the original cancer diagnosis.
Weingessel, Birgit; Richter-Mueksch, Sibylla; Vécsei-Marlovits, Pia V
2011-05-01
To evaluate patients’ maximum acceptable waiting time (MAWT) and to assess the determinants of patient perceptions of MAWT. A total of 500 consecutive patients with cataract were asked to fill out a preoperative questionnaire, addressing patients’ MAWT to undergo cataract surgery. Patients’ visual impairment (VF-14 score), education, profession and social status were evaluated, and an ophthalmologic examination was performed. Univariate analysis included Spearman’s correlation test, unpaired Student’s t-test and the Mann–Whitney U test. Univariate and multivariate associations were calculated using unconditional logistic regression. The mean MAWT was 3.17 ± 2.12 months. The mean VF-14 score was 72.10 ± 22.54. Between VF-14 score and MAWT, there was a significant correlation (r = 0.180, p = 0.004). Patients with higher education (high school, university) accepted significantly longer MAWT (3.92 ± 2.38 months versus 3.02 ± 2.00 months, p = 0.009). Patients who had self-noticed visual impairment were nearly four times (OR: 3.88, 95% CI = 2.07–7.28, p < 0.001) more likely to accept only MAWT of <3 months. Patients with low tolerance for waiting had greater self-reported difficulty with vision. Patients’ acceptance of waiting was not associated with clinical visual acuity measures. Education, ability to work, living independently and taking care of dependents were also strong predictors from patients’ perspective. Considering the implementation of standards for waiting lists, these facts should be taken into account. © 2010 The Authors. Journal compilation © 2010 Acta Ophthalmol.
Characteristics and Prognosis of Oldest Old Subjects with Amyotrophic Lateral Sclerosis.
Dandaba, Meira; Couratier, Philippe; Labrunie, Anaïs; Nicol, Marie; Hamidou, Bello; Raymondeau, Marie; Logroscino, Giancarlo; Preux, Pierre Marie; Marin, Benoît
2017-01-01
Amyotrophic Lateral Sclerosis (ALS) is an age-related neurodegenerative disease with unclear characteristics and prognosis in the oldest old (80 years and over). The aim of this study was to compare the oldest old and younger ALS patients in terms of clinical and socio-demographic characteristics, and prognosis. ALS incident cases from the register of ALS in Limousin (FRALim), diagnosed between January 2000 and July 2013, were included. Descriptive and comparative analyses by age group were carried out. For time to event univariate analysis, Kaplan-Meier estimator and log rank test were used. Univariate and multivariate survival analyses were carried out with Cox's proportional hazard model. Out of 322 patients, 50 (15.5%) were aged 80 or over ("oldest old" ALS) at the time of diagnosis. Among them, the male:female gender-ratio was 1.27, and 32.6% had a bulbar onset (not different from subjects aged less than 80 years). With increasing age, there was a worsening of the clinical state of the patients at time of diagnosis in terms of weight loss, forced vital capacity, ALSFRS-R and manual muscular testing. Access to ALS referral centres decreased with age, and the use of riluzole tended to be lower in the oldest old group. The median survival of oldest old patients appeared to be 10 months shorter than that of subjects aged less than 80 years (7.4 vs. 17.4 months). The survival of oldest old ALS patients is particularly short. It relates to prognostic features at baseline and to an independent effect of advanced age. © 2017 S. Karger AG, Basel.
Anthropometry as a predictor of vertical jump heights derived from an instrumented platform.
Caruso, John F; Daily, Jeremy S; Mason, Melissa L; Shepherd, Catherine M; McLagan, Jessica R; Marshall, Mallory R; Walker, Ron H; West, Jason O
2012-01-01
The current study purpose examined the vertical height-anthropometry relationship with jump data obtained from an instrumented platform. Our methods required college-aged (n = 177) subjects to make 3 visits to our laboratory to measure the following anthropometric variables: height, body mass, upper arm length (UAL), lower arm length, upper leg length, and lower leg length. Per jump, maximum height was measured in 3 ways: from the subjects' takeoff, hang times, and as they landed on the platform. Standard multivariate regression assessed how well anthropometry predicted the criterion variance per gender (men, women, pooled) and jump height method (takeoff, hang time, landing) combination. Z-scores indicated that small amounts of the total data were outliers. The results showed that the majority of outliers were from jump heights calculated as women landed on the platform. With the genders pooled, anthropometry predicted a significant (p < 0.05) amount of variance from jump heights calculated from both takeoff and hang time. The anthropometry-vertical jump relationship was not significant from heights calculated as subjects landed on the platform, likely due to the female outliers. Yet anthropometric data of men did predict a significant amount of variance from heights calculated when they landed on the platform; univariate correlations of men's data revealed that UAL was the best predictor. It was concluded that the large sample of men's data led to greater data heterogeneity and a higher univariate correlation. Because of our sample size and data heterogeneity, practical applications suggest that coaches may find our results best predict performance for a variety of college-aged athletes and vertical jump enthusiasts.
A bivariate model for analyzing recurrent multi-type automobile failures
NASA Astrophysics Data System (ADS)
Sunethra, A. A.; Sooriyarachchi, M. R.
2017-09-01
The failure mechanism in an automobile can be defined as a system of multi-type recurrent failures where failures can occur due to various multi-type failure modes and these failures are repetitive such that more than one failure can occur from each failure mode. In analysing such automobile failures, both the time and type of the failure serve as response variables. However, these two response variables are highly correlated with each other since the timing of failures has an association with the mode of the failure. When there are more than one correlated response variables, the fitting of a multivariate model is more preferable than separate univariate models. Therefore, a bivariate model of time and type of failure becomes appealing for such automobile failure data. When there are multiple failure observations pertaining to a single automobile, such data cannot be treated as independent data because failure instances of a single automobile are correlated with each other while failures among different automobiles can be treated as independent. Therefore, this study proposes a bivariate model consisting time and type of failure as responses adjusted for correlated data. The proposed model was formulated following the approaches of shared parameter models and random effects models for joining the responses and for representing the correlated data respectively. The proposed model is applied to a sample of automobile failures with three types of failure modes and up to five failure recurrences. The parametric distributions that were suitable for the two responses of time to failure and type of failure were Weibull distribution and multinomial distribution respectively. The proposed bivariate model was programmed in SAS Procedure Proc NLMIXED by user programming appropriate likelihood functions. The performance of the bivariate model was compared with separate univariate models fitted for the two responses and it was identified that better performance is secured by the bivariate model. The proposed model can be used to determine the time and type of failure that would occur in the automobiles considered here.
Pons-Estel, Guillermo J.; Alarcón, Graciela S.; González, Luis A.; Zhang, Jie; Vilá, Luis M.; Reveille, John D.; McGwin, Gerald
2010-01-01
Objective To determine the features predictive of time-to-integument damage in patients with systemic lupus erythematosus (SLE) from a multiethnic cohort (LUMINA). Methods SLE LUMINA patients (n=580), age ≥16 years, disease duration ≤5 years at baseline (T0), of African American, Hispanic and Caucasian ethnicity were studied. Integument damage was defined per the SLICC damage index (scarring alopecia, extensive skin scarring and skin ulcers lasting at least six months); factors associated with time-to-its occurrence were examined by Cox proportional univariable and multivariable (main model) hazards regression analyses. Two alternative models were also examined; in model 1 all patients, regardless of when integument damage occurred (n=94), were included; in model 2 a time-varying approach (GEE) was employed. Results Thirty-nine (6.7%) of 580 patients developed integument damage over a mean (SD) total disease duration of 5.9 (3.7) years and were included in the main multivariable regression model. After adjusting for discoid rash, nailfold infarcts, photosensitivity and Raynaud’s phenomenon (significant in the univariable analyses), disease activity over time [Hazard ratio (HR)=1.17; 95% Confidence interval (CI) 1.09–1.26)] was associated with a shorter time-to-integument damage whereas hydroxychloroquine use (HR=0.23, 95% CI 0.12–0.47) and Texan-Hispanic (HR=0.35; 95% CI 0.14–0.87) and Caucasian ethnicities (HR=0.37; 95% CI 0.14–0.99) were associated with a longer time. Results of the alternative models were consistent with those of the main model albeit in model 2 the association with hydroxychloroquine was not significant. Conclusions Our data indicate that hydroxychloroquine use is possibly associated with a delay in integument damage development in patients with SLE. PMID:20391486
Estimating Risk of Natural Gas Portfolios by Using GARCH-EVT-Copula Model.
Tang, Jiechen; Zhou, Chao; Yuan, Xinyu; Sriboonchitta, Songsak
2015-01-01
This paper concentrates on estimating the risk of Title Transfer Facility (TTF) Hub natural gas portfolios by using the GARCH-EVT-copula model. We first use the univariate ARMA-GARCH model to model each natural gas return series. Second, the extreme value distribution (EVT) is fitted to the tails of the residuals to model marginal residual distributions. Third, multivariate Gaussian copula and Student t-copula are employed to describe the natural gas portfolio risk dependence structure. Finally, we simulate N portfolios and estimate value at risk (VaR) and conditional value at risk (CVaR). Our empirical results show that, for an equally weighted portfolio of five natural gases, the VaR and CVaR values obtained from the Student t-copula are larger than those obtained from the Gaussian copula. Moreover, when minimizing the portfolio risk, the optimal natural gas portfolio weights are found to be similar across the multivariate Gaussian copula and Student t-copula and different confidence levels.
Estimating Risk of Natural Gas Portfolios by Using GARCH-EVT-Copula Model
Tang, Jiechen; Zhou, Chao; Yuan, Xinyu; Sriboonchitta, Songsak
2015-01-01
This paper concentrates on estimating the risk of Title Transfer Facility (TTF) Hub natural gas portfolios by using the GARCH-EVT-copula model. We first use the univariate ARMA-GARCH model to model each natural gas return series. Second, the extreme value distribution (EVT) is fitted to the tails of the residuals to model marginal residual distributions. Third, multivariate Gaussian copula and Student t-copula are employed to describe the natural gas portfolio risk dependence structure. Finally, we simulate N portfolios and estimate value at risk (VaR) and conditional value at risk (CVaR). Our empirical results show that, for an equally weighted portfolio of five natural gases, the VaR and CVaR values obtained from the Student t-copula are larger than those obtained from the Gaussian copula. Moreover, when minimizing the portfolio risk, the optimal natural gas portfolio weights are found to be similar across the multivariate Gaussian copula and Student t-copula and different confidence levels. PMID:26351652
Malboeuf-Hurtubise, Catherine; Lacourse, Eric; Herba, Catherine; Taylor, Geneviève; Amor, Leila Ben
2017-01-01
Mindfulness-based interventions constitute a promising option to address anxiety and depression in elementary school students. This study evaluated the effect of a mindfulness-based intervention on anxiety and depression in elementary school students with a diagnosis of anxiety or depression disorder. A single-subject experimental A-B-A design was used. Participants were three elementary school students from grades three and four, along with their teacher. Anxiety and depression were measured on 10 occasions at baseline, during the intervention, and at follow-up. Primary hypotheses were tested using a univariate single case multilevel modeling strategy and visual analysis. Following intervention, 2 participants reported improvements on anxiety and depression, while their teachers reported deteriorating scores on these variables. Results from this n-of-1 trial design is consistent with other work suggesting caution with regard to the overall impact and efficacy of mindfulness-based interventions as a universal treatment option for youth. Future research is warranted. PMID:28853297
Trend time-series modeling and forecasting with neural networks.
Qi, Min; Zhang, G Peter
2008-05-01
Despite its great importance, there has been no general consensus on how to model the trends in time-series data. Compared to traditional approaches, neural networks (NNs) have shown some promise in time-series forecasting. This paper investigates how to best model trend time series using NNs. Four different strategies (raw data, raw data with time index, detrending, and differencing) are used to model various trend patterns (linear, nonlinear, deterministic, stochastic, and breaking trend). We find that with NNs differencing often gives meritorious results regardless of the underlying data generating processes (DGPs). This finding is also confirmed by the real gross national product (GNP) series.
Syndromic surveillance system based on near real-time cattle mortality monitoring.
Torres, G; Ciaravino, V; Ascaso, S; Flores, V; Romero, L; Simón, F
2015-05-01
Early detection of an infectious disease incursion will minimize the impact of outbreaks in livestock. Syndromic surveillance based on the analysis of readily available data can enhance traditional surveillance systems and allow veterinary authorities to react in a timely manner. This study was based on monitoring the number of cattle carcasses sent for rendering in the veterinary unit of Talavera de la Reina (Spain). The aim was to develop a system to detect deviations from expected values which would signal unexpected health events. Historical weekly collected dead cattle (WCDC) time series stabilized by the Box-Cox transformation and adjusted by the minimum least squares method were used to build the univariate cycling regression model based on a Fourier transformation. Three different models, according to type of production system, were built to estimate the baseline expected number of WCDC. Two types of risk signals were generated: point risk signals when the observed value was greater than the upper 95% confidence interval of the expected baseline, and cumulative risk signals, generated by a modified cumulative sum algorithm, when the cumulative sums of reported deaths were above the cumulative sum of expected deaths. Data from 2011 were used to prospectively validate the model generating seven risk signals. None of them were correlated to infectious disease events but some coincided, in time, with very high climatic temperatures recorded in the region. The harvest effect was also observed during the first week of the study year. Establishing appropriate risk signal thresholds is a limiting factor of predictive models; it needs to be adjusted based on experience gained during the use of the models. To increase the sensitivity and specificity of the predictions epidemiological interpretation of non-specific risk signals should be complemented by other sources of information. The methodology developed in this study can enhance other existing early detection surveillance systems. Syndromic surveillance based on mortality monitoring can reduce the detection time for certain disease outbreaks associated with mild mortality only detected at regional level. The methodology can be adapted to monitor other parameters routinely collected at farm level which can be influenced by communicable diseases. Copyright © 2015 Elsevier B.V. All rights reserved.
Adaptive time-variant models for fuzzy-time-series forecasting.
Wong, Wai-Keung; Bai, Enjian; Chu, Alice Wai-Ching
2010-12-01
A fuzzy time series has been applied to the prediction of enrollment, temperature, stock indices, and other domains. Related studies mainly focus on three factors, namely, the partition of discourse, the content of forecasting rules, and the methods of defuzzification, all of which greatly influence the prediction accuracy of forecasting models. These studies use fixed analysis window sizes for forecasting. In this paper, an adaptive time-variant fuzzy-time-series forecasting model (ATVF) is proposed to improve forecasting accuracy. The proposed model automatically adapts the analysis window size of fuzzy time series based on the prediction accuracy in the training phase and uses heuristic rules to generate forecasting values in the testing phase. The performance of the ATVF model is tested using both simulated and actual time series including the enrollments at the University of Alabama, Tuscaloosa, and the Taiwan Stock Exchange Capitalization Weighted Stock Index (TAIEX). The experiment results show that the proposed ATVF model achieves a significant improvement in forecasting accuracy as compared to other fuzzy-time-series forecasting models.
Filter-based multiscale entropy analysis of complex physiological time series.
Xu, Yuesheng; Zhao, Liang
2013-08-01
Multiscale entropy (MSE) has been widely and successfully used in analyzing the complexity of physiological time series. We reinterpret the averaging process in MSE as filtering a time series by a filter of a piecewise constant type. From this viewpoint, we introduce filter-based multiscale entropy (FME), which filters a time series to generate multiple frequency components, and then we compute the blockwise entropy of the resulting components. By choosing filters adapted to the feature of a given time series, FME is able to better capture its multiscale information and to provide more flexibility for studying its complexity. Motivated by the heart rate turbulence theory, which suggests that the human heartbeat interval time series can be described in piecewise linear patterns, we propose piecewise linear filter multiscale entropy (PLFME) for the complexity analysis of the time series. Numerical results from PLFME are more robust to data of various lengths than those from MSE. The numerical performance of the adaptive piecewise constant filter multiscale entropy without prior information is comparable to that of PLFME, whose design takes prior information into account.
Characteristics of the transmission of autoregressive sub-patterns in financial time series
NASA Astrophysics Data System (ADS)
Gao, Xiangyun; An, Haizhong; Fang, Wei; Huang, Xuan; Li, Huajiao; Zhong, Weiqiong
2014-09-01
There are many types of autoregressive patterns in financial time series, and they form a transmission process. Here, we define autoregressive patterns quantitatively through an econometrical regression model. We present a computational algorithm that sets the autoregressive patterns as nodes and transmissions between patterns as edges, and then converts the transmission process of autoregressive patterns in a time series into a network. We utilised daily Shanghai (securities) composite index time series to study the transmission characteristics of autoregressive patterns. We found statistically significant evidence that the financial market is not random and that there are similar characteristics between parts and whole time series. A few types of autoregressive sub-patterns and transmission patterns drive the oscillations of the financial market. A clustering effect on fluctuations appears in the transmission process, and certain non-major autoregressive sub-patterns have high media capabilities in the financial time series. Different stock indexes exhibit similar characteristics in the transmission of fluctuation information. This work not only proposes a distinctive perspective for analysing financial time series but also provides important information for investors.
Characteristics of the transmission of autoregressive sub-patterns in financial time series
Gao, Xiangyun; An, Haizhong; Fang, Wei; Huang, Xuan; Li, Huajiao; Zhong, Weiqiong
2014-01-01
There are many types of autoregressive patterns in financial time series, and they form a transmission process. Here, we define autoregressive patterns quantitatively through an econometrical regression model. We present a computational algorithm that sets the autoregressive patterns as nodes and transmissions between patterns as edges, and then converts the transmission process of autoregressive patterns in a time series into a network. We utilised daily Shanghai (securities) composite index time series to study the transmission characteristics of autoregressive patterns. We found statistically significant evidence that the financial market is not random and that there are similar characteristics between parts and whole time series. A few types of autoregressive sub-patterns and transmission patterns drive the oscillations of the financial market. A clustering effect on fluctuations appears in the transmission process, and certain non-major autoregressive sub-patterns have high media capabilities in the financial time series. Different stock indexes exhibit similar characteristics in the transmission of fluctuation information. This work not only proposes a distinctive perspective for analysing financial time series but also provides important information for investors. PMID:25189200
Predicting long-term catchment nutrient export: the use of nonlinear time series models
NASA Astrophysics Data System (ADS)
Valent, Peter; Howden, Nicholas J. K.; Szolgay, Jan; Komornikova, Magda
2010-05-01
After the Second World War the nitrate concentrations in European water bodies changed significantly as the result of increased nitrogen fertilizer use and changes in land use. However, in the last decades, as a consequence of the implementation of nitrate-reducing measures in Europe, the nitrate concentrations in water bodies slowly decrease. This causes that the mean and variance of the observed time series also changes with time (nonstationarity and heteroscedascity). In order to detect changes and properly describe the behaviour of such time series by time series analysis, linear models (such as autoregressive (AR), moving average (MA) and autoregressive moving average models (ARMA)), are no more suitable. Time series with sudden changes in statistical characteristics can cause various problems in the calibration of traditional water quality models and thus give biased predictions. Proper statistical analysis of these non-stationary and heteroscedastic time series with the aim of detecting and subsequently explaining the variations in their statistical characteristics requires the use of nonlinear time series models. This information can be then used to improve the model building and calibration of conceptual water quality model or to select right calibration periods in order to produce reliable predictions. The objective of this contribution is to analyze two long time series of nitrate concentrations of the rivers Ouse and Stour with advanced nonlinear statistical modelling techniques and compare their performance with traditional linear models of the ARMA class in order to identify changes in the time series characteristics. The time series were analysed with nonlinear models with multiple regimes represented by self-exciting threshold autoregressive (SETAR) and Markov-switching models (MSW). The analysis showed that, based on the value of residual sum of squares (RSS) in both datasets, SETAR and MSW models described the time-series better than models of the ARMA class. In most cases the relative improvement of SETAR models against AR models of first order was low ranging between 1% and 4% with the exception of the three-regime model for the River Stour time-series where the improvement was 48.9%. In comparison, the relative improvement of MSW models was between 44.6% and 52.5 for two-regime and from 60.4% to 75% for three-regime models. However, the visual assessment of models plotted against original datasets showed that despite a high value of RSS, some ARMA models could describe the analyzed time-series better than AR, MA and SETAR models with lower values of RSS. In both datasets MSW models provided a very good visual fit describing most of the extreme values.
Documentation of a spreadsheet for time-series analysis and drawdown estimation
Halford, Keith J.
2006-01-01
Drawdowns during aquifer tests can be obscured by barometric pressure changes, earth tides, regional pumping, and recharge events in the water-level record. These stresses can create water-level fluctuations that should be removed from observed water levels prior to estimating drawdowns. Simple models have been developed for estimating unpumped water levels during aquifer tests that are referred to as synthetic water levels. These models sum multiple time series such as barometric pressure, tidal potential, and background water levels to simulate non-pumping water levels. The amplitude and phase of each time series are adjusted so that synthetic water levels match measured water levels during periods unaffected by an aquifer test. Differences between synthetic and measured water levels are minimized with a sum-of-squares objective function. Root-mean-square errors during fitting and prediction periods were compared multiple times at four geographically diverse sites. Prediction error equaled fitting error when fitting periods were greater than or equal to four times prediction periods. The proposed drawdown estimation approach has been implemented in a spreadsheet application. Measured time series are independent so that collection frequencies can differ and sampling times can be asynchronous. Time series can be viewed selectively and magnified easily. Fitting and prediction periods can be defined graphically or entered directly. Synthetic water levels for each observation well are created with earth tides, measured time series, moving averages of time series, and differences between measured and moving averages of time series. Selected series and fitting parameters for synthetic water levels are stored and drawdowns are estimated for prediction periods. Drawdowns can be viewed independently and adjusted visually if an anomaly skews initial drawdowns away from 0. The number of observations in a drawdown time series can be reduced by averaging across user-defined periods. Raw or reduced drawdown estimates can be copied from the spreadsheet application or written to tab-delimited ASCII files.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-07-17
... comparable to the obligations proposed in this filing: Market-Makers % Time % Series Classes CBOE (current........ 60 All classes collectively. PMMs % Time % Series Classes CBOE (current rule) 99% of the time... % Time % Series Classes CBOE (current rule) 99% of the time required to 90% * Class-by-class. provide...
Multiscale multifractal time irreversibility analysis of stock markets
NASA Astrophysics Data System (ADS)
Jiang, Chenguang; Shang, Pengjian; Shi, Wenbin
2016-11-01
Time irreversibility is one of the most important properties of nonstationary time series. Complex time series often demonstrate even multiscale time irreversibility, such that not only the original but also coarse-grained time series are asymmetric over a wide range of scales. We study the multiscale time irreversibility of time series. In this paper, we develop a method called multiscale multifractal time irreversibility analysis (MMRA), which allows us to extend the description of time irreversibility to include the dependence on the segment size and statistical moments. We test the effectiveness of MMRA in detecting multifractality and time irreversibility of time series generated from delayed Henon map and binomial multifractal model. Then we employ our method to the time irreversibility analysis of stock markets in different regions. We find that the emerging market has higher multifractality degree and time irreversibility compared with developed markets. In this sense, the MMRA method may provide new angles in assessing the evolution stage of stock markets.
Wang, Fang; Wang, Lin; Chen, Yuming
2017-08-31
In order to investigate the time-dependent cross-correlations of fine particulate (PM2.5) series among neighboring cities in Northern China, in this paper, we propose a new cross-correlation coefficient, the time-lagged q-L dependent height crosscorrelation coefficient (denoted by p q (τ, L)), which incorporates the time-lag factor and the fluctuation amplitude information into the analogous height cross-correlation analysis coefficient. Numerical tests are performed to illustrate that the newly proposed coefficient ρ q (τ, L) can be used to detect cross-correlations between two series with time lags and to identify different range of fluctuations at which two series possess cross-correlations. Applying the new coefficient to analyze the time-dependent cross-correlations of PM2.5 series between Beijing and the three neighboring cities of Tianjin, Zhangjiakou, and Baoding, we find that time lags between the PM2.5 series with larger fluctuations are longer than those between PM2.5 series withsmaller fluctuations. Our analysis also shows that cross-correlations between the PM2.5 series of two neighboring cities are significant and the time lags between two PM2.5 series of neighboring cities are significantly non-zero. These findings providenew scientific support on the view that air pollution in neighboring cities can affect one another not simultaneously but with a time lag.
D'Amico, E J; Neilands, T B; Zambarano, R
2001-11-01
Although power analysis is an important component in the planning and implementation of research designs, it is often ignored. Computer programs for performing power analysis are available, but most have limitations, particularly for complex multivariate designs. An SPSS procedure is presented that can be used for calculating power for univariate, multivariate, and repeated measures models with and without time-varying and time-constant covariates. Three examples provide a framework for calculating power via this method: an ANCOVA, a MANOVA, and a repeated measures ANOVA with two or more groups. The benefits and limitations of this procedure are discussed.
Nonlinear dynamics of the atmospheric pollutants in Mexico City
NASA Astrophysics Data System (ADS)
Muñoz-Diosdado, Alejandro; Barrera-Ferrer, Amilcar; Angulo-Brown, Fernando
2014-05-01
The atmospheric pollution in the Metropolitan Zone of Mexico City (MZMC) is a serious problem with social, economical and political consequences, in virtue that it is the region which concentrates both the greatest country population and a great part of commercial and industrial activities. According to the World Health Organization, maximum permissible concentrations of atmospheric pollutants are exceeded frequently. In the MZMC, the environmental monitoring has been limited to criteria pollutants, named in this way due to when their levels are measured in the atmosphere, they indicate in a precise way the air quality. The Automatic Atmospheric Monitoring Network monitors and registers the values of pollutants concentration in air in the MZMC. Actually, it is integrated by approximately 35 automatic-equipped remote stations, which report an every-hour register. Local and global invariant quantities have been widely used to describe the fractal properties of diverse time series. In the study of certain time series, many times it is assumed that they are monofractal, which means that they can be described only with one fractal dimension. But this hypothesis is unrealistic because a lot of time series are heterogeneous and non stationary, so their scaling properties are not the same throughout time and therefore they may require more fractal dimensions for their description. Complexity of the atmospheric pollutants dynamics suggests us to analyze its time series of hourly concentration registers with the multifractal formalism. So, in this work, air concentration time series of MZMC criteria pollutants were studied with the proposed method. The chosen pollutants to perform this analysis are ozone, sulfur dioxide, carbon monoxide, nitrogen dioxide and PM10 (particles less than 10 micrometers). We found that pollutants air concentration time series are multifractal. When we calculate the degree of multifractality for each time series we know that while more multifractal are the time series, there is more complexity both in the time series and in the system from which the measurements were obtained. We studied the variation of the degree of multifractality over time, by calculating the multifractal spectra of the time series for each year; we see the variation in each monitoring station from 1990 until 2013. Multifractal analysis can tell us what kinds of correlations are present in the time series, and it is interesting to consider how these correlations vary over time. Our results show that for all the pollutants and all the monitoring stations the time series have long range correlations and they are highly persistent.
A time-series approach to dynamical systems from classical and quantum worlds
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fossion, Ruben
2014-01-08
This contribution discusses some recent applications of time-series analysis in Random Matrix Theory (RMT), and applications of RMT in the statistial analysis of eigenspectra of correlation matrices of multivariate time series.
Evaluation of nonlinearity and validity of nonlinear modeling for complex time series.
Suzuki, Tomoya; Ikeguchi, Tohru; Suzuki, Masuo
2007-10-01
Even if an original time series exhibits nonlinearity, it is not always effective to approximate the time series by a nonlinear model because such nonlinear models have high complexity from the viewpoint of information criteria. Therefore, we propose two measures to evaluate both the nonlinearity of a time series and validity of nonlinear modeling applied to it by nonlinear predictability and information criteria. Through numerical simulations, we confirm that the proposed measures effectively detect the nonlinearity of an observed time series and evaluate the validity of the nonlinear model. The measures are also robust against observational noises. We also analyze some real time series: the difference of the number of chickenpox and measles patients, the number of sunspots, five Japanese vowels, and the chaotic laser. We can confirm that the nonlinear model is effective for the Japanese vowel /a/, the difference of the number of measles patients, and the chaotic laser.
Time series analysis of the developed financial markets' integration using visibility graphs
NASA Astrophysics Data System (ADS)
Zhuang, Enyu; Small, Michael; Feng, Gang
2014-09-01
A time series representing the developed financial markets' segmentation from 1973 to 2012 is studied. The time series reveals an obvious market integration trend. To further uncover the features of this time series, we divide it into seven windows and generate seven visibility graphs. The measuring capabilities of the visibility graphs provide means to quantitatively analyze the original time series. It is found that the important historical incidents that influenced market integration coincide with variations in the measured graphical node degree. Through the measure of neighborhood span, the frequencies of the historical incidents are disclosed. Moreover, it is also found that large "cycles" and significant noise in the time series are linked to large and small communities in the generated visibility graphs. For large cycles, how historical incidents significantly affected market integration is distinguished by density and compactness of the corresponding communities.
Evaluation of nonlinearity and validity of nonlinear modeling for complex time series
NASA Astrophysics Data System (ADS)
Suzuki, Tomoya; Ikeguchi, Tohru; Suzuki, Masuo
2007-10-01
Even if an original time series exhibits nonlinearity, it is not always effective to approximate the time series by a nonlinear model because such nonlinear models have high complexity from the viewpoint of information criteria. Therefore, we propose two measures to evaluate both the nonlinearity of a time series and validity of nonlinear modeling applied to it by nonlinear predictability and information criteria. Through numerical simulations, we confirm that the proposed measures effectively detect the nonlinearity of an observed time series and evaluate the validity of the nonlinear model. The measures are also robust against observational noises. We also analyze some real time series: the difference of the number of chickenpox and measles patients, the number of sunspots, five Japanese vowels, and the chaotic laser. We can confirm that the nonlinear model is effective for the Japanese vowel /a/, the difference of the number of measles patients, and the chaotic laser.
Change point detection of the Persian Gulf sea surface temperature
NASA Astrophysics Data System (ADS)
Shirvani, A.
2017-01-01
In this study, the Student's t parametric and Mann-Whitney nonparametric change point models (CPMs) were applied to detect change point in the annual Persian Gulf sea surface temperature anomalies (PGSSTA) time series for the period 1951-2013. The PGSSTA time series, which were serially correlated, were transformed to produce an uncorrelated pre-whitened time series. The pre-whitened PGSSTA time series were utilized as the input file of change point models. Both the applied parametric and nonparametric CPMs estimated the change point in the PGSSTA in 1992. The PGSSTA follow the normal distribution up to 1992 and thereafter, but with a different mean value after year 1992. The estimated slope of linear trend in PGSSTA time series for the period 1951-1992 was negative; however, that was positive after the detected change point. Unlike the PGSSTA, the applied CPMs suggested no change point in the Niño3.4SSTA time series.
A high-fidelity weather time series generator using the Markov Chain process on a piecewise level
NASA Astrophysics Data System (ADS)
Hersvik, K.; Endrerud, O.-E. V.
2017-12-01
A method is developed for generating a set of unique weather time-series based on an existing weather series. The method allows statistically valid weather variations to take place within repeated simulations of offshore operations. The numerous generated time series need to share the same statistical qualities as the original time series. Statistical qualities here refer mainly to the distribution of weather windows available for work, including durations and frequencies of such weather windows, and seasonal characteristics. The method is based on the Markov chain process. The core new development lies in how the Markov Process is used, specifically by joining small pieces of random length time series together rather than joining individual weather states, each from a single time step, which is a common solution found in the literature. This new Markov model shows favorable characteristics with respect to the requirements set forth and all aspects of the validation performed.
NASA Astrophysics Data System (ADS)
Casdagli, M. C.
1997-09-01
We show that recurrence plots (RPs) give detailed characterizations of time series generated by dynamical systems driven by slowly varying external forces. For deterministic systems we show that RPs of the time series can be used to reconstruct the RP of the driving force if it varies sufficiently slowly. If the driving force is one-dimensional, its functional form can then be inferred up to an invertible coordinate transformation. The same results hold for stochastic systems if the RP of the time series is suitably averaged and transformed. These results are used to investigate the nonlinear prediction of time series generated by dynamical systems driven by slowly varying external forces. We also consider the problem of detecting a small change in the driving force, and propose a surrogate data technique for assessing statistical significance. Numerically simulated time series and a time series of respiration rates recorded from a subject with sleep apnea are used as illustrative examples.
A simple and fast representation space for classifying complex time series
NASA Astrophysics Data System (ADS)
Zunino, Luciano; Olivares, Felipe; Bariviera, Aurelio F.; Rosso, Osvaldo A.
2017-03-01
In the context of time series analysis considerable effort has been directed towards the implementation of efficient discriminating statistical quantifiers. Very recently, a simple and fast representation space has been introduced, namely the number of turning points versus the Abbe value. It is able to separate time series from stationary and non-stationary processes with long-range dependences. In this work we show that this bidimensional approach is useful for distinguishing complex time series: different sets of financial and physiological data are efficiently discriminated. Additionally, a multiscale generalization that takes into account the multiple time scales often involved in complex systems has been also proposed. This multiscale analysis is essential to reach a higher discriminative power between physiological time series in health and disease.
Using SAR satellite data time series for regional glacier mapping
NASA Astrophysics Data System (ADS)
Winsvold, Solveig H.; Kääb, Andreas; Nuth, Christopher; Andreassen, Liss M.; van Pelt, Ward J. J.; Schellenberger, Thomas
2018-03-01
With dense SAR satellite data time series it is possible to map surface and subsurface glacier properties that vary in time. On Sentinel-1A and RADARSAT-2 backscatter time series images over mainland Norway and Svalbard, we outline how to map glaciers using descriptive methods. We present five application scenarios. The first shows potential for tracking transient snow lines with SAR backscatter time series and correlates with both optical satellite images (Sentinel-2A and Landsat 8) and equilibrium line altitudes derived from in situ surface mass balance data. In the second application scenario, time series representation of glacier facies corresponding to SAR glacier zones shows potential for a more accurate delineation of the zones and how they change in time. The third application scenario investigates the firn evolution using dense SAR backscatter time series together with a coupled energy balance and multilayer firn model. We find strong correlation between backscatter signals with both the modeled firn air content and modeled wetness in the firn. In the fourth application scenario, we highlight how winter rain events can be detected in SAR time series, revealing important information about the area extent of internal accumulation. In the last application scenario, averaged summer SAR images were found to have potential in assisting the process of mapping glaciers outlines, especially in the presence of seasonal snow. Altogether we present examples of how to map glaciers and to further understand glaciological processes using the existing and future massive amount of multi-sensor time series data.
Gómez-Extremera, Manuel; Carpena, Pedro; Ivanov, Plamen Ch; Bernaola-Galván, Pedro A
2016-04-01
We systematically study the scaling properties of the magnitude and sign of the fluctuations in correlated time series, which is a simple and useful approach to distinguish between systems with different dynamical properties but the same linear correlations. First, we decompose artificial long-range power-law linearly correlated time series into magnitude and sign series derived from the consecutive increments in the original series, and we study their correlation properties. We find analytical expressions for the correlation exponent of the sign series as a function of the exponent of the original series. Such expressions are necessary for modeling surrogate time series with desired scaling properties. Next, we study linear and nonlinear correlation properties of series composed as products of independent magnitude and sign series. These surrogate series can be considered as a zero-order approximation to the analysis of the coupling of magnitude and sign in real data, a problem still open in many fields. We find analytical results for the scaling behavior of the composed series as a function of the correlation exponents of the magnitude and sign series used in the composition, and we determine the ranges of magnitude and sign correlation exponents leading to either single scaling or to crossover behaviors. Finally, we obtain how the linear and nonlinear properties of the composed series depend on the correlation exponents of their magnitude and sign series. Based on this information we propose a method to generate surrogate series with controlled correlation exponent and multifractal spectrum.
Multivariate generalized multifactor dimensionality reduction to detect gene-gene interactions
2013-01-01
Background Recently, one of the greatest challenges in genome-wide association studies is to detect gene-gene and/or gene-environment interactions for common complex human diseases. Ritchie et al. (2001) proposed multifactor dimensionality reduction (MDR) method for interaction analysis. MDR is a combinatorial approach to reduce multi-locus genotypes into high-risk and low-risk groups. Although MDR has been widely used for case-control studies with binary phenotypes, several extensions have been proposed. One of these methods, a generalized MDR (GMDR) proposed by Lou et al. (2007), allows adjusting for covariates and applying to both dichotomous and continuous phenotypes. GMDR uses the residual score of a generalized linear model of phenotypes to assign either high-risk or low-risk group, while MDR uses the ratio of cases to controls. Methods In this study, we propose multivariate GMDR, an extension of GMDR for multivariate phenotypes. Jointly analysing correlated multivariate phenotypes may have more power to detect susceptible genes and gene-gene interactions. We construct generalized estimating equations (GEE) with multivariate phenotypes to extend generalized linear models. Using the score vectors from GEE we discriminate high-risk from low-risk groups. We applied the multivariate GMDR method to the blood pressure data of the 7,546 subjects from the Korean Association Resource study: systolic blood pressure (SBP) and diastolic blood pressure (DBP). We compare the results of multivariate GMDR for SBP and DBP to the results from separate univariate GMDR for SBP and DBP, respectively. We also applied the multivariate GMDR method to the repeatedly measured hypertension status from 5,466 subjects and compared its result with those of univariate GMDR at each time point. Results Results from the univariate GMDR and multivariate GMDR in two-locus model with both blood pressures and hypertension phenotypes indicate best combinations of SNPs whose interaction has significant association with risk for high blood pressures or hypertension. Although the test balanced accuracy (BA) of multivariate analysis was not always greater than that of univariate analysis, the multivariate BAs were more stable with smaller standard deviations. Conclusions In this study, we have developed multivariate GMDR method using GEE approach. It is useful to use multivariate GMDR with correlated multiple phenotypes of interests. PMID:24565370
Compression based entropy estimation of heart rate variability on multiple time scales.
Baumert, Mathias; Voss, Andreas; Javorka, Michal
2013-01-01
Heart rate fluctuates beat by beat in a complex manner. The aim of this study was to develop a framework for entropy assessment of heart rate fluctuations on multiple time scales. We employed the Lempel-Ziv algorithm for lossless data compression to investigate the compressibility of RR interval time series on different time scales, using a coarse-graining procedure. We estimated the entropy of RR interval time series of 20 young and 20 old subjects and also investigated the compressibility of randomly shuffled surrogate RR time series. The original RR time series displayed significantly smaller compression entropy values than randomized RR interval data. The RR interval time series of older subjects showed significantly different entropy characteristics over multiple time scales than those of younger subjects. In conclusion, data compression may be useful approach for multiscale entropy assessment of heart rate variability.
Ghisi, Enedir; Cardoso, Karla Albino; Rupp, Ricardo Forgiarini
2012-06-15
The main objective of this article is to assess the possibility of using short-term instead of long-term rainfall time series to evaluate the potential for potable water savings by using rainwater in houses. The analysis was performed considering rainfall data from 1960 to 1995 for the city of Santa Bárbara do Oeste, located in the state of São Paulo, southeastern Brazil. The influence of the rainfall time series, roof area, potable water demand and percentage rainwater demand on the potential for potable water savings was evaluated. The potential for potable water savings was estimated using computer simulations considering a set of long-term rainfall time series and different sets of short-term rainfall time series. The ideal rainwater tank capacity was also assessed for some cases. It was observed that the higher the percentage rainwater demand and the shorter the rainfall time series, the larger the difference between the potential for potable water savings and the greater the variation in the ideal rainwater tank size. The sets of short-term rainfall time series considered adequate for different scenarios ranged from 1 to 13 years depending on the roof area, percentage rainwater demand and potable water demand. The main finding of the research is that sets of short-term rainfall time series can be used to assess the potential for potable water savings by using rainwater, as the results obtained are similar to those obtained from the long-term rainfall time series. Copyright © 2012 Elsevier Ltd. All rights reserved.
Carrilho, Flair J; Moraes, Cleusa R; Pinho, João RR; Mello, Isabel MVGC; Bertolini, Dennis A; Lemos, Marcílio F; Moreira, Regina C; Bassit, Leda C; Cardoso, Rita A; Ribeiro-dos-Santos, Gabriela; Da Silva, Luiz C
2004-01-01
Background Patients under haemodialysis are considered at high risk to acquire hepatitis B virus (HBV) infection. Since few data are reported from Brazil, our aim was to assess the frequency and risk factors for HBV infection in haemodialysis patients from 22 Dialysis Centres from Santa Catarina State, south of Brazil. Methods This study includes 813 patients, 149 haemodialysis workers and 772 healthy controls matched by sex and age. Serum samples were assayed for HBV markers and viraemia was detected by nested PCR. HBV was genotyped by partial S gene sequencing. Univariate and multivariate statistical analyses with stepwise logistic regression analysis were carried out to analyse the relationship between HBV infection and the characteristics of patients and their Dialysis Units. Results Frequency of HBV infection was 10.0%, 2.7% and 2.7% among patients, haemodialysis workers and controls, respectively. Amidst patients, the most frequent HBV genotypes were A (30.6%), D (57.1%) and F (12.2%). Univariate analysis showed association between HBV infection and total time in haemodialysis, type of dialysis equipment, hygiene and sterilization of equipment, number of times reusing the dialysis lines and filters, number of patients per care-worker and current HCV infection. The logistic regression model showed that total time in haemodialysis, number of times of reusing the dialysis lines and filters, and number of patients per worker were significantly related to HBV infection. Conclusions Frequency of HBV infection among haemodialysis patients at Santa Catarina state is very high. The most frequent HBV genotypes were A, D and F. The risk for a patient to become HBV positive increase 1.47 times each month of haemodialysis; 1.96 times if the dialysis unit reuses the lines and filters ≥ 10 times compared with haemodialysis units which reuse < 10 times; 3.42 times if the number of patients per worker is more than five. Sequence similarity among the HBV S gene from isolates of different patients pointed out to nosocomial transmission. PMID:15113436
Duffy, Sonia A.; Ronis, David L.; McLean, Scott; Fowler, Karen E.; Gruber, Stephen B.; Wolf, Gregory T.; Terrell, Jeffrey E.
2009-01-01
Purpose Our prior work has shown that the health behaviors of head and neck cancer patients are interrelated and are associated with quality of life; however, other than smoking, the relationship between health behaviors and survival is unclear. Patients and Methods A prospective cohort study was conducted to determine the relationship between five pretreatment health behaviors (smoking, alcohol, diet, physical activity, and sleep) and all-cause survival among 504 head and neck cancer patients. Results Smoking status was the strongest predictor of survival, with both current smokers (hazard ratio [HR] = 2.4; 95% CI, 1.3 to 4.4) and former smokers (HR = 2.0; 95% CI, 1.2 to 3.5) showing significant associations with poor survival. Problem drinking was associated with survival in the univariate analysis (HR = 1.4; 95% CI, 1.0 to 2.0) but lost significance when controlling for other factors. Low fruit intake was negatively associated with survival in the univariate analysis only (HR = 1.6; 95% CI, 1.1 to 2.1), whereas vegetable intake was not significant in either univariate or multivariate analyses. Although physical activity was associated with survival in the univariate analysis (HR = 0.95; 95% CI, 0.93 to 0.97), it was not significant in the multivariate model. Sleep was not significantly associated with survival in either univariate or multivariate analysis. Control variables that were also independently associated with survival in the multivariate analysis were age, education, tumor site, cancer stage, and surgical treatment. Conclusion Variation in selected pretreatment health behaviors (eg, smoking, fruit intake, and physical activity) in this population is associated with variation in survival. PMID:19289626
Varadhan, Ravi; Wang, Sue-Jane
2016-01-01
Treatment effect heterogeneity is a well-recognized phenomenon in randomized controlled clinical trials. In this paper, we discuss subgroup analyses with prespecified subgroups of clinical or biological importance. We explore various alternatives to the naive (the traditional univariate) subgroup analyses to address the issues of multiplicity and confounding. Specifically, we consider a model-based Bayesian shrinkage (Bayes-DS) and a nonparametric, empirical Bayes shrinkage approach (Emp-Bayes) to temper the optimism of traditional univariate subgroup analyses; a standardization approach (standardization) that accounts for correlation between baseline covariates; and a model-based maximum likelihood estimation (MLE) approach. The Bayes-DS and Emp-Bayes methods model the variation in subgroup-specific treatment effect rather than testing the null hypothesis of no difference between subgroups. The standardization approach addresses the issue of confounding in subgroup analyses. The MLE approach is considered only for comparison in simulation studies as the “truth” since the data were generated from the same model. Using the characteristics of a hypothetical large outcome trial, we perform simulation studies and articulate the utilities and potential limitations of these estimators. Simulation results indicate that Bayes-DS and Emp-Bayes can protect against optimism present in the naïve approach. Due to its simplicity, the naïve approach should be the reference for reporting univariate subgroup-specific treatment effect estimates from exploratory subgroup analyses. Standardization, although it tends to have a larger variance, is suggested when it is important to address the confounding of univariate subgroup effects due to correlation between baseline covariates. The Bayes-DS approach is available as an R package (DSBayes). PMID:26485117
Pleiotropy Analysis of Quantitative Traits at Gene Level by Multivariate Functional Linear Models
Wang, Yifan; Liu, Aiyi; Mills, James L.; Boehnke, Michael; Wilson, Alexander F.; Bailey-Wilson, Joan E.; Xiong, Momiao; Wu, Colin O.; Fan, Ruzong
2015-01-01
In genetics, pleiotropy describes the genetic effect of a single gene on multiple phenotypic traits. A common approach is to analyze the phenotypic traits separately using univariate analyses and combine the test results through multiple comparisons. This approach may lead to low power. Multivariate functional linear models are developed to connect genetic variant data to multiple quantitative traits adjusting for covariates for a unified analysis. Three types of approximate F-distribution tests based on Pillai–Bartlett trace, Hotelling–Lawley trace, and Wilks’s Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants in one genetic region. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and optimal sequence kernel association test (SKAT-O). Extensive simulations were performed to evaluate the false positive rates and power performance of the proposed models and tests. We show that the approximate F-distribution tests control the type I error rates very well. Overall, simultaneous analysis of multiple traits can increase power performance compared to an individual test of each trait. The proposed methods were applied to analyze (1) four lipid traits in eight European cohorts, and (2) three biochemical traits in the Trinity Students Study. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and SKAT-O for the three biochemical traits. The approximate F-distribution tests of the proposed functional linear models are more sensitive than those of the traditional multivariate linear models that in turn are more sensitive than SKAT-O in the univariate case. The analysis of the four lipid traits and the three biochemical traits detects more association than SKAT-O in the univariate case. PMID:25809955
Pleiotropy analysis of quantitative traits at gene level by multivariate functional linear models.
Wang, Yifan; Liu, Aiyi; Mills, James L; Boehnke, Michael; Wilson, Alexander F; Bailey-Wilson, Joan E; Xiong, Momiao; Wu, Colin O; Fan, Ruzong
2015-05-01
In genetics, pleiotropy describes the genetic effect of a single gene on multiple phenotypic traits. A common approach is to analyze the phenotypic traits separately using univariate analyses and combine the test results through multiple comparisons. This approach may lead to low power. Multivariate functional linear models are developed to connect genetic variant data to multiple quantitative traits adjusting for covariates for a unified analysis. Three types of approximate F-distribution tests based on Pillai-Bartlett trace, Hotelling-Lawley trace, and Wilks's Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants in one genetic region. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and optimal sequence kernel association test (SKAT-O). Extensive simulations were performed to evaluate the false positive rates and power performance of the proposed models and tests. We show that the approximate F-distribution tests control the type I error rates very well. Overall, simultaneous analysis of multiple traits can increase power performance compared to an individual test of each trait. The proposed methods were applied to analyze (1) four lipid traits in eight European cohorts, and (2) three biochemical traits in the Trinity Students Study. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and SKAT-O for the three biochemical traits. The approximate F-distribution tests of the proposed functional linear models are more sensitive than those of the traditional multivariate linear models that in turn are more sensitive than SKAT-O in the univariate case. The analysis of the four lipid traits and the three biochemical traits detects more association than SKAT-O in the univariate case. © 2015 WILEY PERIODICALS, INC.
Time series decomposition methods were applied to meteorological and air quality data and their numerical model estimates. Decomposition techniques express a time series as the sum of a small number of independent modes which hypothetically represent identifiable forcings, thereb...
A new method for reconstruction of solar irradiance
NASA Astrophysics Data System (ADS)
Privalsky, Victor
2018-07-01
The purpose of this research is to show how time series should be reconstructed using an example with the data on total solar irradiation (TSI) of the Earth and on sunspot numbers (SSN) since 1749. The traditional approach through regression equation(s) is designed for time-invariant vectors of random variables and is not applicable to time series, which present random functions of time. The autoregressive reconstruction (ARR) method suggested here requires fitting a multivariate stochastic difference equation to the target/proxy time series. The reconstruction is done through the scalar equation for the target time series with the white noise term excluded. The time series approach is shown to provide a better reconstruction of TSI than the correlation/regression method. A reconstruction criterion is introduced which allows one to define in advance the achievable level of success in the reconstruction. The conclusion is that time series, including the total solar irradiance, cannot be reconstructed properly if the data are not treated as sample records of random processes and analyzed in both time and frequency domains.