Sample records for random cluster models

  1. Prediction models for clustered data: comparison of a random intercept and standard regression model

    PubMed Central

    2013-01-01

    Background When study data are clustered, standard regression analysis is considered inappropriate and analytical techniques for clustered data need to be used. For prediction research in which the interest of predictor effects is on the patient level, random effect regression models are probably preferred over standard regression analysis. It is well known that the random effect parameter estimates and the standard logistic regression parameter estimates are different. Here, we compared random effect and standard logistic regression models for their ability to provide accurate predictions. Methods Using an empirical study on 1642 surgical patients at risk of postoperative nausea and vomiting, who were treated by one of 19 anesthesiologists (clusters), we developed prognostic models either with standard or random intercept logistic regression. External validity of these models was assessed in new patients from other anesthesiologists. We supported our results with simulation studies using intra-class correlation coefficients (ICC) of 5%, 15%, or 30%. Standard performance measures and measures adapted for the clustered data structure were estimated. Results The model developed with random effect analysis showed better discrimination than the standard approach, if the cluster effects were used for risk prediction (standard c-index of 0.69 versus 0.66). In the external validation set, both models showed similar discrimination (standard c-index 0.68 versus 0.67). The simulation study confirmed these results. For datasets with a high ICC (≥15%), model calibration was only adequate in external subjects, if the used performance measure assumed the same data structure as the model development method: standard calibration measures showed good calibration for the standard developed model, calibration measures adapting the clustered data structure showed good calibration for the prediction model with random intercept. Conclusion The models with random intercept discriminate better than the standard model only if the cluster effect is used for predictions. The prediction model with random intercept had good calibration within clusters. PMID:23414436

  2. Prediction models for clustered data: comparison of a random intercept and standard regression model.

    PubMed

    Bouwmeester, Walter; Twisk, Jos W R; Kappen, Teus H; van Klei, Wilton A; Moons, Karel G M; Vergouwe, Yvonne

    2013-02-15

    When study data are clustered, standard regression analysis is considered inappropriate and analytical techniques for clustered data need to be used. For prediction research in which the interest of predictor effects is on the patient level, random effect regression models are probably preferred over standard regression analysis. It is well known that the random effect parameter estimates and the standard logistic regression parameter estimates are different. Here, we compared random effect and standard logistic regression models for their ability to provide accurate predictions. Using an empirical study on 1642 surgical patients at risk of postoperative nausea and vomiting, who were treated by one of 19 anesthesiologists (clusters), we developed prognostic models either with standard or random intercept logistic regression. External validity of these models was assessed in new patients from other anesthesiologists. We supported our results with simulation studies using intra-class correlation coefficients (ICC) of 5%, 15%, or 30%. Standard performance measures and measures adapted for the clustered data structure were estimated. The model developed with random effect analysis showed better discrimination than the standard approach, if the cluster effects were used for risk prediction (standard c-index of 0.69 versus 0.66). In the external validation set, both models showed similar discrimination (standard c-index 0.68 versus 0.67). The simulation study confirmed these results. For datasets with a high ICC (≥15%), model calibration was only adequate in external subjects, if the used performance measure assumed the same data structure as the model development method: standard calibration measures showed good calibration for the standard developed model, calibration measures adapting the clustered data structure showed good calibration for the prediction model with random intercept. The models with random intercept discriminate better than the standard model only if the cluster effect is used for predictions. The prediction model with random intercept had good calibration within clusters.

  3. Percolation of the site random-cluster model by Monte Carlo method

    NASA Astrophysics Data System (ADS)

    Wang, Songsong; Zhang, Wanzhou; Ding, Chengxiang

    2015-08-01

    We propose a site random-cluster model by introducing an additional cluster weight in the partition function of the traditional site percolation. To simulate the model on a square lattice, we combine the color-assignation and the Swendsen-Wang methods to design a highly efficient cluster algorithm with a small critical slowing-down phenomenon. To verify whether or not it is consistent with the bond random-cluster model, we measure several quantities, such as the wrapping probability Re, the percolating cluster density P∞, and the magnetic susceptibility per site χp, as well as two exponents, such as the thermal exponent yt and the fractal dimension yh of the percolating cluster. We find that for different exponents of cluster weight q =1.5 , 2, 2.5 , 3, 3.5 , and 4, the numerical estimation of the exponents yt and yh are consistent with the theoretical values. The universalities of the site random-cluster model and the bond random-cluster model are completely identical. For larger values of q , we find obvious signatures of the first-order percolation transition by the histograms and the hysteresis loops of percolating cluster density and the energy per site. Our results are helpful for the understanding of the percolation of traditional statistical models.

  4. Handling Correlations between Covariates and Random Slopes in Multilevel Models

    ERIC Educational Resources Information Center

    Bates, Michael David; Castellano, Katherine E.; Rabe-Hesketh, Sophia; Skrondal, Anders

    2014-01-01

    This article discusses estimation of multilevel/hierarchical linear models that include cluster-level random intercepts and random slopes. Viewing the models as structural, the random intercepts and slopes represent the effects of omitted cluster-level covariates that may be correlated with included covariates. The resulting correlations between…

  5. The Effects of Including Observed Means or Latent Means as Covariates in Multilevel Models for Cluster Randomized Trials

    ERIC Educational Resources Information Center

    Aydin, Burak; Leite, Walter L.; Algina, James

    2016-01-01

    We investigated methods of including covariates in two-level models for cluster randomized trials to increase power to detect the treatment effect. We compared multilevel models that included either an observed cluster mean or a latent cluster mean as a covariate, as well as the effect of including Level 1 deviation scores in the model. A Monte…

  6. Baseline adjustments for binary data in repeated cross-sectional cluster randomized trials.

    PubMed

    Nixon, R M; Thompson, S G

    2003-09-15

    Analysis of covariance models, which adjust for a baseline covariate, are often used to compare treatment groups in a controlled trial in which individuals are randomized. Such analysis adjusts for any baseline imbalance and usually increases the precision of the treatment effect estimate. We assess the value of such adjustments in the context of a cluster randomized trial with repeated cross-sectional design and a binary outcome. In such a design, a new sample of individuals is taken from the clusters at each measurement occasion, so that baseline adjustment has to be at the cluster level. Logistic regression models are used to analyse the data, with cluster level random effects to allow for different outcome probabilities in each cluster. We compare the estimated treatment effect and its precision in models that incorporate a covariate measuring the cluster level probabilities at baseline and those that do not. In two data sets, taken from a cluster randomized trial in the treatment of menorrhagia, the value of baseline adjustment is only evident when the number of subjects per cluster is large. We assess the generalizability of these findings by undertaking a simulation study, and find that increased precision of the treatment effect requires both large cluster sizes and substantial heterogeneity between clusters at baseline, but baseline imbalance arising by chance in a randomized study can always be effectively adjusted for. Copyright 2003 John Wiley & Sons, Ltd.

  7. How large are the consequences of covariate imbalance in cluster randomized trials: a simulation study with a continuous outcome and a binary covariate at the cluster level.

    PubMed

    Moerbeek, Mirjam; van Schie, Sander

    2016-07-11

    The number of clusters in a cluster randomized trial is often low. It is therefore likely random assignment of clusters to treatment conditions results in covariate imbalance. There are no studies that quantify the consequences of covariate imbalance in cluster randomized trials on parameter and standard error bias and on power to detect treatment effects. The consequences of covariance imbalance in unadjusted and adjusted linear mixed models are investigated by means of a simulation study. The factors in this study are the degree of imbalance, the covariate effect size, the cluster size and the intraclass correlation coefficient. The covariate is binary and measured at the cluster level; the outcome is continuous and measured at the individual level. The results show covariate imbalance results in negligible parameter bias and small standard error bias in adjusted linear mixed models. Ignoring the possibility of covariate imbalance while calculating the sample size at the cluster level may result in a loss in power of at most 25 % in the adjusted linear mixed model. The results are more severe for the unadjusted linear mixed model: parameter biases up to 100 % and standard error biases up to 200 % may be observed. Power levels based on the unadjusted linear mixed model are often too low. The consequences are most severe for large clusters and/or small intraclass correlation coefficients since then the required number of clusters to achieve a desired power level is smallest. The possibility of covariate imbalance should be taken into account while calculating the sample size of a cluster randomized trial. Otherwise more sophisticated methods to randomize clusters to treatments should be used, such as stratification or balance algorithms. All relevant covariates should be carefully identified, be actually measured and included in the statistical model to avoid severe levels of parameter and standard error bias and insufficient power levels.

  8. Fast Constrained Spectral Clustering and Cluster Ensemble with Random Projection

    PubMed Central

    Liu, Wenfen

    2017-01-01

    Constrained spectral clustering (CSC) method can greatly improve the clustering accuracy with the incorporation of constraint information into spectral clustering and thus has been paid academic attention widely. In this paper, we propose a fast CSC algorithm via encoding landmark-based graph construction into a new CSC model and applying random sampling to decrease the data size after spectral embedding. Compared with the original model, the new algorithm has the similar results with the increase of its model size asymptotically; compared with the most efficient CSC algorithm known, the new algorithm runs faster and has a wider range of suitable data sets. Meanwhile, a scalable semisupervised cluster ensemble algorithm is also proposed via the combination of our fast CSC algorithm and dimensionality reduction with random projection in the process of spectral ensemble clustering. We demonstrate by presenting theoretical analysis and empirical results that the new cluster ensemble algorithm has advantages in terms of efficiency and effectiveness. Furthermore, the approximate preservation of random projection in clustering accuracy proved in the stage of consensus clustering is also suitable for the weighted k-means clustering and thus gives the theoretical guarantee to this special kind of k-means clustering where each point has its corresponding weight. PMID:29312447

  9. Clustering of time-course gene expression profiles using normal mixture models with autoregressive random effects

    PubMed Central

    2012-01-01

    Background Time-course gene expression data such as yeast cell cycle data may be periodically expressed. To cluster such data, currently used Fourier series approximations of periodic gene expressions have been found not to be sufficiently adequate to model the complexity of the time-course data, partly due to their ignoring the dependence between the expression measurements over time and the correlation among gene expression profiles. We further investigate the advantages and limitations of available models in the literature and propose a new mixture model with autoregressive random effects of the first order for the clustering of time-course gene-expression profiles. Some simulations and real examples are given to demonstrate the usefulness of the proposed models. Results We illustrate the applicability of our new model using synthetic and real time-course datasets. We show that our model outperforms existing models to provide more reliable and robust clustering of time-course data. Our model provides superior results when genetic profiles are correlated. It also gives comparable results when the correlation between the gene profiles is weak. In the applications to real time-course data, relevant clusters of coregulated genes are obtained, which are supported by gene-function annotation databases. Conclusions Our new model under our extension of the EMMIX-WIRE procedure is more reliable and robust for clustering time-course data because it adopts a random effects model that allows for the correlation among observations at different time points. It postulates gene-specific random effects with an autocorrelation variance structure that models coregulation within the clusters. The developed R package is flexible in its specification of the random effects through user-input parameters that enables improved modelling and consequent clustering of time-course data. PMID:23151154

  10. Search for Directed Networks by Different Random Walk Strategies

    NASA Astrophysics Data System (ADS)

    Zhu, Zi-Qi; Jin, Xiao-Ling; Huang, Zhi-Long

    2012-03-01

    A comparative study is carried out on the efficiency of five different random walk strategies searching on directed networks constructed based on several typical complex networks. Due to the difference in search efficiency of the strategies rooted in network clustering, the clustering coefficient in a random walker's eye on directed networks is defined and computed to be half of the corresponding undirected networks. The search processes are performed on the directed networks based on Erdös—Rényi model, Watts—Strogatz model, Barabási—Albert model and clustered scale-free network model. It is found that self-avoiding random walk strategy is the best search strategy for such directed networks. Compared to unrestricted random walk strategy, path-iteration-avoiding random walks can also make the search process much more efficient. However, no-triangle-loop and no-quadrangle-loop random walks do not improve the search efficiency as expected, which is different from those on undirected networks since the clustering coefficient of directed networks are smaller than that of undirected networks.

  11. Selection of Variables in Cluster Analysis: An Empirical Comparison of Eight Procedures

    ERIC Educational Resources Information Center

    Steinley, Douglas; Brusco, Michael J.

    2008-01-01

    Eight different variable selection techniques for model-based and non-model-based clustering are evaluated across a wide range of cluster structures. It is shown that several methods have difficulties when non-informative variables (i.e., random noise) are included in the model. Furthermore, the distribution of the random noise greatly impacts the…

  12. Multilevel Analysis Methods for Partially Nested Cluster Randomized Trials

    ERIC Educational Resources Information Center

    Sanders, Elizabeth A.

    2011-01-01

    This paper explores multilevel modeling approaches for 2-group randomized experiments in which a treatment condition involving clusters of individuals is compared to a control condition involving only ungrouped individuals, otherwise known as partially nested cluster randomized designs (PNCRTs). Strategies for comparing groups from a PNCRT in the…

  13. Bayesian hierarchical models for cost-effectiveness analyses that use data from cluster randomized trials.

    PubMed

    Grieve, Richard; Nixon, Richard; Thompson, Simon G

    2010-01-01

    Cost-effectiveness analyses (CEA) may be undertaken alongside cluster randomized trials (CRTs) where randomization is at the level of the cluster (for example, the hospital or primary care provider) rather than the individual. Costs (and outcomes) within clusters may be correlated so that the assumption made by standard bivariate regression models, that observations are independent, is incorrect. This study develops a flexible modeling framework to acknowledge the clustering in CEA that use CRTs. The authors extend previous Bayesian bivariate models for CEA of multicenter trials to recognize the specific form of clustering in CRTs. They develop new Bayesian hierarchical models (BHMs) that allow mean costs and outcomes, and also variances, to differ across clusters. They illustrate how each model can be applied using data from a large (1732 cases, 70 primary care providers) CRT evaluating alternative interventions for reducing postnatal depression. The analyses compare cost-effectiveness estimates from BHMs with standard bivariate regression models that ignore the data hierarchy. The BHMs show high levels of cost heterogeneity across clusters (intracluster correlation coefficient, 0.17). Compared with standard regression models, the BHMs yield substantially increased uncertainty surrounding the cost-effectiveness estimates, and altered point estimates. The authors conclude that ignoring clustering can lead to incorrect inferences. The BHMs that they present offer a flexible modeling framework that can be applied more generally to CEA that use CRTs.

  14. General Framework for Effect Sizes in Cluster Randomized Experiments

    ERIC Educational Resources Information Center

    VanHoudnos, Nathan

    2016-01-01

    Cluster randomized experiments are ubiquitous in modern education research. Although a variety of modeling approaches are used to analyze these data, perhaps the most common methodology is a normal mixed effects model where some effects, such as the treatment effect, are regarded as fixed, and others, such as the effect of group random assignment…

  15. Finite-sample corrected generalized estimating equation of population average treatment effects in stepped wedge cluster randomized trials.

    PubMed

    Scott, JoAnna M; deCamp, Allan; Juraska, Michal; Fay, Michael P; Gilbert, Peter B

    2017-04-01

    Stepped wedge designs are increasingly commonplace and advantageous for cluster randomized trials when it is both unethical to assign placebo, and it is logistically difficult to allocate an intervention simultaneously to many clusters. We study marginal mean models fit with generalized estimating equations for assessing treatment effectiveness in stepped wedge cluster randomized trials. This approach has advantages over the more commonly used mixed models that (1) the population-average parameters have an important interpretation for public health applications and (2) they avoid untestable assumptions on latent variable distributions and avoid parametric assumptions about error distributions, therefore, providing more robust evidence on treatment effects. However, cluster randomized trials typically have a small number of clusters, rendering the standard generalized estimating equation sandwich variance estimator biased and highly variable and hence yielding incorrect inferences. We study the usual asymptotic generalized estimating equation inferences (i.e., using sandwich variance estimators and asymptotic normality) and four small-sample corrections to generalized estimating equation for stepped wedge cluster randomized trials and for parallel cluster randomized trials as a comparison. We show by simulation that the small-sample corrections provide improvement, with one correction appearing to provide at least nominal coverage even with only 10 clusters per group. These results demonstrate the viability of the marginal mean approach for both stepped wedge and parallel cluster randomized trials. We also study the comparative performance of the corrected methods for stepped wedge and parallel designs, and describe how the methods can accommodate interval censoring of individual failure times and incorporate semiparametric efficient estimators.

  16. A pattern-mixture model approach for handling missing continuous outcome data in longitudinal cluster randomized trials.

    PubMed

    Fiero, Mallorie H; Hsu, Chiu-Hsieh; Bell, Melanie L

    2017-11-20

    We extend the pattern-mixture approach to handle missing continuous outcome data in longitudinal cluster randomized trials, which randomize groups of individuals to treatment arms, rather than the individuals themselves. Individuals who drop out at the same time point are grouped into the same dropout pattern. We approach extrapolation of the pattern-mixture model by applying multilevel multiple imputation, which imputes missing values while appropriately accounting for the hierarchical data structure found in cluster randomized trials. To assess parameters of interest under various missing data assumptions, imputed values are multiplied by a sensitivity parameter, k, which increases or decreases imputed values. Using simulated data, we show that estimates of parameters of interest can vary widely under differing missing data assumptions. We conduct a sensitivity analysis using real data from a cluster randomized trial by increasing k until the treatment effect inference changes. By performing a sensitivity analysis for missing data, researchers can assess whether certain missing data assumptions are reasonable for their cluster randomized trial. Copyright © 2017 John Wiley & Sons, Ltd.

  17. Measurement Error Correction Formula for Cluster-Level Group Differences in Cluster Randomized and Observational Studies

    ERIC Educational Resources Information Center

    Cho, Sun-Joo; Preacher, Kristopher J.

    2016-01-01

    Multilevel modeling (MLM) is frequently used to detect cluster-level group differences in cluster randomized trial and observational studies. Group differences on the outcomes (posttest scores) are detected by controlling for the covariate (pretest scores) as a proxy variable for unobserved factors that predict future attributes. The pretest and…

  18. Estimating overall exposure effects for the clustered and censored outcome using random effect Tobit regression models.

    PubMed

    Wang, Wei; Griswold, Michael E

    2016-11-30

    The random effect Tobit model is a regression model that accommodates both left- and/or right-censoring and within-cluster dependence of the outcome variable. Regression coefficients of random effect Tobit models have conditional interpretations on a constructed latent dependent variable and do not provide inference of overall exposure effects on the original outcome scale. Marginalized random effects model (MREM) permits likelihood-based estimation of marginal mean parameters for the clustered data. For random effect Tobit models, we extend the MREM to marginalize over both the random effects and the normal space and boundary components of the censored response to estimate overall exposure effects at population level. We also extend the 'Average Predicted Value' method to estimate the model-predicted marginal means for each person under different exposure status in a designated reference group by integrating over the random effects and then use the calculated difference to assess the overall exposure effect. The maximum likelihood estimation is proposed utilizing a quasi-Newton optimization algorithm with Gauss-Hermite quadrature to approximate the integration of the random effects. We use these methods to carefully analyze two real datasets. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  19. Ferromagnetic clusters induced by a nonmagnetic random disorder in diluted magnetic semiconductors

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bui, Dinh-Hoi; Physics Department, Hue University’s College of Education, 34 Le Loi, Hue; Phan, Van-Nham, E-mail: phanvannham@dtu.edu.vn

    In this work, we analyze the nonmagnetic random disorder leading to a formation of ferromagnetic clusters in diluted magnetic semiconductors. The nonmagnetic random disorder arises from randomness in the host lattice. Including the disorder to the Kondo lattice model with random distribution of magnetic dopants, the ferromagnetic–paramagnetic transition in the system is investigated in the framework of dynamical mean-field theory. At a certain low temperature one finds a fraction of ferromagnetic sites transiting to the paramagnetic state. Enlarging the nonmagnetic random disorder strength, the paramagnetic regimes expand resulting in the formation of the ferromagnetic clusters.

  20. Quenched Large Deviations for Simple Random Walks on Percolation Clusters Including Long-Range Correlations

    NASA Astrophysics Data System (ADS)

    Berger, Noam; Mukherjee, Chiranjib; Okamura, Kazuki

    2018-03-01

    We prove a quenched large deviation principle (LDP) for a simple random walk on a supercritical percolation cluster (SRWPC) on {Z^d} ({d ≥ 2}). The models under interest include classical Bernoulli bond and site percolation as well as models that exhibit long range correlations, like the random cluster model, the random interlacement and the vacant set of random interlacements (for {d ≥ 3}) and the level sets of the Gaussian free field ({d≥ 3}). Inspired by the methods developed by Kosygina et al. (Commun Pure Appl Math 59:1489-1521, 2006) for proving quenched LDP for elliptic diffusions with a random drift, and by Yilmaz (Commun Pure Appl Math 62(8):1033-1075, 2009) and Rosenbluth (Quenched large deviations for multidimensional random walks in a random environment: a variational formula. Ph.D. thesis, NYU, arXiv:0804.1444v1) for similar results regarding elliptic random walks in random environment, we take the point of view of the moving particle and prove a large deviation principle for the quenched distribution of the pair empirical measures of the environment Markov chain in the non-elliptic case of SRWPC. Via a contraction principle, this reduces easily to a quenched LDP for the distribution of the mean velocity of the random walk and both rate functions admit explicit variational formulas. The main difficulty in our set up lies in the inherent non-ellipticity as well as the lack of translation-invariance stemming from conditioning on the fact that the origin belongs to the infinite cluster. We develop a unifying approach for proving quenched large deviations for SRWPC based on exploiting coercivity properties of the relative entropies in the context of convex variational analysis, combined with input from ergodic theory and invoking geometric properties of the supercritical percolation cluster.

  1. Quenched Large Deviations for Simple Random Walks on Percolation Clusters Including Long-Range Correlations

    NASA Astrophysics Data System (ADS)

    Berger, Noam; Mukherjee, Chiranjib; Okamura, Kazuki

    2017-12-01

    We prove a quenched large deviation principle (LDP) for a simple random walk on a supercritical percolation cluster (SRWPC) on {Z^d} ({d ≥ 2} ). The models under interest include classical Bernoulli bond and site percolation as well as models that exhibit long range correlations, like the random cluster model, the random interlacement and the vacant set of random interlacements (for {d ≥ 3} ) and the level sets of the Gaussian free field ({d≥ 3} ). Inspired by the methods developed by Kosygina et al. (Commun Pure Appl Math 59:1489-1521, 2006) for proving quenched LDP for elliptic diffusions with a random drift, and by Yilmaz (Commun Pure Appl Math 62(8):1033-1075, 2009) and Rosenbluth (Quenched large deviations for multidimensional random walks in a random environment: a variational formula. Ph.D. thesis, NYU, arXiv:0804.1444v1) for similar results regarding elliptic random walks in random environment, we take the point of view of the moving particle and prove a large deviation principle for the quenched distribution of the pair empirical measures of the environment Markov chain in the non-elliptic case of SRWPC. Via a contraction principle, this reduces easily to a quenched LDP for the distribution of the mean velocity of the random walk and both rate functions admit explicit variational formulas. The main difficulty in our set up lies in the inherent non-ellipticity as well as the lack of translation-invariance stemming from conditioning on the fact that the origin belongs to the infinite cluster. We develop a unifying approach for proving quenched large deviations for SRWPC based on exploiting coercivity properties of the relative entropies in the context of convex variational analysis, combined with input from ergodic theory and invoking geometric properties of the supercritical percolation cluster.

  2. Quantifying the impact of fixed effects modeling of clusters in multiple imputation for cluster randomized trials

    PubMed Central

    Andridge, Rebecca. R.

    2011-01-01

    In cluster randomized trials (CRTs), identifiable clusters rather than individuals are randomized to study groups. Resulting data often consist of a small number of clusters with correlated observations within a treatment group. Missing data often present a problem in the analysis of such trials, and multiple imputation (MI) has been used to create complete data sets, enabling subsequent analysis with well-established analysis methods for CRTs. We discuss strategies for accounting for clustering when multiply imputing a missing continuous outcome, focusing on estimation of the variance of group means as used in an adjusted t-test or ANOVA. These analysis procedures are congenial to (can be derived from) a mixed effects imputation model; however, this imputation procedure is not yet available in commercial statistical software. An alternative approach that is readily available and has been used in recent studies is to include fixed effects for cluster, but the impact of using this convenient method has not been studied. We show that under this imputation model the MI variance estimator is positively biased and that smaller ICCs lead to larger overestimation of the MI variance. Analytical expressions for the bias of the variance estimator are derived in the case of data missing completely at random (MCAR), and cases in which data are missing at random (MAR) are illustrated through simulation. Finally, various imputation methods are applied to data from the Detroit Middle School Asthma Project, a recent school-based CRT, and differences in inference are compared. PMID:21259309

  3. Random phase approximation and cluster mean field studies of hard core Bose Hubbard model

    NASA Astrophysics Data System (ADS)

    Alavani, Bhargav K.; Gaude, Pallavi P.; Pai, Ramesh V.

    2018-04-01

    We investigate zero temperature and finite temperature properties of the Bose Hubbard Model in the hard core limit using Random Phase Approximation (RPA) and Cluster Mean Field Theory (CMFT). We show that our RPA calculations are able to capture quantum and thermal fluctuations significantly better than CMFT.

  4. Inference from clustering with application to gene-expression microarrays.

    PubMed

    Dougherty, Edward R; Barrera, Junior; Brun, Marcel; Kim, Seungchan; Cesar, Roberto M; Chen, Yidong; Bittner, Michael; Trent, Jeffrey M

    2002-01-01

    There are many algorithms to cluster sample data points based on nearness or a similarity measure. Often the implication is that points in different clusters come from different underlying classes, whereas those in the same cluster come from the same class. Stochastically, the underlying classes represent different random processes. The inference is that clusters represent a partition of the sample points according to which process they belong. This paper discusses a model-based clustering toolbox that evaluates cluster accuracy. Each random process is modeled as its mean plus independent noise, sample points are generated, the points are clustered, and the clustering error is the number of points clustered incorrectly according to the generating random processes. Various clustering algorithms are evaluated based on process variance and the key issue of the rate at which algorithmic performance improves with increasing numbers of experimental replications. The model means can be selected by hand to test the separability of expected types of biological expression patterns. Alternatively, the model can be seeded by real data to test the expected precision of that output or the extent of improvement in precision that replication could provide. In the latter case, a clustering algorithm is used to form clusters, and the model is seeded with the means and variances of these clusters. Other algorithms are then tested relative to the seeding algorithm. Results are averaged over various seeds. Output includes error tables and graphs, confusion matrices, principal-component plots, and validation measures. Five algorithms are studied in detail: K-means, fuzzy C-means, self-organizing maps, hierarchical Euclidean-distance-based and correlation-based clustering. The toolbox is applied to gene-expression clustering based on cDNA microarrays using real data. Expression profile graphics are generated and error analysis is displayed within the context of these profile graphics. A large amount of generated output is available over the web.

  5. Clustering, randomness and regularity in cloud fields. I - Theoretical considerations. II - Cumulus cloud fields

    NASA Technical Reports Server (NTRS)

    Weger, R. C.; Lee, J.; Zhu, Tianri; Welch, R. M.

    1992-01-01

    The current controversy existing in reference to the regularity vs. clustering in cloud fields is examined by means of analysis and simulation studies based upon nearest-neighbor cumulative distribution statistics. It is shown that the Poisson representation of random point processes is superior to pseudorandom-number-generated models and that pseudorandom-number-generated models bias the observed nearest-neighbor statistics towards regularity. Interpretation of this nearest-neighbor statistics is discussed for many cases of superpositions of clustering, randomness, and regularity. A detailed analysis is carried out of cumulus cloud field spatial distributions based upon Landsat, AVHRR, and Skylab data, showing that, when both large and small clouds are included in the cloud field distributions, the cloud field always has a strong clustering signal.

  6. Bootstrap-based methods for estimating standard errors in Cox's regression analyses of clustered event times.

    PubMed

    Xiao, Yongling; Abrahamowicz, Michal

    2010-03-30

    We propose two bootstrap-based methods to correct the standard errors (SEs) from Cox's model for within-cluster correlation of right-censored event times. The cluster-bootstrap method resamples, with replacement, only the clusters, whereas the two-step bootstrap method resamples (i) the clusters, and (ii) individuals within each selected cluster, with replacement. In simulations, we evaluate both methods and compare them with the existing robust variance estimator and the shared gamma frailty model, which are available in statistical software packages. We simulate clustered event time data, with latent cluster-level random effects, which are ignored in the conventional Cox's model. For cluster-level covariates, both proposed bootstrap methods yield accurate SEs, and type I error rates, and acceptable coverage rates, regardless of the true random effects distribution, and avoid serious variance under-estimation by conventional Cox-based standard errors. However, the two-step bootstrap method over-estimates the variance for individual-level covariates. We also apply the proposed bootstrap methods to obtain confidence bands around flexible estimates of time-dependent effects in a real-life analysis of cluster event times.

  7. Clustering and Phase Transitions on a Neutral Landscape

    NASA Astrophysics Data System (ADS)

    Scott, Adam; King, Dawn; Maric, Nevena; Bahar, Sonya

    2012-02-01

    The problem of speciation and species aggregation on a neutral landscape, subject to random mutational fluctuations rather than selective drive, has been a focus of research since the seminal work of Kimura on genetic drift. These ideas have received increased attention due to the more recent development of a neutral ecological theory by Hubbell. De Aguiar et al. recently demonstrated, in a computational model, that speciation can occur under neutral conditions; this study bears some comparison with more mathematical studies of clustering on neutral landscapes in the context of branching and annihilating random walks. Here, we show that clustering can occur on a neutral landscape where the dimensions specify the simulated organisms' phenotypes. Unlike the De Aguiar et al. model, we simulate sympatric speciation: the organisms cluster phenotypically, but are not spatially separated. Moreover, we find that clustering occurs not only in the case of assortative mating, but also in the case of asexual fission. Clustering is not observed in a control case where organisms can mate randomly. We find that the population size and the number of clusters undergo phase-transition-like behavior as the maximum mutation size is varied.

  8. MODEL-BASED CLUSTERING FOR CLASSIFICATION OF AQUATIC SYSTEMS AND DIAGNOSIS OF ECOLOGICAL STRESS

    EPA Science Inventory

    Clustering approaches were developed using the classification likelihood, the mixture likelihood, and also using a randomization approach with a model index. Using a clustering approach based on the mixture and classification likelihoods, we have developed an algorithm that...

  9. Bayesian network meta-analysis for cluster randomized trials with binary outcomes.

    PubMed

    Uhlmann, Lorenz; Jensen, Katrin; Kieser, Meinhard

    2017-06-01

    Network meta-analysis is becoming a common approach to combine direct and indirect comparisons of several treatment arms. In recent research, there have been various developments and extensions of the standard methodology. Simultaneously, cluster randomized trials are experiencing an increased popularity, especially in the field of health services research, where, for example, medical practices are the units of randomization but the outcome is measured at the patient level. Combination of the results of cluster randomized trials is challenging. In this tutorial, we examine and compare different approaches for the incorporation of cluster randomized trials in a (network) meta-analysis. Furthermore, we provide practical insight on the implementation of the models. In simulation studies, it is shown that some of the examined approaches lead to unsatisfying results. However, there are alternatives which are suitable to combine cluster randomized trials in a network meta-analysis as they are unbiased and reach accurate coverage rates. In conclusion, the methodology can be extended in such a way that an adequate inclusion of the results obtained in cluster randomized trials becomes feasible. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  10. A polymer, random walk model for the size-distribution of large DNA fragments after high linear energy transfer radiation

    NASA Technical Reports Server (NTRS)

    Ponomarev, A. L.; Brenner, D.; Hlatky, L. R.; Sachs, R. K.

    2000-01-01

    DNA double-strand breaks (DSBs) produced by densely ionizing radiation are not located randomly in the genome: recent data indicate DSB clustering along chromosomes. Stochastic DSB clustering at large scales, from > 100 Mbp down to < 0.01 Mbp, is modeled using computer simulations and analytic equations. A random-walk, coarse-grained polymer model for chromatin is combined with a simple track structure model in Monte Carlo software called DNAbreak and is applied to data on alpha-particle irradiation of V-79 cells. The chromatin model neglects molecular details but systematically incorporates an increase in average spatial separation between two DNA loci as the number of base-pairs between the loci increases. Fragment-size distributions obtained using DNAbreak match data on large fragments about as well as distributions previously obtained with a less mechanistic approach. Dose-response relations, linear at small doses of high linear energy transfer (LET) radiation, are obtained. They are found to be non-linear when the dose becomes so large that there is a significant probability of overlapping or close juxtaposition, along one chromosome, for different DSB clusters from different tracks. The non-linearity is more evident for large fragments than for small. The DNAbreak results furnish an example of the RLC (randomly located clusters) analytic formalism, which generalizes the broken-stick fragment-size distribution of the random-breakage model that is often applied to low-LET data.

  11. Using Cluster Bootstrapping to Analyze Nested Data With a Few Clusters.

    PubMed

    Huang, Francis L

    2018-04-01

    Cluster randomized trials involving participants nested within intact treatment and control groups are commonly performed in various educational, psychological, and biomedical studies. However, recruiting and retaining intact groups present various practical, financial, and logistical challenges to evaluators and often, cluster randomized trials are performed with a low number of clusters (~20 groups). Although multilevel models are often used to analyze nested data, researchers may be concerned of potentially biased results due to having only a few groups under study. Cluster bootstrapping has been suggested as an alternative procedure when analyzing clustered data though it has seen very little use in educational and psychological studies. Using a Monte Carlo simulation that varied the number of clusters, average cluster size, and intraclass correlations, we compared standard errors using cluster bootstrapping with those derived using ordinary least squares regression and multilevel models. Results indicate that cluster bootstrapping, though more computationally demanding, can be used as an alternative procedure for the analysis of clustered data when treatment effects at the group level are of primary interest. Supplementary material showing how to perform cluster bootstrapped regressions using R is also provided.

  12. Detecting Intervention Effects in a Cluster-Randomized Design Using Multilevel Structural Equation Modeling for Binary Responses

    ERIC Educational Resources Information Center

    Cho, Sun-Joo; Preacher, Kristopher J.; Bottge, Brian A.

    2015-01-01

    Multilevel modeling (MLM) is frequently used to detect group differences, such as an intervention effect in a pre-test--post-test cluster-randomized design. Group differences on the post-test scores are detected by controlling for pre-test scores as a proxy variable for unobserved factors that predict future attributes. The pre-test and post-test…

  13. Bias and inference from misspecified mixed-effect models in stepped wedge trial analysis.

    PubMed

    Thompson, Jennifer A; Fielding, Katherine L; Davey, Calum; Aiken, Alexander M; Hargreaves, James R; Hayes, Richard J

    2017-10-15

    Many stepped wedge trials (SWTs) are analysed by using a mixed-effect model with a random intercept and fixed effects for the intervention and time periods (referred to here as the standard model). However, it is not known whether this model is robust to misspecification. We simulated SWTs with three groups of clusters and two time periods; one group received the intervention during the first period and two groups in the second period. We simulated period and intervention effects that were either common-to-all or varied-between clusters. Data were analysed with the standard model or with additional random effects for period effect or intervention effect. In a second simulation study, we explored the weight given to within-cluster comparisons by simulating a larger intervention effect in the group of the trial that experienced both the control and intervention conditions and applying the three analysis models described previously. Across 500 simulations, we computed bias and confidence interval coverage of the estimated intervention effect. We found up to 50% bias in intervention effect estimates when period or intervention effects varied between clusters and were treated as fixed effects in the analysis. All misspecified models showed undercoverage of 95% confidence intervals, particularly the standard model. A large weight was given to within-cluster comparisons in the standard model. In the SWTs simulated here, mixed-effect models were highly sensitive to departures from the model assumptions, which can be explained by the high dependence on within-cluster comparisons. Trialists should consider including a random effect for time period in their SWT analysis model. © 2017 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd. © 2017 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

  14. Bias and inference from misspecified mixed‐effect models in stepped wedge trial analysis

    PubMed Central

    Fielding, Katherine L.; Davey, Calum; Aiken, Alexander M.; Hargreaves, James R.; Hayes, Richard J.

    2017-01-01

    Many stepped wedge trials (SWTs) are analysed by using a mixed‐effect model with a random intercept and fixed effects for the intervention and time periods (referred to here as the standard model). However, it is not known whether this model is robust to misspecification. We simulated SWTs with three groups of clusters and two time periods; one group received the intervention during the first period and two groups in the second period. We simulated period and intervention effects that were either common‐to‐all or varied‐between clusters. Data were analysed with the standard model or with additional random effects for period effect or intervention effect. In a second simulation study, we explored the weight given to within‐cluster comparisons by simulating a larger intervention effect in the group of the trial that experienced both the control and intervention conditions and applying the three analysis models described previously. Across 500 simulations, we computed bias and confidence interval coverage of the estimated intervention effect. We found up to 50% bias in intervention effect estimates when period or intervention effects varied between clusters and were treated as fixed effects in the analysis. All misspecified models showed undercoverage of 95% confidence intervals, particularly the standard model. A large weight was given to within‐cluster comparisons in the standard model. In the SWTs simulated here, mixed‐effect models were highly sensitive to departures from the model assumptions, which can be explained by the high dependence on within‐cluster comparisons. Trialists should consider including a random effect for time period in their SWT analysis model. © 2017 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd. PMID:28556355

  15. A Bayesian, generalized frailty model for comet assays.

    PubMed

    Ghebretinsae, Aklilu Habteab; Faes, Christel; Molenberghs, Geert; De Boeck, Marlies; Geys, Helena

    2013-05-01

    This paper proposes a flexible modeling approach for so-called comet assay data regularly encountered in preclinical research. While such data consist of non-Gaussian outcomes in a multilevel hierarchical structure, traditional analyses typically completely or partly ignore this hierarchical nature by summarizing measurements within a cluster. Non-Gaussian outcomes are often modeled using exponential family models. This is true not only for binary and count data, but also for, example, time-to-event outcomes. Two important reasons for extending this family are for (1) the possible occurrence of overdispersion, meaning that the variability in the data may not be adequately described by the models, which often exhibit a prescribed mean-variance link, and (2) the accommodation of a hierarchical structure in the data, owing to clustering in the data. The first issue is dealt with through so-called overdispersion models. Clustering is often accommodated through the inclusion of random subject-specific effects. Though not always, one conventionally assumes such random effects to be normally distributed. In the case of time-to-event data, one encounters, for example, the gamma frailty model (Duchateau and Janssen, 2007 ). While both of these issues may occur simultaneously, models combining both are uncommon. Molenberghs et al. ( 2010 ) proposed a broad class of generalized linear models accommodating overdispersion and clustering through two separate sets of random effects. Here, we use this method to model data from a comet assay with a three-level hierarchical structure. Although a conjugate gamma random effect is used for the overdispersion random effect, both gamma and normal random effects are considered for the hierarchical random effect. Apart from model formulation, we place emphasis on Bayesian estimation. Our proposed method has an upper hand over the traditional analysis in that it (1) uses the appropriate distribution stipulated in the literature; (2) deals with the complete hierarchical nature; and (3) uses all information instead of summary measures. The fit of the model to the comet assay is compared against the background of more conventional model fits. Results indicate the toxicity of 1,2-dimethylhydrazine dihydrochloride at different dose levels (low, medium, and high).

  16. Disentangling giant component and finite cluster contributions in sparse random matrix spectra.

    PubMed

    Kühn, Reimer

    2016-04-01

    We describe a method for disentangling giant component and finite cluster contributions to sparse random matrix spectra, using sparse symmetric random matrices defined on Erdős-Rényi graphs as an example and test bed. Our methods apply to sparse matrices defined in terms of arbitrary graphs in the configuration model class, as long as they have finite mean degree.

  17. Multivariate generalized hidden Markov regression models with random covariates: Physical exercise in an elderly population.

    PubMed

    Punzo, Antonio; Ingrassia, Salvatore; Maruotti, Antonello

    2018-04-22

    A time-varying latent variable model is proposed to jointly analyze multivariate mixed-support longitudinal data. The proposal can be viewed as an extension of hidden Markov regression models with fixed covariates (HMRMFCs), which is the state of the art for modelling longitudinal data, with a special focus on the underlying clustering structure. HMRMFCs are inadequate for applications in which a clustering structure can be identified in the distribution of the covariates, as the clustering is independent from the covariates distribution. Here, hidden Markov regression models with random covariates are introduced by explicitly specifying state-specific distributions for the covariates, with the aim of improving the recovering of the clusters in the data with respect to a fixed covariates paradigm. The hidden Markov regression models with random covariates class is defined focusing on the exponential family, in a generalized linear model framework. Model identifiability conditions are sketched, an expectation-maximization algorithm is outlined for parameter estimation, and various implementation and operational issues are discussed. Properties of the estimators of the regression coefficients, as well as of the hidden path parameters, are evaluated through simulation experiments and compared with those of HMRMFCs. The method is applied to physical activity data. Copyright © 2018 John Wiley & Sons, Ltd.

  18. Simulating star clusters with the AMUSE software framework. I. Dependence of cluster lifetimes on model assumptions and cluster dissolution modes

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Whitehead, Alfred J.; McMillan, Stephen L. W.; Vesperini, Enrico

    2013-12-01

    We perform a series of simulations of evolving star clusters using the Astrophysical Multipurpose Software Environment (AMUSE), a new community-based multi-physics simulation package, and compare our results to existing work. These simulations model a star cluster beginning with a King model distribution and a selection of power-law initial mass functions and contain a tidal cutoff. They are evolved using collisional stellar dynamics and include mass loss due to stellar evolution. After studying and understanding that the differences between AMUSE results and results from previous studies are understood, we explored the variation in cluster lifetimes due to the random realization noisemore » introduced by transforming a King model to specific initial conditions. This random realization noise can affect the lifetime of a simulated star cluster by up to 30%. Two modes of star cluster dissolution were identified: a mass evolution curve that contains a runaway cluster dissolution with a sudden loss of mass, and a dissolution mode that does not contain this feature. We refer to these dissolution modes as 'dynamical' and 'relaxation' dominated, respectively. For Salpeter-like initial mass functions, we determined the boundary between these two modes in terms of the dynamical and relaxation timescales.« less

  19. Temporal clustering of tropical cyclones and its ecosystem impacts

    PubMed Central

    Mumby, Peter J.; Vitolo, Renato; Stephenson, David B.

    2011-01-01

    Tropical cyclones have massive economic, social, and ecological impacts, and models of their occurrence influence many planning activities from setting insurance premiums to conservation planning. Most impact models allow for geographically varying cyclone rates but assume that individual storm events occur randomly with constant rate in time. This study analyzes the statistical properties of Atlantic tropical cyclones and shows that local cyclone counts vary in time, with periods of elevated activity followed by relative quiescence. Such temporal clustering is particularly strong in the Caribbean Sea, along the coasts of Belize, Honduras, Costa Rica, Jamaica, the southwest of Haiti, and in the main hurricane development region in the North Atlantic between Africa and the Caribbean. Failing to recognize this natural nonstationarity in cyclone rates can give inaccurate impact predictions. We demonstrate this by exploring cyclone impacts on coral reefs. For a given cyclone rate, we find that clustered events have a less detrimental impact than independent random events. Predictions using a standard random hurricane model were overly pessimistic, predicting reef degradation more than a decade earlier than that expected under clustered disturbance. The presence of clustering allows coral reefs more time to recover to healthier states, but the impacts of clustering will vary from one ecosystem to another. PMID:22006300

  20. Mathematical modelling of complex contagion on clustered networks

    NASA Astrophysics Data System (ADS)

    O'sullivan, David J.; O'Keeffe, Gary; Fennell, Peter; Gleeson, James

    2015-09-01

    The spreading of behavior, such as the adoption of a new innovation, is influenced bythe structure of social networks that interconnect the population. In the experiments of Centola (Science, 2010), adoption of new behavior was shown to spread further and faster across clustered-lattice networks than across corresponding random networks. This implies that the “complex contagion” effects of social reinforcement are important in such diffusion, in contrast to “simple” contagion models of disease-spread which predict that epidemics would grow more efficiently on random networks than on clustered networks. To accurately model complex contagion on clustered networks remains a challenge because the usual assumptions (e.g. of mean-field theory) regarding tree-like networks are invalidated by the presence of triangles in the network; the triangles are, however, crucial to the social reinforcement mechanism, which posits an increased probability of a person adopting behavior that has been adopted by two or more neighbors. In this paper we modify the analytical approach that was introduced by Hebert-Dufresne et al. (Phys. Rev. E, 2010), to study disease-spread on clustered networks. We show how the approximation method can be adapted to a complex contagion model, and confirm the accuracy of the method with numerical simulations. The analytical results of the model enable us to quantify the level of social reinforcement that is required to observe—as in Centola’s experiments—faster diffusion on clustered topologies than on random networks.

  1. Quantifying opening-mode fracture spatial organization in horizontal wellbore image logs, core and outcrop: Application to Upper Cretaceous Frontier Formation tight gas sandstones, USA

    NASA Astrophysics Data System (ADS)

    Li, J. Z.; Laubach, S. E.; Gale, J. F. W.; Marrett, R. A.

    2018-03-01

    The Upper Cretaceous Frontier Formation is a naturally fractured gas-producing sandstone in Wyoming. Regionally, random and statistically more clustered than random patterns exist in the same upper to lower shoreface depositional facies. East-west- and north-south-striking regional fractures sampled using image logs and cores from three horizontal wells exhibit clustered patterns, whereas data collected from east-west-striking fractures in outcrop have patterns that are indistinguishable from random. Image log data analyzed with the correlation count method shows clusters ∼35 m wide and spaced ∼50 to 90 m apart as well as clusters up to 12 m wide with periodic inter-cluster spacings. A hierarchy of cluster sizes exists; organization within clusters is likely fractal. These rocks have markedly different structural and burial histories, so regional differences in degree of clustering are unsurprising. Clustered patterns correspond to fractures having core quartz deposition contemporaneous with fracture opening, circumstances that some models suggest might affect spacing patterns by interfering with fracture growth. Our results show that quantifying and identifying patterns as statistically more or less clustered than random delineates differences in fracture patterns that are not otherwise apparent but that may influence gas and water production, and therefore may be economically important.

  2. On aggregation in CA models in biology

    NASA Astrophysics Data System (ADS)

    Alber, Mark S.; Kiskowski, Audi

    2001-12-01

    Aggregation of randomly distributed particles into clusters of aligned particles is modeled using a cellular automata (CA) approach. The CA model accounts for interactions between more than one type of particle, in which pressures for angular alignment with neighbors compete with pressures for grouping by cell type. In the case of only one particle type clusters tend to unite into one big cluster. In the case of several types of particles the dynamics of clusters is more complicated and for specific choices of parameters particle sorting occurs simultaneously with the formation of clusters of aligned particles.

  3. Models of epidemics: when contact repetition and clustering should be included

    PubMed Central

    Smieszek, Timo; Fiebig, Lena; Scholz, Roland W

    2009-01-01

    Background The spread of infectious disease is determined by biological factors, e.g. the duration of the infectious period, and social factors, e.g. the arrangement of potentially contagious contacts. Repetitiveness and clustering of contacts are known to be relevant factors influencing the transmission of droplet or contact transmitted diseases. However, we do not yet completely know under what conditions repetitiveness and clustering should be included for realistically modelling disease spread. Methods We compare two different types of individual-based models: One assumes random mixing without repetition of contacts, whereas the other assumes that the same contacts repeat day-by-day. The latter exists in two variants, with and without clustering. We systematically test and compare how the total size of an outbreak differs between these model types depending on the key parameters transmission probability, number of contacts per day, duration of the infectious period, different levels of clustering and varying proportions of repetitive contacts. Results The simulation runs under different parameter constellations provide the following results: The difference between both model types is highest for low numbers of contacts per day and low transmission probabilities. The number of contacts and the transmission probability have a higher influence on this difference than the duration of the infectious period. Even when only minor parts of the daily contacts are repetitive and clustered can there be relevant differences compared to a purely random mixing model. Conclusion We show that random mixing models provide acceptable estimates of the total outbreak size if the number of contacts per day is high or if the per-contact transmission probability is high, as seen in typical childhood diseases such as measles. In the case of very short infectious periods, for instance, as in Norovirus, models assuming repeating contacts will also behave similarly as random mixing models. If the number of daily contacts or the transmission probability is low, as assumed for MRSA or Ebola, particular consideration should be given to the actual structure of potentially contagious contacts when designing the model. PMID:19563624

  4. Recommendations for choosing an analysis method that controls Type I error for unbalanced cluster sample designs with Gaussian outcomes.

    PubMed

    Johnson, Jacqueline L; Kreidler, Sarah M; Catellier, Diane J; Murray, David M; Muller, Keith E; Glueck, Deborah H

    2015-11-30

    We used theoretical and simulation-based approaches to study Type I error rates for one-stage and two-stage analytic methods for cluster-randomized designs. The one-stage approach uses the observed data as outcomes and accounts for within-cluster correlation using a general linear mixed model. The two-stage model uses the cluster specific means as the outcomes in a general linear univariate model. We demonstrate analytically that both one-stage and two-stage models achieve exact Type I error rates when cluster sizes are equal. With unbalanced data, an exact size α test does not exist, and Type I error inflation may occur. Via simulation, we compare the Type I error rates for four one-stage and six two-stage hypothesis testing approaches for unbalanced data. With unbalanced data, the two-stage model, weighted by the inverse of the estimated theoretical variance of the cluster means, and with variance constrained to be positive, provided the best Type I error control for studies having at least six clusters per arm. The one-stage model with Kenward-Roger degrees of freedom and unconstrained variance performed well for studies having at least 14 clusters per arm. The popular analytic method of using a one-stage model with denominator degrees of freedom appropriate for balanced data performed poorly for small sample sizes and low intracluster correlation. Because small sample sizes and low intracluster correlation are common features of cluster-randomized trials, the Kenward-Roger method is the preferred one-stage approach. Copyright © 2015 John Wiley & Sons, Ltd.

  5. Cascades on a class of clustered random networks

    NASA Astrophysics Data System (ADS)

    Hackett, Adam; Melnik, Sergey; Gleeson, James P.

    2011-05-01

    We present an analytical approach to determining the expected cascade size in a broad range of dynamical models on the class of random networks with arbitrary degree distribution and nonzero clustering introduced previously in [M. E. J. Newman, Phys. Rev. Lett. PRLTAO0031-900710.1103/PhysRevLett.103.058701103, 058701 (2009)]. A condition for the existence of global cascades is derived as well as a general criterion that determines whether increasing the level of clustering will increase, or decrease, the expected cascade size. Applications, examples of which are provided, include site percolation, bond percolation, and Watts’ threshold model; in all cases analytical results give excellent agreement with numerical simulations.

  6. Motivational Pathways to Leisure-Time Physical Activity Participation in Urban Physical Education: A Cluster-Randomized Trial

    ERIC Educational Resources Information Center

    Yli-Piipari, Sami; Layne, Todd; Hinson, Janet; Irwin, Carol

    2018-01-01

    Purpose: Grounded in the trans-contextual model of motivation framework, this cluster-randomized trial examined the effectiveness of an autonomy supportive physical education (PE) instruction on student motivation and physical activity (PA). Method: The study comprised six middle schools and 408 students (M[subscript age] = 12.29), with primary…

  7. Outcome-Driven Cluster Analysis with Application to Microarray Data.

    PubMed

    Hsu, Jessie J; Finkelstein, Dianne M; Schoenfeld, David A

    2015-01-01

    One goal of cluster analysis is to sort characteristics into groups (clusters) so that those in the same group are more highly correlated to each other than they are to those in other groups. An example is the search for groups of genes whose expression of RNA is correlated in a population of patients. These genes would be of greater interest if their common level of RNA expression were additionally predictive of the clinical outcome. This issue arose in the context of a study of trauma patients on whom RNA samples were available. The question of interest was whether there were groups of genes that were behaving similarly, and whether each gene in the cluster would have a similar effect on who would recover. For this, we develop an algorithm to simultaneously assign characteristics (genes) into groups of highly correlated genes that have the same effect on the outcome (recovery). We propose a random effects model where the genes within each group (cluster) equal the sum of a random effect, specific to the observation and cluster, and an independent error term. The outcome variable is a linear combination of the random effects of each cluster. To fit the model, we implement a Markov chain Monte Carlo algorithm based on the likelihood of the observed data. We evaluate the effect of including outcome in the model through simulation studies and describe a strategy for prediction. These methods are applied to trauma data from the Inflammation and Host Response to Injury research program, revealing a clustering of the genes that are informed by the recovery outcome.

  8. Modeling of correlated data with informative cluster sizes: An evaluation of joint modeling and within-cluster resampling approaches.

    PubMed

    Zhang, Bo; Liu, Wei; Zhang, Zhiwei; Qu, Yanping; Chen, Zhen; Albert, Paul S

    2017-08-01

    Joint modeling and within-cluster resampling are two approaches that are used for analyzing correlated data with informative cluster sizes. Motivated by a developmental toxicity study, we examined the performances and validity of these two approaches in testing covariate effects in generalized linear mixed-effects models. We show that the joint modeling approach is robust to the misspecification of cluster size models in terms of Type I and Type II errors when the corresponding covariates are not included in the random effects structure; otherwise, statistical tests may be affected. We also evaluate the performance of the within-cluster resampling procedure and thoroughly investigate the validity of it in modeling correlated data with informative cluster sizes. We show that within-cluster resampling is a valid alternative to joint modeling for cluster-specific covariates, but it is invalid for time-dependent covariates. The two methods are applied to a developmental toxicity study that investigated the effect of exposure to diethylene glycol dimethyl ether.

  9. A Bimodal Hybrid Model for Time-Dependent Probabilistic Seismic Hazard Analysis

    NASA Astrophysics Data System (ADS)

    Yaghmaei-Sabegh, Saman; Shoaeifar, Nasser; Shoaeifar, Parva

    2018-03-01

    The evaluation of evidence provided by geological studies and historical catalogs indicates that in some seismic regions and faults, multiple large earthquakes occur in cluster. Then, the occurrences of large earthquakes confront with quiescence and only the small-to-moderate earthquakes take place. Clustering of large earthquakes is the most distinguishable departure from the assumption of constant hazard of random occurrence of earthquakes in conventional seismic hazard analysis. In the present study, a time-dependent recurrence model is proposed to consider a series of large earthquakes that occurs in clusters. The model is flexible enough to better reflect the quasi-periodic behavior of large earthquakes with long-term clustering, which can be used in time-dependent probabilistic seismic hazard analysis with engineering purposes. In this model, the time-dependent hazard results are estimated by a hazard function which comprises three parts. A decreasing hazard of last large earthquake cluster and an increasing hazard of the next large earthquake cluster, along with a constant hazard of random occurrence of small-to-moderate earthquakes. In the final part of the paper, the time-dependent seismic hazard of the New Madrid Seismic Zone at different time intervals has been calculated for illustrative purpose.

  10. Relaxation dynamics of maximally clustered networks

    NASA Astrophysics Data System (ADS)

    Klaise, Janis; Johnson, Samuel

    2018-01-01

    We study the relaxation dynamics of fully clustered networks (maximal number of triangles) to an unclustered state under two different edge dynamics—the double-edge swap, corresponding to degree-preserving randomization of the configuration model, and single edge replacement, corresponding to full randomization of the Erdős-Rényi random graph. We derive expressions for the time evolution of the degree distribution, edge multiplicity distribution and clustering coefficient. We show that under both dynamics networks undergo a continuous phase transition in which a giant connected component is formed. We calculate the position of the phase transition analytically using the Erdős-Rényi phenomenology.

  11. Cluster-cluster correlations and constraints on the correlation hierarchy

    NASA Technical Reports Server (NTRS)

    Hamilton, A. J. S.; Gott, J. R., III

    1988-01-01

    The hypothesis that galaxies cluster around clusters at least as strongly as they cluster around galaxies imposes constraints on the hierarchy of correlation amplitudes in hierachical clustering models. The distributions which saturate these constraints are the Rayleigh-Levy random walk fractals proposed by Mandelbrot; for these fractal distributions cluster-cluster correlations are all identically equal to galaxy-galaxy correlations. If correlation amplitudes exceed the constraints, as is observed, then cluster-cluster correlations must exceed galaxy-galaxy correlations, as is observed.

  12. Spread of information and infection on finite random networks

    NASA Astrophysics Data System (ADS)

    Isham, Valerie; Kaczmarska, Joanna; Nekovee, Maziar

    2011-04-01

    The modeling of epidemic-like processes on random networks has received considerable attention in recent years. While these processes are inherently stochastic, most previous work has been focused on deterministic models that ignore important fluctuations that may persist even in the infinite network size limit. In a previous paper, for a class of epidemic and rumor processes, we derived approximate models for the full probability distribution of the final size of the epidemic, as opposed to only mean values. In this paper we examine via direct simulations the adequacy of the approximate model to describe stochastic epidemics and rumors on several random network topologies: homogeneous networks, Erdös-Rényi (ER) random graphs, Barabasi-Albert scale-free networks, and random geometric graphs. We find that the approximate model is reasonably accurate in predicting the probability of spread. However, the position of the threshold and the conditional mean of the final size for processes near the threshold are not well described by the approximate model even in the case of homogeneous networks. We attribute this failure to the presence of other structural properties beyond degree-degree correlations, and in particular clustering, which are present in any finite network but are not incorporated in the approximate model. In order to test this “hypothesis” we perform additional simulations on a set of ER random graphs where degree-degree correlations and clustering are separately and independently introduced using recently proposed algorithms from the literature. Our results show that even strong degree-degree correlations have only weak effects on the position of the threshold and the conditional mean of the final size. On the other hand, the introduction of clustering greatly affects both the position of the threshold and the conditional mean. Similar analysis for the Barabasi-Albert scale-free network confirms the significance of clustering on the dynamics of rumor spread. For this network, though, with its highly skewed degree distribution, the addition of positive correlation had a much stronger effect on the final size distribution than was found for the simple random graph.

  13. Regional SAR Image Segmentation Based on Fuzzy Clustering with Gamma Mixture Model

    NASA Astrophysics Data System (ADS)

    Li, X. L.; Zhao, Q. H.; Li, Y.

    2017-09-01

    Most of stochastic based fuzzy clustering algorithms are pixel-based, which can not effectively overcome the inherent speckle noise in SAR images. In order to deal with the problem, a regional SAR image segmentation algorithm based on fuzzy clustering with Gamma mixture model is proposed in this paper. First, initialize some generating points randomly on the image, the image domain is divided into many sub-regions using Voronoi tessellation technique. Each sub-region is regarded as a homogeneous area in which the pixels share the same cluster label. Then, assume the probability of the pixel to be a Gamma mixture model with the parameters respecting to the cluster which the pixel belongs to. The negative logarithm of the probability represents the dissimilarity measure between the pixel and the cluster. The regional dissimilarity measure of one sub-region is defined as the sum of the measures of pixels in the region. Furthermore, the Markov Random Field (MRF) model is extended from pixels level to Voronoi sub-regions, and then the regional objective function is established under the framework of fuzzy clustering. The optimal segmentation results can be obtained by the solution of model parameters and generating points. Finally, the effectiveness of the proposed algorithm can be proved by the qualitative and quantitative analysis from the segmentation results of the simulated and real SAR images.

  14. Assessing variation in life-history tactics within a population using mixture regression models: a practical guide for evolutionary ecologists.

    PubMed

    Hamel, Sandra; Yoccoz, Nigel G; Gaillard, Jean-Michel

    2017-05-01

    Mixed models are now well-established methods in ecology and evolution because they allow accounting for and quantifying within- and between-individual variation. However, the required normal distribution of the random effects can often be violated by the presence of clusters among subjects, which leads to multi-modal distributions. In such cases, using what is known as mixture regression models might offer a more appropriate approach. These models are widely used in psychology, sociology, and medicine to describe the diversity of trajectories occurring within a population over time (e.g. psychological development, growth). In ecology and evolution, however, these models are seldom used even though understanding changes in individual trajectories is an active area of research in life-history studies. Our aim is to demonstrate the value of using mixture models to describe variation in individual life-history tactics within a population, and hence to promote the use of these models by ecologists and evolutionary ecologists. We first ran a set of simulations to determine whether and when a mixture model allows teasing apart latent clustering, and to contrast the precision and accuracy of estimates obtained from mixture models versus mixed models under a wide range of ecological contexts. We then used empirical data from long-term studies of large mammals to illustrate the potential of using mixture models for assessing within-population variation in life-history tactics. Mixture models performed well in most cases, except for variables following a Bernoulli distribution and when sample size was small. The four selection criteria we evaluated [Akaike information criterion (AIC), Bayesian information criterion (BIC), and two bootstrap methods] performed similarly well, selecting the right number of clusters in most ecological situations. We then showed that the normality of random effects implicitly assumed by evolutionary ecologists when using mixed models was often violated in life-history data. Mixed models were quite robust to this violation in the sense that fixed effects were unbiased at the population level. However, fixed effects at the cluster level and random effects were better estimated using mixture models. Our empirical analyses demonstrated that using mixture models facilitates the identification of the diversity of growth and reproductive tactics occurring within a population. Therefore, using this modelling framework allows testing for the presence of clusters and, when clusters occur, provides reliable estimates of fixed and random effects for each cluster of the population. In the presence or expectation of clusters, using mixture models offers a suitable extension of mixed models, particularly when evolutionary ecologists aim at identifying how ecological and evolutionary processes change within a population. Mixture regression models therefore provide a valuable addition to the statistical toolbox of evolutionary ecologists. As these models are complex and have their own limitations, we provide recommendations to guide future users. © 2016 Cambridge Philosophical Society.

  15. A Tutorial on Multilevel Survival Analysis: Methods, Models and Applications

    PubMed Central

    Austin, Peter C.

    2017-01-01

    Summary Data that have a multilevel structure occur frequently across a range of disciplines, including epidemiology, health services research, public health, education and sociology. We describe three families of regression models for the analysis of multilevel survival data. First, Cox proportional hazards models with mixed effects incorporate cluster-specific random effects that modify the baseline hazard function. Second, piecewise exponential survival models partition the duration of follow-up into mutually exclusive intervals and fit a model that assumes that the hazard function is constant within each interval. This is equivalent to a Poisson regression model that incorporates the duration of exposure within each interval. By incorporating cluster-specific random effects, generalised linear mixed models can be used to analyse these data. Third, after partitioning the duration of follow-up into mutually exclusive intervals, one can use discrete time survival models that use a complementary log–log generalised linear model to model the occurrence of the outcome of interest within each interval. Random effects can be incorporated to account for within-cluster homogeneity in outcomes. We illustrate the application of these methods using data consisting of patients hospitalised with a heart attack. We illustrate the application of these methods using three statistical programming languages (R, SAS and Stata). PMID:29307954

  16. A Tutorial on Multilevel Survival Analysis: Methods, Models and Applications.

    PubMed

    Austin, Peter C

    2017-08-01

    Data that have a multilevel structure occur frequently across a range of disciplines, including epidemiology, health services research, public health, education and sociology. We describe three families of regression models for the analysis of multilevel survival data. First, Cox proportional hazards models with mixed effects incorporate cluster-specific random effects that modify the baseline hazard function. Second, piecewise exponential survival models partition the duration of follow-up into mutually exclusive intervals and fit a model that assumes that the hazard function is constant within each interval. This is equivalent to a Poisson regression model that incorporates the duration of exposure within each interval. By incorporating cluster-specific random effects, generalised linear mixed models can be used to analyse these data. Third, after partitioning the duration of follow-up into mutually exclusive intervals, one can use discrete time survival models that use a complementary log-log generalised linear model to model the occurrence of the outcome of interest within each interval. Random effects can be incorporated to account for within-cluster homogeneity in outcomes. We illustrate the application of these methods using data consisting of patients hospitalised with a heart attack. We illustrate the application of these methods using three statistical programming languages (R, SAS and Stata).

  17. The topology of large-scale structure. I - Topology and the random phase hypothesis. [galactic formation models

    NASA Technical Reports Server (NTRS)

    Weinberg, David H.; Gott, J. Richard, III; Melott, Adrian L.

    1987-01-01

    Many models for the formation of galaxies and large-scale structure assume a spectrum of random phase (Gaussian), small-amplitude density fluctuations as initial conditions. In such scenarios, the topology of the galaxy distribution on large scales relates directly to the topology of the initial density fluctuations. Here a quantitative measure of topology - the genus of contours in a smoothed density distribution - is described and applied to numerical simulations of galaxy clustering, to a variety of three-dimensional toy models, and to a volume-limited sample of the CfA redshift survey. For random phase distributions the genus of density contours exhibits a universal dependence on threshold density. The clustering simulations show that a smoothing length of 2-3 times the mass correlation length is sufficient to recover the topology of the initial fluctuations from the evolved galaxy distribution. Cold dark matter and white noise models retain a random phase topology at shorter smoothing lengths, but massive neutrino models develop a cellular topology.

  18. Application of Multiple Imputation for Missing Values in Three-Way Three-Mode Multi-Environment Trial Data

    PubMed Central

    Tian, Ting; McLachlan, Geoffrey J.; Dieters, Mark J.; Basford, Kaye E.

    2015-01-01

    It is a common occurrence in plant breeding programs to observe missing values in three-way three-mode multi-environment trial (MET) data. We proposed modifications of models for estimating missing observations for these data arrays, and developed a novel approach in terms of hierarchical clustering. Multiple imputation (MI) was used in four ways, multiple agglomerative hierarchical clustering, normal distribution model, normal regression model, and predictive mean match. The later three models used both Bayesian analysis and non-Bayesian analysis, while the first approach used a clustering procedure with randomly selected attributes and assigned real values from the nearest neighbour to the one with missing observations. Different proportions of data entries in six complete datasets were randomly selected to be missing and the MI methods were compared based on the efficiency and accuracy of estimating those values. The results indicated that the models using Bayesian analysis had slightly higher accuracy of estimation performance than those using non-Bayesian analysis but they were more time-consuming. However, the novel approach of multiple agglomerative hierarchical clustering demonstrated the overall best performances. PMID:26689369

  19. Application of Multiple Imputation for Missing Values in Three-Way Three-Mode Multi-Environment Trial Data.

    PubMed

    Tian, Ting; McLachlan, Geoffrey J; Dieters, Mark J; Basford, Kaye E

    2015-01-01

    It is a common occurrence in plant breeding programs to observe missing values in three-way three-mode multi-environment trial (MET) data. We proposed modifications of models for estimating missing observations for these data arrays, and developed a novel approach in terms of hierarchical clustering. Multiple imputation (MI) was used in four ways, multiple agglomerative hierarchical clustering, normal distribution model, normal regression model, and predictive mean match. The later three models used both Bayesian analysis and non-Bayesian analysis, while the first approach used a clustering procedure with randomly selected attributes and assigned real values from the nearest neighbour to the one with missing observations. Different proportions of data entries in six complete datasets were randomly selected to be missing and the MI methods were compared based on the efficiency and accuracy of estimating those values. The results indicated that the models using Bayesian analysis had slightly higher accuracy of estimation performance than those using non-Bayesian analysis but they were more time-consuming. However, the novel approach of multiple agglomerative hierarchical clustering demonstrated the overall best performances.

  20. A Simulation Study Comparing Epidemic Dynamics on Exponential Random Graph and Edge-Triangle Configuration Type Contact Network Models

    PubMed Central

    Rolls, David A.; Wang, Peng; McBryde, Emma; Pattison, Philippa; Robins, Garry

    2015-01-01

    We compare two broad types of empirically grounded random network models in terms of their abilities to capture both network features and simulated Susceptible-Infected-Recovered (SIR) epidemic dynamics. The types of network models are exponential random graph models (ERGMs) and extensions of the configuration model. We use three kinds of empirical contact networks, chosen to provide both variety and realistic patterns of human contact: a highly clustered network, a bipartite network and a snowball sampled network of a “hidden population”. In the case of the snowball sampled network we present a novel method for fitting an edge-triangle model. In our results, ERGMs consistently capture clustering as well or better than configuration-type models, but the latter models better capture the node degree distribution. Despite the additional computational requirements to fit ERGMs to empirical networks, the use of ERGMs provides only a slight improvement in the ability of the models to recreate epidemic features of the empirical network in simulated SIR epidemics. Generally, SIR epidemic results from using configuration-type models fall between those from a random network model (i.e., an Erdős-Rényi model) and an ERGM. The addition of subgraphs of size four to edge-triangle type models does improve agreement with the empirical network for smaller densities in clustered networks. Additional subgraphs do not make a noticeable difference in our example, although we would expect the ability to model cliques to be helpful for contact networks exhibiting household structure. PMID:26555701

  1. Testing prediction methods: Earthquake clustering versus the Poisson model

    USGS Publications Warehouse

    Michael, A.J.

    1997-01-01

    Testing earthquake prediction methods requires statistical techniques that compare observed success to random chance. One technique is to produce simulated earthquake catalogs and measure the relative success of predicting real and simulated earthquakes. The accuracy of these tests depends on the validity of the statistical model used to simulate the earthquakes. This study tests the effect of clustering in the statistical earthquake model on the results. Three simulation models were used to produce significance levels for a VLF earthquake prediction method. As the degree of simulated clustering increases, the statistical significance drops. Hence, the use of a seismicity model with insufficient clustering can lead to overly optimistic results. A successful method must pass the statistical tests with a model that fully replicates the observed clustering. However, a method can be rejected based on tests with a model that contains insufficient clustering. U.S. copyright. Published in 1997 by the American Geophysical Union.

  2. Magnetic cluster expansion model for random and ordered magnetic face-centered cubic Fe-Ni-Cr alloys

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lavrentiev, M. Yu., E-mail: Mikhail.Lavrentiev@ukaea.uk; Nguyen-Manh, D.; Dudarev, S. L.

    A Magnetic Cluster Expansion model for ternary face-centered cubic Fe-Ni-Cr alloys has been developed, using DFT data spanning binary and ternary alloy configurations. Using this Magnetic Cluster Expansion model Hamiltonian, we perform Monte Carlo simulations and explore magnetic structures of alloys over the entire range of compositions, considering both random and ordered alloy structures. In random alloys, the removal of magnetic collinearity constraint reduces the total magnetic moment but does not affect the predicted range of compositions where the alloys adopt low-temperature ferromagnetic configurations. During alloying of ordered fcc Fe-Ni compounds with Cr, chromium atoms tend to replace nickel rathermore » than iron atoms. Replacement of Ni by Cr in ordered alloys with high iron content increases the Curie temperature of the alloys. This can be explained by strong antiferromagnetic Fe-Cr coupling, similar to that found in bcc Fe-Cr solutions, where the Curie temperature increase, predicted by simulations as a function of Cr concentration, is confirmed by experimental observations. In random alloys, both magnetization and the Curie temperature decrease abruptly with increasing chromium content, in agreement with experiment.« less

  3. The formation of magnetic silicide Fe3Si clusters during ion implantation

    NASA Astrophysics Data System (ADS)

    Balakirev, N.; Zhikharev, V.; Gumarov, G.

    2014-05-01

    A simple two-dimensional model of the formation of magnetic silicide Fe3Si clusters during high-dose Fe ion implantation into silicon has been proposed and the cluster growth process has been computer simulated. The model takes into account the interaction between the cluster magnetization and magnetic moments of Fe atoms random walking in the implanted layer. If the clusters are formed in the presence of the external magnetic field parallel to the implanted layer, the model predicts the elongation of the growing cluster in the field direction. It has been proposed that the cluster elongation results in the uniaxial magnetic anisotropy in the plane of the implanted layer, which is observed in iron silicide films ion-beam synthesized in the external magnetic field.

  4. Epidemiological characteristics of reported sporadic and outbreak cases of E. coli O157 in people from Alberta, Canada (2000-2002): methodological challenges of comparing clustered to unclustered data.

    PubMed

    Pearl, D L; Louie, M; Chui, L; Doré, K; Grimsrud, K M; Martin, S W; Michel, P; Svenson, L W; McEwen, S A

    2008-04-01

    Using multivariable models, we compared whether there were significant differences between reported outbreak and sporadic cases in terms of their sex, age, and mode and site of disease transmission. We also determined the potential role of administrative, temporal, and spatial factors within these models. We compared a variety of approaches to account for clustering of cases in outbreaks including weighted logistic regression, random effects models, general estimating equations, robust variance estimates, and the random selection of one case from each outbreak. Age and mode of transmission were the only epidemiologically and statistically significant covariates in our final models using the above approaches. Weighing observations in a logistic regression model by the inverse of their outbreak size appeared to be a relatively robust and valid means for modelling these data. Some analytical techniques, designed to account for clustering, had difficulty converging or producing realistic measures of association.

  5. Multilevel covariance regression with correlated random effects in the mean and variance structure.

    PubMed

    Quintero, Adrian; Lesaffre, Emmanuel

    2017-09-01

    Multivariate regression methods generally assume a constant covariance matrix for the observations. In case a heteroscedastic model is needed, the parametric and nonparametric covariance regression approaches can be restrictive in the literature. We propose a multilevel regression model for the mean and covariance structure, including random intercepts in both components and allowing for correlation between them. The implied conditional covariance function can be different across clusters as a result of the random effect in the variance structure. In addition, allowing for correlation between the random intercepts in the mean and covariance makes the model convenient for skewedly distributed responses. Furthermore, it permits us to analyse directly the relation between the mean response level and the variability in each cluster. Parameter estimation is carried out via Gibbs sampling. We compare the performance of our model to other covariance modelling approaches in a simulation study. Finally, the proposed model is applied to the RN4CAST dataset to identify the variables that impact burnout of nurses in Belgium. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. Comparing cluster-level dynamic treatment regimens using sequential, multiple assignment, randomized trials: Regression estimation and sample size considerations.

    PubMed

    NeCamp, Timothy; Kilbourne, Amy; Almirall, Daniel

    2017-08-01

    Cluster-level dynamic treatment regimens can be used to guide sequential treatment decision-making at the cluster level in order to improve outcomes at the individual or patient-level. In a cluster-level dynamic treatment regimen, the treatment is potentially adapted and re-adapted over time based on changes in the cluster that could be impacted by prior intervention, including aggregate measures of the individuals or patients that compose it. Cluster-randomized sequential multiple assignment randomized trials can be used to answer multiple open questions preventing scientists from developing high-quality cluster-level dynamic treatment regimens. In a cluster-randomized sequential multiple assignment randomized trial, sequential randomizations occur at the cluster level and outcomes are observed at the individual level. This manuscript makes two contributions to the design and analysis of cluster-randomized sequential multiple assignment randomized trials. First, a weighted least squares regression approach is proposed for comparing the mean of a patient-level outcome between the cluster-level dynamic treatment regimens embedded in a sequential multiple assignment randomized trial. The regression approach facilitates the use of baseline covariates which is often critical in the analysis of cluster-level trials. Second, sample size calculators are derived for two common cluster-randomized sequential multiple assignment randomized trial designs for use when the primary aim is a between-dynamic treatment regimen comparison of the mean of a continuous patient-level outcome. The methods are motivated by the Adaptive Implementation of Effective Programs Trial which is, to our knowledge, the first-ever cluster-randomized sequential multiple assignment randomized trial in psychiatry.

  7. Estimators for Clustered Education RCTs Using the Neyman Model for Causal Inference

    ERIC Educational Resources Information Center

    Schochet, Peter Z.

    2013-01-01

    This article examines the estimation of two-stage clustered designs for education randomized control trials (RCTs) using the nonparametric Neyman causal inference framework that underlies experiments. The key distinction between the considered causal models is whether potential treatment and control group outcomes are considered to be fixed for…

  8. Choosing appropriate analysis methods for cluster randomised cross-over trials with a binary outcome.

    PubMed

    Morgan, Katy E; Forbes, Andrew B; Keogh, Ruth H; Jairath, Vipul; Kahan, Brennan C

    2017-01-30

    In cluster randomised cross-over (CRXO) trials, clusters receive multiple treatments in a randomised sequence over time. In such trials, there is usual correlation between patients in the same cluster. In addition, within a cluster, patients in the same period may be more similar to each other than to patients in other periods. We demonstrate that it is necessary to account for these correlations in the analysis to obtain correct Type I error rates. We then use simulation to compare different methods of analysing a binary outcome from a two-period CRXO design. Our simulations demonstrated that hierarchical models without random effects for period-within-cluster, which do not account for any extra within-period correlation, performed poorly with greatly inflated Type I errors in many scenarios. In scenarios where extra within-period correlation was present, a hierarchical model with random effects for cluster and period-within-cluster only had correct Type I errors when there were large numbers of clusters; with small numbers of clusters, the error rate was inflated. We also found that generalised estimating equations did not give correct error rates in any scenarios considered. An unweighted cluster-level summary regression performed best overall, maintaining an error rate close to 5% for all scenarios, although it lost power when extra within-period correlation was present, especially for small numbers of clusters. Results from our simulation study show that it is important to model both levels of clustering in CRXO trials, and that any extra within-period correlation should be accounted for. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  9. Determining the impact of a new physiotherapist-led primary care model for back pain: protocol for a pilot cluster randomized controlled trial.

    PubMed

    Miller, Jordan; Barber, David; Donnelly, Catherine; French, Simon; Green, Michael; Hill, Jonathan; MacDermid, Joy; Marsh, Jacquelyn; Norman, Kathleen; Richardson, Julie; Taljaard, Monica; Wideman, Timothy; Cooper, Lynn; McPhee, Colleen

    2017-11-09

    Back pain is a leading contributor to disability, healthcare costs, and lost work. Family physicians are the most common first point of contact in the healthcare system for people with back pain, but physiotherapists (PTs) may be able to support the primary care team through evidence-based primary care. A cluster randomized trial is needed to determine the clinical, health system, and societal impact of a primary care model that integrates physiotherapists at the first visit for people with back pain. Prior to conducting a future fully powered cluster randomized trial, we need to demonstrate feasibility of the methods. Therefore, the purpose of this pilot study will be to: 1) Determine feasibility of patient recruitment, assessment procedures, and retention. 2) Determine the feasibility of training and implementation of a new PT-led primary care model for low back pain (LBP) 3) Explore the perspectives of patients and healthcare providers (HCPs) related to their experiences and attitudes towards the new service delivery model, barriers/facilitators to implementation, perceived satisfaction, perceived value, and impact on clinic processes and patient outcomes. This pilot cluster randomized controlled trial will enroll four sites and randomize them to implement a new PT-led primary care model for back pain or a usual physician-led primary care model. All adults booking a primary care visit for back pain will be invited to participate. Feasibility outcomes will include: recruitment and retention rates, completeness of assessment data, PT training participation and confidence after training, and PT treatment fidelity. Secondary outcomes will include the clinical, health system, cost, and process outcomes planned for the future fully powered cluster trial. Results will be analyzed and reported descriptively and qualitatively. To explore perspectives of both HCPs and patients, we will conduct semi-structured qualitative interviews with patients and focus groups with HCPs from participants in the PT-led primary care sites. If this pilot demonstrates feasibility, a fully powered trial will provide evidence that has the potential to transform primary care for back pain. The full trial will inform future service design, whether these models should be more widely implemented, and training agendas. ClinicalTrials.gov, NCT03320148 . Submitted for registration on 17 September 2017.

  10. Grouping by proximity and the visual impression of approximate number in random dot arrays.

    PubMed

    Im, Hee Yeon; Zhong, Sheng-Hua; Halberda, Justin

    2016-09-01

    We address the challenges of how to model human perceptual grouping in random dot arrays and how perceptual grouping affects human number estimation in these arrays. We introduce a modeling approach relying on a modified k-means clustering algorithm to formally describe human observers' grouping behavior. We found that a default grouping window size of approximately 4° of visual angle describes human grouping judgments across a range of random dot arrays (i.e., items within 4° are grouped together). This window size was highly consistent across observers and images, and was also stable across stimulus durations, suggesting that the k-means model captured a robust signature of perceptual grouping. Further, the k-means model outperformed other models (e.g., CODE) at describing human grouping behavior. Next, we found that the more the dots in a display are clustered together, the more human observers tend to underestimate the numerosity of the dots. We demonstrate that this effect is independent of density, and the modified k-means model can predict human observers' numerosity judgments and underestimation. Finally, we explored the robustness of the relationship between clustering and dot number underestimation and found that the effects of clustering remain, but are greatly reduced, when participants receive feedback on every trial. Together, this work suggests some promising avenues for formal models of human grouping behavior, and it highlights the importance of a 4° window of perceptual grouping. Lastly, it reveals a robust, somewhat plastic, relationship between perceptual grouping and number estimation. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. A two-stage model of fracture of rocks

    USGS Publications Warehouse

    Kuksenko, V.; Tomilin, N.; Damaskinskaya, E.; Lockner, D.

    1996-01-01

    In this paper we propose a two-stage model of rock fracture. In the first stage, cracks or local regions of failure are uncorrelated occur randomly throughout the rock in response to loading of pre-existing flaws. As damage accumulates in the rock, there is a gradual increase in the probability that large clusters of closely spaced cracks or local failure sites will develop. Based on statistical arguments, a critical density of damage will occur where clusters of flaws become large enough to lead to larger-scale failure of the rock (stage two). While crack interaction and cooperative failure is expected to occur within clusters of closely spaced cracks, the initial development of clusters is predicted based on the random variation in pre-existing Saw populations. Thus the onset of the unstable second stage in the model can be computed from the generation of random, uncorrelated damage. The proposed model incorporates notions of the kinetic (and therefore time-dependent) nature of the strength of solids as well as the discrete hierarchic structure of rocks and the flaw populations that lead to damage accumulation. The advantage offered by this model is that its salient features are valid for fracture processes occurring over a wide range of scales including earthquake processes. A notion of the rank of fracture (fracture size) is introduced, and criteria are presented for both fracture nucleation and the transition of the failure process from one scale to another.

  12. Quantifying Biomass from Point Clouds by Connecting Representations of Ecosystem Structure

    NASA Astrophysics Data System (ADS)

    Hendryx, S. M.; Barron-Gafford, G.

    2017-12-01

    Quantifying terrestrial ecosystem biomass is an essential part of monitoring carbon stocks and fluxes within the global carbon cycle and optimizing natural resource management. Point cloud data such as from lidar and structure from motion can be effective for quantifying biomass over large areas, but significant challenges remain in developing effective models that allow for such predictions. Inference models that estimate biomass from point clouds are established in many environments, yet, are often scale-dependent, needing to be fitted and applied at the same spatial scale and grid size at which they were developed. Furthermore, training such models typically requires large in situ datasets that are often prohibitively costly or time-consuming to obtain. We present here a scale- and sensor-invariant framework for efficiently estimating biomass from point clouds. Central to this framework, we present a new algorithm, assignPointsToExistingClusters, that has been developed for finding matches between in situ data and clusters in remotely-sensed point clouds. The algorithm can be used for assessing canopy segmentation accuracy and for training and validating machine learning models for predicting biophysical variables. We demonstrate the algorithm's efficacy by using it to train a random forest model of above ground biomass in a shrubland environment in Southern Arizona. We show that by learning a nonlinear function to estimate biomass from segmented canopy features we can reduce error, especially in the presence of inaccurate clusterings, when compared to a traditional, deterministic technique to estimate biomass from remotely measured canopies. Our random forest on cluster features model extends established methods of training random forest regressions to predict biomass of subplots but requires significantly less training data and is scale invariant. The random forest on cluster features model reduced mean absolute error, when evaluated on all test data in leave one out cross validation, by 40.6% from deterministic mesquite allometry and 35.9% from the inferred ecosystem-state allometric function. Our framework should allow for the inference of biomass more efficiently than common subplot methods and more accurately than individual tree segmentation methods in densely vegetated environments.

  13. Relative efficiency of unequal versus equal cluster sizes in cluster randomized trials using generalized estimating equation models.

    PubMed

    Liu, Jingxia; Colditz, Graham A

    2018-05-01

    There is growing interest in conducting cluster randomized trials (CRTs). For simplicity in sample size calculation, the cluster sizes are assumed to be identical across all clusters. However, equal cluster sizes are not guaranteed in practice. Therefore, the relative efficiency (RE) of unequal versus equal cluster sizes has been investigated when testing the treatment effect. One of the most important approaches to analyze a set of correlated data is the generalized estimating equation (GEE) proposed by Liang and Zeger, in which the "working correlation structure" is introduced and the association pattern depends on a vector of association parameters denoted by ρ. In this paper, we utilize GEE models to test the treatment effect in a two-group comparison for continuous, binary, or count data in CRTs. The variances of the estimator of the treatment effect are derived for the different types of outcome. RE is defined as the ratio of variance of the estimator of the treatment effect for equal to unequal cluster sizes. We discuss a commonly used structure in CRTs-exchangeable, and derive the simpler formula of RE with continuous, binary, and count outcomes. Finally, REs are investigated for several scenarios of cluster size distributions through simulation studies. We propose an adjusted sample size due to efficiency loss. Additionally, we also propose an optimal sample size estimation based on the GEE models under a fixed budget for known and unknown association parameter (ρ) in the working correlation structure within the cluster. © 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  14. MCMC Sampling for a Multilevel Model with Nonindependent Residuals within and between Cluster Units

    ERIC Educational Resources Information Center

    Browne, William; Goldstein, Harvey

    2010-01-01

    In this article, we discuss the effect of removing the independence assumptions between the residuals in two-level random effect models. We first consider removing the independence between the Level 2 residuals and instead assume that the vector of all residuals at the cluster level follows a general multivariate normal distribution. We…

  15. Confidence intervals for a difference between lognormal means in cluster randomization trials.

    PubMed

    Poirier, Julia; Zou, G Y; Koval, John

    2017-04-01

    Cluster randomization trials, in which intact social units are randomized to different interventions, have become popular in the last 25 years. Outcomes from these trials in many cases are positively skewed, following approximately lognormal distributions. When inference is focused on the difference between treatment arm arithmetic means, existent confidence interval procedures either make restricting assumptions or are complex to implement. We approach this problem by assuming log-transformed outcomes from each treatment arm follow a one-way random effects model. The treatment arm means are functions of multiple parameters for which separate confidence intervals are readily available, suggesting that the method of variance estimates recovery may be applied to obtain closed-form confidence intervals. A simulation study showed that this simple approach performs well in small sample sizes in terms of empirical coverage, relatively balanced tail errors, and interval widths as compared to existing methods. The methods are illustrated using data arising from a cluster randomization trial investigating a critical pathway for the treatment of community acquired pneumonia.

  16. Small Modifications to Network Topology Can Induce Stochastic Bistable Spiking Dynamics in a Balanced Cortical Model

    PubMed Central

    McDonnell, Mark D.; Ward, Lawrence M.

    2014-01-01

    Abstract Directed random graph models frequently are used successfully in modeling the population dynamics of networks of cortical neurons connected by chemical synapses. Experimental results consistently reveal that neuronal network topology is complex, however, in the sense that it differs statistically from a random network, and differs for classes of neurons that are physiologically different. This suggests that complex network models whose subnetworks have distinct topological structure may be a useful, and more biologically realistic, alternative to random networks. Here we demonstrate that the balanced excitation and inhibition frequently observed in small cortical regions can transiently disappear in otherwise standard neuronal-scale models of fluctuation-driven dynamics, solely because the random network topology was replaced by a complex clustered one, whilst not changing the in-degree of any neurons. In this network, a small subset of cells whose inhibition comes only from outside their local cluster are the cause of bistable population dynamics, where different clusters of these cells irregularly switch back and forth from a sparsely firing state to a highly active state. Transitions to the highly active state occur when a cluster of these cells spikes sufficiently often to cause strong unbalanced positive feedback to each other. Transitions back to the sparsely firing state rely on occasional large fluctuations in the amount of non-local inhibition received. Neurons in the model are homogeneous in their intrinsic dynamics and in-degrees, but differ in the abundance of various directed feedback motifs in which they participate. Our findings suggest that (i) models and simulations should take into account complex structure that varies for neuron and synapse classes; (ii) differences in the dynamics of neurons with similar intrinsic properties may be caused by their membership in distinctive local networks; (iii) it is important to identify neurons that share physiological properties and location, but differ in their connectivity. PMID:24743633

  17. Individualization as Driving Force of Clustering Phenomena in Humans

    PubMed Central

    Mäs, Michael; Flache, Andreas; Helbing, Dirk

    2010-01-01

    One of the most intriguing dynamics in biological systems is the emergence of clustering, in the sense that individuals self-organize into separate agglomerations in physical or behavioral space. Several theories have been developed to explain clustering in, for instance, multi-cellular organisms, ant colonies, bee hives, flocks of birds, schools of fish, and animal herds. A persistent puzzle, however, is the clustering of opinions in human populations, particularly when opinions vary continuously, such as the degree to which citizens are in favor of or against a vaccination program. Existing continuous opinion formation models predict “monoculture” in the long run, unless subsets of the population are perfectly separated from each other. Yet, social diversity is a robust empirical phenomenon, although perfect separation is hardly possible in an increasingly connected world. Considering randomness has not overcome the theoretical shortcomings so far. Small perturbations of individual opinions trigger social influence cascades that inevitably lead to monoculture, while larger noise disrupts opinion clusters and results in rampant individualism without any social structure. Our solution to the puzzle builds on recent empirical research, combining the integrative tendencies of social influence with the disintegrative effects of individualization. A key element of the new computational model is an adaptive kind of noise. We conduct computer simulation experiments demonstrating that with this kind of noise a third phase besides individualism and monoculture becomes possible, characterized by the formation of metastable clusters with diversity between and consensus within clusters. When clusters are small, individualization tendencies are too weak to prohibit a fusion of clusters. When clusters grow too large, however, individualization increases in strength, which promotes their splitting. In summary, the new model can explain cultural clustering in human societies. Strikingly, model predictions are not only robust to “noise”—randomness is actually the central mechanism that sustains pluralism and clustering. PMID:20975937

  18. Power Calculations for Moderators in Multi-Site Cluster Randomized Trials

    ERIC Educational Resources Information Center

    Spybrook, Jessaca; Kelcey, Ben; Dong, Nianbo

    2016-01-01

    Cluster randomized trials (CRTs), or studies in which intact groups of individuals are randomly assigned to a condition, are becoming more common in evaluation studies of educational programs. A specific type of CRT in which clusters are randomly assigned to treatment within blocks or sites, known as multisite cluster randomized trials (MSCRTs),…

  19. Estimation of Complex Generalized Linear Mixed Models for Measurement and Growth

    ERIC Educational Resources Information Center

    Jeon, Minjeong

    2012-01-01

    Maximum likelihood (ML) estimation of generalized linear mixed models (GLMMs) is technically challenging because of the intractable likelihoods that involve high dimensional integrations over random effects. The problem is magnified when the random effects have a crossed design and thus the data cannot be reduced to small independent clusters. A…

  20. Longitudinal Evaluation of a Scale-up Model for Teaching Mathematics with Trajectories and Technologies: Persistence of Effects in the Third Year

    ERIC Educational Resources Information Center

    Clements, Douglas H.; Sarama, Julie; Wolfe, Christopher B.; Spitler, Mary Elaine

    2013-01-01

    Using a cluster randomized trial design, we evaluated the persistence of effects of a research-based model for scaling up educational interventions. The model was implemented in 42 schools in two city districts serving low-resource communities, randomly assigned to three conditions. In pre-kindergarten, the two experimental interventions were…

  1. Computational lymphatic node models in pediatric and adult hybrid phantoms for radiation dosimetry

    NASA Astrophysics Data System (ADS)

    Lee, Choonsik; Lamart, Stephanie; Moroz, Brian E.

    2013-03-01

    We developed models of lymphatic nodes for six pediatric and two adult hybrid computational phantoms to calculate the lymphatic node dose estimates from external and internal radiation exposures. We derived the number of lymphatic nodes from the recommendations in International Commission on Radiological Protection (ICRP) Publications 23 and 89 at 16 cluster locations for the lymphatic nodes: extrathoracic, cervical, thoracic (upper and lower), breast (left and right), mesentery (left and right), axillary (left and right), cubital (left and right), inguinal (left and right) and popliteal (left and right), for different ages (newborn, 1-, 5-, 10-, 15-year-old and adult). We modeled each lymphatic node within the voxel format of the hybrid phantoms by assuming that all nodes have identical size derived from published data except narrow cluster sites. The lymph nodes were generated by the following algorithm: (1) selection of the lymph node site among the 16 cluster sites; (2) random sampling of the location of the lymph node within a spherical space centered at the chosen cluster site; (3) creation of the sphere or ovoid of tissue representing the node based on lymphatic node characteristics defined in ICRP Publications 23 and 89. We created lymph nodes until the pre-defined number of lymphatic nodes at the selected cluster site was reached. This algorithm was applied to pediatric (newborn, 1-, 5-and 10-year-old male, and 15-year-old males) and adult male and female ICRP-compliant hybrid phantoms after voxelization. To assess the performance of our models for internal dosimetry, we calculated dose conversion coefficients, called S values, for selected organs and tissues with Iodine-131 distributed in six lymphatic node cluster sites using MCNPX2.6, a well validated Monte Carlo radiation transport code. Our analysis of the calculations indicates that the S values were significantly affected by the location of the lymph node clusters and that the values increased for smaller phantoms due to the shorter inter-organ distances compared to the bigger phantoms. By testing sensitivity of S values to random sampling and voxel resolution, we confirmed that the lymph node model is reasonably stable and consistent for different random samplings and voxel resolutions.

  2. Three estimates of the association between linear growth failure and cognitive ability.

    PubMed

    Cheung, Y B; Lam, K F

    2009-09-01

    To compare three estimators of association between growth stunting as measured by height-for-age Z-score and cognitive ability in children, and to examine the extent statistical adjustment for covariates is useful for removing confounding due to socio-economic status. Three estimators, namely random-effects, within- and between-cluster estimators, for panel data were used to estimate the association in a survey of 1105 pairs of siblings who were assessed for anthropometry and cognition. Furthermore, a 'combined' model was formulated to simultaneously provide the within- and between-cluster estimates. Random-effects and between-cluster estimators showed strong association between linear growth and cognitive ability, even after adjustment for a range of socio-economic variables. In contrast, the within-cluster estimator showed a much more modest association: For every increase of one Z-score in linear growth, cognitive ability increased by about 0.08 standard deviation (P < 0.001). The combined model verified that the between-cluster estimate was significantly larger than the within-cluster estimate (P = 0.004). Residual confounding by socio-economic situations may explain a substantial proportion of the observed association between linear growth and cognition in studies that attempt to control the confounding by means of multivariable regression analysis. The within-cluster estimator provides more convincing and modest results about the strength of association.

  3. A new approach to hierarchical data analysis: Targeted maximum likelihood estimation for the causal effect of a cluster-level exposure.

    PubMed

    Balzer, Laura B; Zheng, Wenjing; van der Laan, Mark J; Petersen, Maya L

    2018-01-01

    We often seek to estimate the impact of an exposure naturally occurring or randomly assigned at the cluster-level. For example, the literature on neighborhood determinants of health continues to grow. Likewise, community randomized trials are applied to learn about real-world implementation, sustainability, and population effects of interventions with proven individual-level efficacy. In these settings, individual-level outcomes are correlated due to shared cluster-level factors, including the exposure, as well as social or biological interactions between individuals. To flexibly and efficiently estimate the effect of a cluster-level exposure, we present two targeted maximum likelihood estimators (TMLEs). The first TMLE is developed under a non-parametric causal model, which allows for arbitrary interactions between individuals within a cluster. These interactions include direct transmission of the outcome (i.e. contagion) and influence of one individual's covariates on another's outcome (i.e. covariate interference). The second TMLE is developed under a causal sub-model assuming the cluster-level and individual-specific covariates are sufficient to control for confounding. Simulations compare the alternative estimators and illustrate the potential gains from pairing individual-level risk factors and outcomes during estimation, while avoiding unwarranted assumptions. Our results suggest that estimation under the sub-model can result in bias and misleading inference in an observational setting. Incorporating working assumptions during estimation is more robust than assuming they hold in the underlying causal model. We illustrate our approach with an application to HIV prevention and treatment.

  4. Leveraging contact network structure in the design of cluster randomized trials.

    PubMed

    Harling, Guy; Wang, Rui; Onnela, Jukka-Pekka; De Gruttola, Victor

    2017-02-01

    In settings like the Ebola epidemic, where proof-of-principle trials have provided evidence of efficacy but questions remain about the effectiveness of different possible modes of implementation, it may be useful to conduct trials that not only generate information about intervention effects but also themselves provide public health benefit. Cluster randomized trials are of particular value for infectious disease prevention research by virtue of their ability to capture both direct and indirect effects of intervention, the latter of which depends heavily on the nature of contact networks within and across clusters. By leveraging information about these networks-in particular the degree of connection across randomized units, which can be obtained at study baseline-we propose a novel class of connectivity-informed cluster trial designs that aim both to improve public health impact (speed of epidemic control) and to preserve the ability to detect intervention effects. We several designs for cluster randomized trials with staggered enrollment, in each of which the order of enrollment is based on the total number of ties (contacts) from individuals within a cluster to individuals in other clusters. Our designs can accommodate connectivity based either on the total number of external connections at baseline or on connections only to areas yet to receive the intervention. We further consider a "holdback" version of the designs in which control clusters are held back from re-randomization for some time interval. We investigate the performance of these designs in terms of epidemic control outcomes (time to end of epidemic and cumulative incidence) and power to detect intervention effect, by simulating vaccination trials during an SEIR-type epidemic outbreak using a network-structured agent-based model. We compare results to those of a traditional Stepped Wedge trial. In our simulation studies, connectivity-informed designs lead to a 20% reduction in cumulative incidence compared to comparable traditional study designs, but have little impact on epidemic length. Power to detect intervention effect is reduced in all connectivity-informed designs, but "holdback" versions provide power that is very close to that of a traditional Stepped Wedge approach. Incorporating information about cluster connectivity in the design of cluster randomized trials can increase their public health impact, especially in acute outbreak settings. Using this information helps control outbreaks-by minimizing the number of cross-cluster infections-with very modest cost in terms of power to detect effectiveness.

  5. MIXREG: a computer program for mixed-effects regression analysis with autocorrelated errors.

    PubMed

    Hedeker, D; Gibbons, R D

    1996-05-01

    MIXREG is a program that provides estimates for a mixed-effects regression model (MRM) for normally-distributed response data including autocorrelated errors. This model can be used for analysis of unbalanced longitudinal data, where individuals may be measured at a different number of timepoints, or even at different timepoints. Autocorrelated errors of a general form or following an AR(1), MA(1), or ARMA(1,1) form are allowable. This model can also be used for analysis of clustered data, where the mixed-effects model assumes data within clusters are dependent. The degree of dependency is estimated jointly with estimates of the usual model parameters, thus adjusting for clustering. MIXREG uses maximum marginal likelihood estimation, utilizing both the EM algorithm and a Fisher-scoring solution. For the scoring solution, the covariance matrix of the random effects is expressed in its Gaussian decomposition, and the diagonal matrix reparameterized using the exponential transformation. Estimation of the individual random effects is accomplished using an empirical Bayes approach. Examples illustrating usage and features of MIXREG are provided.

  6. Radiative Feedback of Forming Star Clusters on Their GMC Environments: Theory and Simulation

    NASA Astrophysics Data System (ADS)

    Howard, C. S.; Pudritz, R. E.; Harris, W. E.

    2013-07-01

    Star clusters form from dense clumps within a molecular cloud. Radiation from these newly formed clusters feeds back on their natal molecular cloud through heating and ionization which ultimately stops gas accretion into the cluster. Recent studies suggest that radiative feedback effects from a single cluster may be sufficient to disrupt an entire cloud over a short timescale. Simulating cluster formation on a large scale, however, is computationally demanding due to the high number of stars involved. For this reason, we present a model for representing the radiative output of an entire cluster which involves randomly sampling an initial mass function (IMF) as the cluster accretes mass. We show that this model is able to reproduce the star formation histories of observed clusters. To examine the degree to which radiative feedback shapes the evolution of a molecular cloud, we use the FLASH adaptive-mesh refinement hydrodynamics code to simulate cluster formation in a turbulent cloud. Unlike previous studies, sink particles are used to represent a forming cluster rather than individual stars. Our cluster model is then coupled with a raytracing scheme to treat radiative transfer as the clusters grow in mass. This poster will outline the details of our model and present preliminary results from our 3D hydrodynamical simulations.

  7. Power and money in cluster randomized trials: when is it worth measuring a covariate?

    PubMed

    Moerbeek, Mirjam

    2006-08-15

    The power to detect a treatment effect in cluster randomized trials can be increased by increasing the number of clusters. An alternative is to include covariates into the regression model that relates treatment condition to outcome. In this paper, formulae are derived in order to evaluate both strategies on basis of their costs. It is shown that the strategy that uses covariates is more cost-efficient in detecting a treatment effect when the costs to measure these covariates are small and the correlation between the covariates and outcome is sufficiently large. The minimum required correlation depends on the cluster size, and the costs to recruit a cluster and to measure the covariate, relative to the costs to recruit a person. Measuring a covariate that varies at the person level only is recommended when cluster sizes are small and the costs to recruit and measure a cluster are large. Measuring a cluster level covariate is recommended when cluster sizes are large and the costs to recruit and measure a cluster are small. An illustrative example shows the use of the formulae in a practical setting. Copyright 2006 John Wiley & Sons, Ltd.

  8. Developing appropriate methods for cost-effectiveness analysis of cluster randomized trials.

    PubMed

    Gomes, Manuel; Ng, Edmond S-W; Grieve, Richard; Nixon, Richard; Carpenter, James; Thompson, Simon G

    2012-01-01

    Cost-effectiveness analyses (CEAs) may use data from cluster randomized trials (CRTs), where the unit of randomization is the cluster, not the individual. However, most studies use analytical methods that ignore clustering. This article compares alternative statistical methods for accommodating clustering in CEAs of CRTs. Our simulation study compared the performance of statistical methods for CEAs of CRTs with 2 treatment arms. The study considered a method that ignored clustering--seemingly unrelated regression (SUR) without a robust standard error (SE)--and 4 methods that recognized clustering--SUR and generalized estimating equations (GEEs), both with robust SE, a "2-stage" nonparametric bootstrap (TSB) with shrinkage correction, and a multilevel model (MLM). The base case assumed CRTs with moderate numbers of balanced clusters (20 per arm) and normally distributed costs. Other scenarios included CRTs with few clusters, imbalanced cluster sizes, and skewed costs. Performance was reported as bias, root mean squared error (rMSE), and confidence interval (CI) coverage for estimating incremental net benefits (INBs). We also compared the methods in a case study. Each method reported low levels of bias. Without the robust SE, SUR gave poor CI coverage (base case: 0.89 v. nominal level: 0.95). The MLM and TSB performed well in each scenario (CI coverage, 0.92-0.95). With few clusters, the GEE and SUR (with robust SE) had coverage below 0.90. In the case study, the mean INBs were similar across all methods, but ignoring clustering underestimated statistical uncertainty and the value of further research. MLMs and the TSB are appropriate analytical methods for CEAs of CRTs with the characteristics described. SUR and GEE are not recommended for studies with few clusters.

  9. Theory-based behavioral intervention increases self-reported physical activity in South African men: a cluster-randomized controlled trial.

    PubMed

    Jemmott, John B; Jemmott, Loretta S; Ngwane, Zolani; Zhang, Jingwen; Heeren, G Anita; Icard, Larry D; O'Leary, Ann; Mtose, Xoliswa; Teitelman, Anne; Carty, Craig

    2014-07-01

    To determine whether a health-promotion intervention increases South African men's adherence to physical-activity guidelines. We utilized a cluster-randomized controlled trial design. Eligible clusters, residential neighborhoods near East London, South Africa, were matched in pairs. Within randomly selected pairs, neighborhoods were randomized to theory-based, culturally congruent health-promotion intervention encouraging physical activity or attention-matched HIV/STI risk-reduction control intervention. Men residing in the neighborhoods and reporting coitus in the previous 3 months were eligible. Primary outcome was self-reported individual-level adherence to physical-activity guidelines averaged over 6-month and 12-month post-intervention assessments. Data were collected in 2007-2010. Data collectors, but not facilitators or participants, were blind to group assignment. Primary outcome intention-to-treat analysis included 22 of 22 clusters and 537 of 572 men in the health-promotion intervention and 22 of 22 clusters and 569 of 609 men in the attention-control intervention. Model-estimated probability of meeting physical-activity guidelines was 51.0% in the health-promotion intervention and 44.7% in attention-matched control (OR=1.34; 95% CI, 1.09-1.63), adjusting for baseline prevalence and clustering from 44 neighborhoods. A theory-based culturally congruent intervention increased South African men's self-reported physical activity, a key contributor to deaths from non-communicable diseases in South Africa. ClinicalTrials.gov Identifier: NCT01490359. Copyright © 2014 Elsevier Inc. All rights reserved.

  10. Corrected Mean-Field Model for Random Sequential Adsorption on Random Geometric Graphs

    NASA Astrophysics Data System (ADS)

    Dhara, Souvik; van Leeuwaarden, Johan S. H.; Mukherjee, Debankur

    2018-03-01

    A notorious problem in mathematics and physics is to create a solvable model for random sequential adsorption of non-overlapping congruent spheres in the d-dimensional Euclidean space with d≥ 2 . Spheres arrive sequentially at uniformly chosen locations in space and are accepted only when there is no overlap with previously deposited spheres. Due to spatial correlations, characterizing the fraction of accepted spheres remains largely intractable. We study this fraction by taking a novel approach that compares random sequential adsorption in Euclidean space to the nearest-neighbor blocking on a sequence of clustered random graphs. This random network model can be thought of as a corrected mean-field model for the interaction graph between the attempted spheres. Using functional limit theorems, we characterize the fraction of accepted spheres and its fluctuations.

  11. Exploring multicollinearity using a random matrix theory approach.

    PubMed

    Feher, Kristen; Whelan, James; Müller, Samuel

    2012-01-01

    Clustering of gene expression data is often done with the latent aim of dimension reduction, by finding groups of genes that have a common response to potentially unknown stimuli. However, what is poorly understood to date is the behaviour of a low dimensional signal embedded in high dimensions. This paper introduces a multicollinear model which is based on random matrix theory results, and shows potential for the characterisation of a gene cluster's correlation matrix. This model projects a one dimensional signal into many dimensions and is based on the spiked covariance model, but rather characterises the behaviour of the corresponding correlation matrix. The eigenspectrum of the correlation matrix is empirically examined by simulation, under the addition of noise to the original signal. The simulation results are then used to propose a dimension estimation procedure of clusters from data. Moreover, the simulation results warn against considering pairwise correlations in isolation, as the model provides a mechanism whereby a pair of genes with `low' correlation may simply be due to the interaction of high dimension and noise. Instead, collective information about all the variables is given by the eigenspectrum.

  12. Outcomes of a pilot hand hygiene randomized cluster trial to reduce communicable infections among US office-based employees.

    PubMed

    Stedman-Smith, Maggie; DuBois, Cathy L Z; Grey, Scott F; Kingsbury, Diana M; Shakya, Sunita; Scofield, Jennifer; Slenkovich, Ken

    2015-04-01

    To determine the effectiveness of an office-based multimodal hand hygiene improvement intervention in reducing self-reported communicable infections and work-related absence. A randomized cluster trial including an electronic training video, hand sanitizer, and educational posters (n = 131, intervention; n = 193, control). Primary outcomes include (1) self-reported acute respiratory infections (ARIs)/influenza-like illness (ILI) and/or gastrointestinal (GI) infections during the prior 30 days; and (2) related lost work days. Incidence rate ratios calculated using generalized linear mixed models with a Poisson distribution, adjusted for confounders and random cluster effects. A 31% relative reduction in self-reported combined ARI-ILI/GI infections (incidence rate ratio: 0.69; 95% confidence interval, 0.49 to 0.98). A 21% nonsignificant relative reduction in lost work days. An office-based multimodal hand hygiene improvement intervention demonstrated a substantive reduction in self-reported combined ARI-ILI/GI infections.

  13. The relationship of dynamical heterogeneity to the Adam-Gibbs and random first-order transition theories of glass formation.

    PubMed

    Starr, Francis W; Douglas, Jack F; Sastry, Srikanth

    2013-03-28

    We carefully examine common measures of dynamical heterogeneity for a model polymer melt and test how these scales compare with those hypothesized by the Adam and Gibbs (AG) and random first-order transition (RFOT) theories of relaxation in glass-forming liquids. To this end, we first analyze clusters of highly mobile particles, the string-like collective motion of these mobile particles, and clusters of relative low mobility. We show that the time scale of the high-mobility clusters and strings is associated with a diffusive time scale, while the low-mobility particles' time scale relates to a structural relaxation time. The difference of the characteristic times for the high- and low-mobility particles naturally explains the well-known decoupling of diffusion and structural relaxation time scales. Despite the inherent difference of dynamics between high- and low-mobility particles, we find a high degree of similarity in the geometrical structure of these particle clusters. In particular, we show that the fractal dimensions of these clusters are consistent with those of swollen branched polymers or branched polymers with screened excluded-volume interactions, corresponding to lattice animals and percolation clusters, respectively. In contrast, the fractal dimension of the strings crosses over from that of self-avoiding walks for small strings, to simple random walks for longer, more strongly interacting, strings, corresponding to flexible polymers with screened excluded-volume interactions. We examine the appropriateness of identifying the size scales of either mobile particle clusters or strings with the size of cooperatively rearranging regions (CRR) in the AG and RFOT theories. We find that the string size appears to be the most consistent measure of CRR for both the AG and RFOT models. Identifying strings or clusters with the "mosaic" length of the RFOT model relaxes the conventional assumption that the "entropic droplets" are compact. We also confirm the validity of the entropy formulation of the AG theory, constraining the exponent values of the RFOT theory. This constraint, together with the analysis of size scales, enables us to estimate the characteristic exponents of RFOT.

  14. Developing Appropriate Methods for Cost-Effectiveness Analysis of Cluster Randomized Trials

    PubMed Central

    Gomes, Manuel; Ng, Edmond S.-W.; Nixon, Richard; Carpenter, James; Thompson, Simon G.

    2012-01-01

    Aim. Cost-effectiveness analyses (CEAs) may use data from cluster randomized trials (CRTs), where the unit of randomization is the cluster, not the individual. However, most studies use analytical methods that ignore clustering. This article compares alternative statistical methods for accommodating clustering in CEAs of CRTs. Methods. Our simulation study compared the performance of statistical methods for CEAs of CRTs with 2 treatment arms. The study considered a method that ignored clustering—seemingly unrelated regression (SUR) without a robust standard error (SE)—and 4 methods that recognized clustering—SUR and generalized estimating equations (GEEs), both with robust SE, a “2-stage” nonparametric bootstrap (TSB) with shrinkage correction, and a multilevel model (MLM). The base case assumed CRTs with moderate numbers of balanced clusters (20 per arm) and normally distributed costs. Other scenarios included CRTs with few clusters, imbalanced cluster sizes, and skewed costs. Performance was reported as bias, root mean squared error (rMSE), and confidence interval (CI) coverage for estimating incremental net benefits (INBs). We also compared the methods in a case study. Results. Each method reported low levels of bias. Without the robust SE, SUR gave poor CI coverage (base case: 0.89 v. nominal level: 0.95). The MLM and TSB performed well in each scenario (CI coverage, 0.92–0.95). With few clusters, the GEE and SUR (with robust SE) had coverage below 0.90. In the case study, the mean INBs were similar across all methods, but ignoring clustering underestimated statistical uncertainty and the value of further research. Conclusions. MLMs and the TSB are appropriate analytical methods for CEAs of CRTs with the characteristics described. SUR and GEE are not recommended for studies with few clusters. PMID:22016450

  15. Radiation breakage of DNA: a model based on random-walk chromatin structure

    NASA Technical Reports Server (NTRS)

    Ponomarev, A. L.; Sachs, R. K.

    2001-01-01

    Monte Carlo computer software, called DNAbreak, has recently been developed to analyze observed non-random clustering of DNA double strand breaks in chromatin after exposure to densely ionizing radiation. The software models coarse-grained configurations of chromatin and radiation tracks, small-scale details being suppressed in order to obtain statistical results for larger scales, up to the size of a whole chromosome. We here give an analytic counterpart of the numerical model, useful for benchmarks, for elucidating the numerical results, for analyzing the assumptions of a more general but less mechanistic "randomly-located-clusters" formalism, and, potentially, for speeding up the calculations. The equations characterize multi-track DNA fragment-size distributions in terms of one-track action; an important step in extrapolating high-dose laboratory results to the much lower doses of main interest in environmental or occupational risk estimation. The approach can utilize the experimental information on DNA fragment-size distributions to draw inferences about large-scale chromatin geometry during cell-cycle interphase.

  16. Failure tolerance of spike phase synchronization in coupled neural networks

    NASA Astrophysics Data System (ADS)

    Jalili, Mahdi

    2011-09-01

    Neuronal synchronization plays an important role in the various functionality of nervous system such as binding, cognition, information processing, and computation. In this paper, we investigated how random and intentional failures in the nodes of a network influence its phase synchronization properties. We considered both artificially constructed networks using models such as preferential attachment, Watts-Strogatz, and Erdős-Rényi as well as a number of real neuronal networks. The failure strategy was either random or intentional based on properties of the nodes such as degree, clustering coefficient, betweenness centrality, and vulnerability. Hindmarsh-Rose model was considered as the mathematical model for the individual neurons, and the phase synchronization of the spike trains was monitored as a function of the percentage/number of removed nodes. The numerical simulations were supplemented by considering coupled non-identical Kuramoto oscillators. Failures based on the clustering coefficient, i.e., removing the nodes with high values of the clustering coefficient, had the least effect on the spike synchrony in all of the networks. This was followed by errors where the nodes were removed randomly. However, the behavior of the other three attack strategies was not uniform across the networks, and different strategies were the most influential in different network structure.

  17. Running and rotating: modelling the dynamics of migrating cell clusters

    NASA Astrophysics Data System (ADS)

    Copenhagen, Katherine; Gov, Nir; Gopinathan, Ajay

    Collective motion of cells is a common occurrence in many biological systems, including tissue development and repair, and tumor formation. Recent experiments have shown cells form clusters in a chemical gradient, which display three different phases of motion: translational, rotational, and random. We present a model for cell clusters based loosely on other models seen in the literature that involves a Vicsek-like alignment as well as physical collisions and adhesions between cells. With this model we show that a mechanism for driving rotational motion in this kind of system is an increased motility of rim cells. Further, we examine the details of the relationship between rim and core cells, and find that the phases of the cluster as a whole are correlated with the creation and annihilation of topological defects in the tangential component of the velocity field.

  18. MIXOR: a computer program for mixed-effects ordinal regression analysis.

    PubMed

    Hedeker, D; Gibbons, R D

    1996-03-01

    MIXOR provides maximum marginal likelihood estimates for mixed-effects ordinal probit, logistic, and complementary log-log regression models. These models can be used for analysis of dichotomous and ordinal outcomes from either a clustered or longitudinal design. For clustered data, the mixed-effects model assumes that data within clusters are dependent. The degree of dependency is jointly estimated with the usual model parameters, thus adjusting for dependence resulting from clustering of the data. Similarly, for longitudinal data, the mixed-effects approach can allow for individual-varying intercepts and slopes across time, and can estimate the degree to which these time-related effects vary in the population of individuals. MIXOR uses marginal maximum likelihood estimation, utilizing a Fisher-scoring solution. For the scoring solution, the Cholesky factor of the random-effects variance-covariance matrix is estimated, along with the effects of model covariates. Examples illustrating usage and features of MIXOR are provided.

  19. Task shifting of frontline community health workers for cardiovascular risk reduction: design and rationale of a cluster randomised controlled trial (DISHA study) in India.

    PubMed

    Jeemon, Panniyammakal; Narayanan, Gitanjali; Kondal, Dimple; Kahol, Kashvi; Bharadwaj, Ashok; Purty, Anil; Negi, Prakash; Ladhani, Sulaiman; Sanghvi, Jyoti; Singh, Kuldeep; Kapoor, Deksha; Sobti, Nidhi; Lall, Dorothy; Manimunda, Sathyaprakash; Dwivedi, Supriya; Toteja, Gurudyal; Prabhakaran, Dorairaj

    2016-03-15

    Effective task-shifting interventions targeted at reducing the global cardiovascular disease (CVD) epidemic in low and middle-income countries (LMICs) are urgently needed. DISHA is a cluster randomised controlled trial conducted across 10 sites (5 in phase 1 and 5 in phase 2) in India in 120 clusters. At each site, 12 clusters were randomly selected from a district. A cluster is defined as a small village with 250-300 households and well defined geographical boundaries. They were then randomly allocated to intervention and control clusters in a 1:1 allocation sequence. If any of the intervention and control clusters were <10 km apart, one was dropped and replaced with another randomly selected cluster from the same district. The study included a representative baseline cross-sectional survey, development of a structured intervention model, delivery of intervention for a minimum period of 18 months by trained frontline health workers (mainly Anganwadi workers and ASHA workers) and a post intervention survey in a representative sample. The study staff had no information on intervention allocation until the completion of the baseline survey. In order to ensure comparability of data across sites, the DISHA study follows a common protocol and manual of operation with standardized measurement techniques. Our study is the largest community based cluster randomised trial in low and middle-income country settings designed to test the effectiveness of 'task shifting' interventions involving frontline health workers for cardiovascular risk reduction. CTRI/2013/10/004049 . Registered 7 October 2013.

  20. On the Coupling Time of the Heat-Bath Process for the Fortuin-Kasteleyn Random-Cluster Model

    NASA Astrophysics Data System (ADS)

    Collevecchio, Andrea; Elçi, Eren Metin; Garoni, Timothy M.; Weigel, Martin

    2018-01-01

    We consider the coupling from the past implementation of the random-cluster heat-bath process, and study its random running time, or coupling time. We focus on hypercubic lattices embedded on tori, in dimensions one to three, with cluster fugacity at least one. We make a number of conjectures regarding the asymptotic behaviour of the coupling time, motivated by rigorous results in one dimension and Monte Carlo simulations in dimensions two and three. Amongst our findings, we observe that, for generic parameter values, the distribution of the appropriately standardized coupling time converges to a Gumbel distribution, and that the standard deviation of the coupling time is asymptotic to an explicit universal constant multiple of the relaxation time. Perhaps surprisingly, we observe these results to hold both off criticality, where the coupling time closely mimics the coupon collector's problem, and also at the critical point, provided the cluster fugacity is below the value at which the transition becomes discontinuous. Finally, we consider analogous questions for the single-spin Ising heat-bath process.

  1. Clustering promotes switching dynamics in networks of noisy neurons

    NASA Astrophysics Data System (ADS)

    Franović, Igor; Klinshov, Vladimir

    2018-02-01

    Macroscopic variability is an emergent property of neural networks, typically manifested in spontaneous switching between the episodes of elevated neuronal activity and the quiescent episodes. We investigate the conditions that facilitate switching dynamics, focusing on the interplay between the different sources of noise and heterogeneity of the network topology. We consider clustered networks of rate-based neurons subjected to external and intrinsic noise and derive an effective model where the network dynamics is described by a set of coupled second-order stochastic mean-field systems representing each of the clusters. The model provides an insight into the different contributions to effective macroscopic noise and qualitatively indicates the parameter domains where switching dynamics may occur. By analyzing the mean-field model in the thermodynamic limit, we demonstrate that clustering promotes multistability, which gives rise to switching dynamics in a considerably wider parameter region compared to the case of a non-clustered network with sparse random connection topology.

  2. Large-area imaging reveals biologically driven non-random spatial patterns of corals at a remote reef

    NASA Astrophysics Data System (ADS)

    Edwards, Clinton B.; Eynaud, Yoan; Williams, Gareth J.; Pedersen, Nicole E.; Zgliczynski, Brian J.; Gleason, Arthur C. R.; Smith, Jennifer E.; Sandin, Stuart A.

    2017-12-01

    For sessile organisms such as reef-building corals, differences in the degree of dispersion of individuals across a landscape may result from important differences in life-history strategies or may reflect patterns of habitat availability. Descriptions of spatial patterns can thus be useful not only for the identification of key biological and physical mechanisms structuring an ecosystem, but also by providing the data necessary to generate and test ecological theory. Here, we used an in situ imaging technique to create large-area photomosaics of 16 plots at Palmyra Atoll, central Pacific, each covering 100 m2 of benthic habitat. We mapped the location of 44,008 coral colonies and identified each to the lowest taxonomic level possible. Using metrics of spatial dispersion, we tested for departures from spatial randomness. We also used targeted model fitting to explore candidate processes leading to differences in spatial patterns among taxa. Most taxa were clustered and the degree of clustering varied by taxon. A small number of taxa did not significantly depart from randomness and none revealed evidence of spatial uniformity. Importantly, taxa that readily fragment or tolerate stress through partial mortality were more clustered. With little exception, clustering patterns were consistent with models of fragmentation and dispersal limitation. In some taxa, dispersion was linearly related to abundance, suggesting density dependence of spatial patterning. The spatial patterns of stony corals are non-random and reflect fundamental life-history characteristics of the taxa, suggesting that the reef landscape may, in many cases, have important elements of spatial predictability.

  3. Sample size estimation for alternating logistic regressions analysis of multilevel randomized community trials of under-age drinking.

    PubMed

    Reboussin, Beth A; Preisser, John S; Song, Eun-Young; Wolfson, Mark

    2012-07-01

    Under-age drinking is an enormous public health issue in the USA. Evidence that community level structures may impact on under-age drinking has led to a proliferation of efforts to change the environment surrounding the use of alcohol. Although the focus of these efforts is to reduce drinking by individual youths, environmental interventions are typically implemented at the community level with entire communities randomized to the same intervention condition. A distinct feature of these trials is the tendency of the behaviours of individuals residing in the same community to be more alike than that of others residing in different communities, which is herein called 'clustering'. Statistical analyses and sample size calculations must account for this clustering to avoid type I errors and to ensure an appropriately powered trial. Clustering itself may also be of scientific interest. We consider the alternating logistic regressions procedure within the population-averaged modelling framework to estimate the effect of a law enforcement intervention on the prevalence of under-age drinking behaviours while modelling the clustering at multiple levels, e.g. within communities and within neighbourhoods nested within communities, by using pairwise odds ratios. We then derive sample size formulae for estimating intervention effects when planning a post-test-only or repeated cross-sectional community-randomized trial using the alternating logistic regressions procedure.

  4. SU-G-TeP3-14: Three-Dimensional Cluster Model in Inhomogeneous Dose Distribution

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wei, J; Penagaricano, J; Narayanasamy, G

    2016-06-15

    Purpose: We aim to investigate 3D cluster formation in inhomogeneous dose distribution to search for new models predicting radiation tissue damage and further leading to new optimization paradigm for radiotherapy planning. Methods: The aggregation of higher dose in the organ at risk (OAR) than a preset threshold was chosen as the cluster whose connectivity dictates the cluster structure. Upon the selection of the dose threshold, the fractional density defined as the fraction of voxels in the organ eligible to be part of the cluster was determined according to the dose volume histogram (DVH). A Monte Carlo method was implemented tomore » establish a case pertinent to the corresponding DVH. Ones and zeros were randomly assigned to each OAR voxel with the sampling probability equal to the fractional density. Ten thousand samples were randomly generated to ensure a sufficient number of cluster sets. A recursive cluster searching algorithm was developed to analyze the cluster with various connectivity choices like 1-, 2-, and 3-connectivity. The mean size of the largest cluster (MSLC) from the Monte Carlo samples was taken to be a function of the fractional density. Various OARs from clinical plans were included in the study. Results: Intensive Monte Carlo study demonstrates the inverse relationship between the MSLC and the cluster connectivity as anticipated and the cluster size does not change with fractional density linearly regardless of the connectivity types. An initially-slow-increase to exponential growth transition of the MSLC from low to high density was observed. The cluster sizes were found to vary within a large range and are relatively independent of the OARs. Conclusion: The Monte Carlo study revealed that the cluster size could serve as a suitable index of the tissue damage (percolation cluster) and the clinical outcome of the same DVH might be potentially different.« less

  5. Clustering of longitudinal data by using an extended baseline: A new method for treatment efficacy clustering in longitudinal data.

    PubMed

    Schramm, Catherine; Vial, Céline; Bachoud-Lévi, Anne-Catherine; Katsahian, Sandrine

    2018-01-01

    Heterogeneity in treatment efficacy is a major concern in clinical trials. Clustering may help to identify the treatment responders and the non-responders. In the context of longitudinal cluster analyses, sample size and variability of the times of measurements are the main issues with the current methods. Here, we propose a new two-step method for the Clustering of Longitudinal data by using an Extended Baseline. The first step relies on a piecewise linear mixed model for repeated measurements with a treatment-time interaction. The second step clusters the random predictions and considers several parametric (model-based) and non-parametric (partitioning, ascendant hierarchical clustering) algorithms. A simulation study compares all options of the clustering of longitudinal data by using an extended baseline method with the latent-class mixed model. The clustering of longitudinal data by using an extended baseline method with the two model-based algorithms was the more robust model. The clustering of longitudinal data by using an extended baseline method with all the non-parametric algorithms failed when there were unequal variances of treatment effect between clusters or when the subgroups had unbalanced sample sizes. The latent-class mixed model failed when the between-patients slope variability is high. Two real data sets on neurodegenerative disease and on obesity illustrate the clustering of longitudinal data by using an extended baseline method and show how clustering may help to identify the marker(s) of the treatment response. The application of the clustering of longitudinal data by using an extended baseline method in exploratory analysis as the first stage before setting up stratified designs can provide a better estimation of treatment effect in future clinical trials.

  6. An empirical comparison of methods for analyzing correlated data from a discrete choice survey to elicit patient preference for colorectal cancer screening

    PubMed Central

    2012-01-01

    Background A discrete choice experiment (DCE) is a preference survey which asks participants to make a choice among product portfolios comparing the key product characteristics by performing several choice tasks. Analyzing DCE data needs to account for within-participant correlation because choices from the same participant are likely to be similar. In this study, we empirically compared some commonly-used statistical methods for analyzing DCE data while accounting for within-participant correlation based on a survey of patient preference for colorectal cancer (CRC) screening tests conducted in Hamilton, Ontario, Canada in 2002. Methods A two-stage DCE design was used to investigate the impact of six attributes on participants' preferences for CRC screening test and willingness to undertake the test. We compared six models for clustered binary outcomes (logistic and probit regressions using cluster-robust standard error (SE), random-effects and generalized estimating equation approaches) and three models for clustered nominal outcomes (multinomial logistic and probit regressions with cluster-robust SE and random-effects multinomial logistic model). We also fitted a bivariate probit model with cluster-robust SE treating the choices from two stages as two correlated binary outcomes. The rank of relative importance between attributes and the estimates of β coefficient within attributes were used to assess the model robustness. Results In total 468 participants with each completing 10 choices were analyzed. Similar results were reported for the rank of relative importance and β coefficients across models for stage-one data on evaluating participants' preferences for the test. The six attributes ranked from high to low as follows: cost, specificity, process, sensitivity, preparation and pain. However, the results differed across models for stage-two data on evaluating participants' willingness to undertake the tests. Little within-patient correlation (ICC ≈ 0) was found in stage-one data, but substantial within-patient correlation existed (ICC = 0.659) in stage-two data. Conclusions When small clustering effect presented in DCE data, results remained robust across statistical models. However, results varied when larger clustering effect presented. Therefore, it is important to assess the robustness of the estimates via sensitivity analysis using different models for analyzing clustered data from DCE studies. PMID:22348526

  7. Properties of a new small-world network with spatially biased random shortcuts

    NASA Astrophysics Data System (ADS)

    Matsuzawa, Ryo; Tanimoto, Jun; Fukuda, Eriko

    2017-11-01

    This paper introduces a small-world (SW) network with a power-law distance distribution that differs from conventional models in that it uses completely random shortcuts. By incorporating spatial constraints, we analyze the divergence of the proposed model from conventional models in terms of fundamental network properties such as clustering coefficient, average path length, and degree distribution. We find that when the spatial constraint more strongly prohibits a long shortcut, the clustering coefficient is improved and the average path length increases. We also analyze the spatial prisoner's dilemma (SPD) games played on our new SW network in order to understand its dynamical characteristics. Depending on the basis graph, i.e., whether it is a one-dimensional ring or a two-dimensional lattice, and the parameter controlling the prohibition of long-distance shortcuts, the emergent results can vastly differ.

  8. Integrating data from randomized controlled trials and observational studies to predict the response to pregabalin in patients with painful diabetic peripheral neuropathy.

    PubMed

    Alexander, Joe; Edwards, Roger A; Savoldelli, Alberto; Manca, Luigi; Grugni, Roberto; Emir, Birol; Whalen, Ed; Watt, Stephen; Brodsky, Marina; Parsons, Bruce

    2017-07-20

    More patient-specific medical care is expected as more is learned about variations in patient responses to medical treatments. Analytical tools enable insights by linking treatment responses from different types of studies, such as randomized controlled trials (RCTs) and observational studies. Given the importance of evidence from both types of studies, our goal was to integrate these types of data into a single predictive platform to help predict response to pregabalin in individual patients with painful diabetic peripheral neuropathy (pDPN). We utilized three pivotal RCTs of pregabalin (398 North American patients) and the largest observational study of pregabalin (3159 German patients). We implemented a hierarchical cluster analysis to identify patient clusters in the Observational Study to which RCT patients could be matched using the coarsened exact matching (CEM) technique, thereby creating a matched dataset. We then developed autoregressive moving average models (ARMAXs) to estimate weekly pain scores for pregabalin-treated patients in each cluster in the matched dataset using the maximum likelihood method. Finally, we validated ARMAX models using Observational Study patients who had not matched with RCT patients, using t tests between observed and predicted pain scores. Cluster analysis yielded six clusters (287-777 patients each) with the following clustering variables: gender, age, pDPN duration, body mass index, depression history, pregabalin monotherapy, prior gabapentin use, baseline pain score, and baseline sleep interference. CEM yielded 1528 unique patients in the matched dataset. The reduction in global imbalance scores for the clusters after adding the RCT patients (ranging from 6 to 63% depending on the cluster) demonstrated that the process reduced the bias of covariates in five of the six clusters. ARMAX models of pain score performed well (R 2 : 0.85-0.91; root mean square errors: 0.53-0.57). t tests did not show differences between observed and predicted pain scores in the 1955 patients who had not matched with RCT patients. The combination of cluster analyses, CEM, and ARMAX modeling enabled strong predictive capabilities with respect to pain scores. Integrating RCT and Observational Study data using CEM enabled effective use of Observational Study data to predict patient responses.

  9. CLUMP-3D: Testing ΛCDM with Galaxy Cluster Shapes

    NASA Astrophysics Data System (ADS)

    Sereno, Mauro; Umetsu, Keiichi; Ettori, Stefano; Sayers, Jack; Chiu, I.-Non; Meneghetti, Massimo; Vega-Ferrero, Jesús; Zitrin, Adi

    2018-06-01

    The ΛCDM model of structure formation makes strong predictions on the concentration and shape of dark matter (DM) halos, which are determined by mass accretion processes. Comparison between predicted shapes and observations provides a geometric test of the ΛCDM model. Accurate and precise measurements needs a full three-dimensional (3D) analysis of the cluster mass distribution. We accomplish this with a multi-probe 3D analysis of the X-ray regular Cluster Lensing and Supernova survey with Hubble (CLASH) clusters combining strong and weak lensing, X-ray photometry and spectroscopy, and the Sunyaev–Zel’dovich effect (SZe). The cluster shapes and concentrations are consistent with ΛCDM predictions. The CLASH clusters are randomly oriented, as expected given the sample selection criteria. Shapes agree with numerical results for DM-only halos, which hints at baryonic physics being less effective in making halos rounder.

  10. Detecting Intervention Effects in a Cluster-Randomized Design Using Multilevel Structural Equation Modeling for Binary Responses

    PubMed Central

    Cho, Sun-Joo; Preacher, Kristopher J.; Bottge, Brian A.

    2015-01-01

    Multilevel modeling (MLM) is frequently used to detect group differences, such as an intervention effect in a pre-test–post-test cluster-randomized design. Group differences on the post-test scores are detected by controlling for pre-test scores as a proxy variable for unobserved factors that predict future attributes. The pre-test and post-test scores that are most often used in MLM are summed item responses (or total scores). In prior research, there have been concerns regarding measurement error in the use of total scores in using MLM. To correct for measurement error in the covariate and outcome, a theoretical justification for the use of multilevel structural equation modeling (MSEM) has been established. However, MSEM for binary responses has not been widely applied to detect intervention effects (group differences) in intervention studies. In this article, the use of MSEM for intervention studies is demonstrated and the performance of MSEM is evaluated via a simulation study. Furthermore, the consequences of using MLM instead of MSEM are shown in detecting group differences. Results of the simulation study showed that MSEM performed adequately as the number of clusters, cluster size, and intraclass correlation increased and outperformed MLM for the detection of group differences. PMID:29881032

  11. Modelling conflicts with cluster dynamics in networks

    NASA Astrophysics Data System (ADS)

    Tadić, Bosiljka; Rodgers, G. J.

    2010-12-01

    We introduce cluster dynamical models of conflicts in which only the largest cluster can be involved in an action. This mimics the situations in which an attack is planned by a central body, and the largest attack force is used. We study the model in its annealed random graph version, on a fixed network, and on a network evolving through the actions. The sizes of actions are distributed with a power-law tail, however, the exponent is non-universal and depends on the frequency of actions and sparseness of the available connections between units. Allowing the network reconstruction over time in a self-organized manner, e.g., by adding the links based on previous liaisons between units, we find that the power-law exponent depends on the evolution time of the network. Its lower limit is given by the universal value 5/2, derived analytically for the case of random fragmentation processes. In the temporal patterns behind the size of actions we find long-range correlations in the time series of the number of clusters and the non-trivial distribution of time that a unit waits between two actions. In the case of an evolving network the distribution develops a power-law tail, indicating that through repeated actions, the system develops an internal structure with a hierarchy of units.

  12. Detecting Intervention Effects in a Cluster-Randomized Design Using Multilevel Structural Equation Modeling for Binary Responses.

    PubMed

    Cho, Sun-Joo; Preacher, Kristopher J; Bottge, Brian A

    2015-11-01

    Multilevel modeling (MLM) is frequently used to detect group differences, such as an intervention effect in a pre-test-post-test cluster-randomized design. Group differences on the post-test scores are detected by controlling for pre-test scores as a proxy variable for unobserved factors that predict future attributes. The pre-test and post-test scores that are most often used in MLM are summed item responses (or total scores). In prior research, there have been concerns regarding measurement error in the use of total scores in using MLM. To correct for measurement error in the covariate and outcome, a theoretical justification for the use of multilevel structural equation modeling (MSEM) has been established. However, MSEM for binary responses has not been widely applied to detect intervention effects (group differences) in intervention studies. In this article, the use of MSEM for intervention studies is demonstrated and the performance of MSEM is evaluated via a simulation study. Furthermore, the consequences of using MLM instead of MSEM are shown in detecting group differences. Results of the simulation study showed that MSEM performed adequately as the number of clusters, cluster size, and intraclass correlation increased and outperformed MLM for the detection of group differences.

  13. Evidence for a global seismic-moment release sequence

    USGS Publications Warehouse

    Bufe, C.G.; Perkins, D.M.

    2005-01-01

    Temporal clustering of the larger earthquakes (foreshock-mainshock-aftershock) followed by relative quiescence (stress shadow) are characteristic of seismic cycles along plate boundaries. A global seismic-moment release history, based on a little more than 100 years of instrumental earthquake data in an extended version of the catalog of Pacheco and Sykes (1992), illustrates similar behavior for Earth as a whole. Although the largest earthquakes have occurred in the circum-Pacific region, an analysis of moment release in the hemisphere antipodal to the Pacific plate shows a very similar pattern. Monte Carlo simulations confirm that the global temporal clustering of great shallow earthquakes during 1952-1964 at M ??? 9.0 is highly significant (4% random probability) as is the clustering of the events of M ??? 8.6 (0.2% random probability) during 1950-1965. We have extended the Pacheco and Sykes (1992) catalog from 1989 through 2001 using Harvard moment centroid data. Immediately after the 1950-1965 cluster, significant quiescence at and above M 8.4 begins and continues until 2001 (0.5% random probability). In alternative catalogs derived by correcting for possible random errors in magnitude estimates in the extended Pacheco-Sykes catalog, the clustering of M ??? 9 persists at a significant level. These observations indicate that, for great earthquakes, Earth behaves as a coherent seismotectonic system. A very-large-scale mechanism for global earthquake triggering and/or stress transfer is implied. There are several candidates, but so far only viscoelastic relaxation has been modeled on a global scale.

  14. CHIMERA: Top-down model for hierarchical, overlapping and directed cluster structures in directed and weighted complex networks

    NASA Astrophysics Data System (ADS)

    Franke, R.

    2016-11-01

    In many networks discovered in biology, medicine, neuroscience and other disciplines special properties like a certain degree distribution and hierarchical cluster structure (also called communities) can be observed as general organizing principles. Detecting the cluster structure of an unknown network promises to identify functional subdivisions, hierarchy and interactions on a mesoscale. It is not trivial choosing an appropriate detection algorithm because there are multiple network, cluster and algorithmic properties to be considered. Edges can be weighted and/or directed, clusters overlap or build a hierarchy in several ways. Algorithms differ not only in runtime, memory requirements but also in allowed network and cluster properties. They are based on a specific definition of what a cluster is, too. On the one hand, a comprehensive network creation model is needed to build a large variety of benchmark networks with different reasonable structures to compare algorithms. On the other hand, if a cluster structure is already known, it is desirable to separate effects of this structure from other network properties. This can be done with null model networks that mimic an observed cluster structure to improve statistics on other network features. A third important application is the general study of properties in networks with different cluster structures, possibly evolving over time. Currently there are good benchmark and creation models available. But what is left is a precise sandbox model to build hierarchical, overlapping and directed clusters for undirected or directed, binary or weighted complex random networks on basis of a sophisticated blueprint. This gap shall be closed by the model CHIMERA (Cluster Hierarchy Interconnection Model for Evaluation, Research and Analysis) which will be introduced and described here for the first time.

  15. ODE, RDE and SDE models of cell cycle dynamics and clustering in yeast.

    PubMed

    Boczko, Erik M; Gedeon, Tomas; Stowers, Chris C; Young, Todd R

    2010-07-01

    Biologists have long observed periodic-like oxygen consumption oscillations in yeast populations under certain conditions, and several unsatisfactory explanations for this phenomenon have been proposed. These ‘autonomous oscillations’ have often appeared with periods that are nearly integer divisors of the calculated doubling time of the culture. We hypothesize that these oscillations could be caused by a form of cell cycle synchronization that we call clustering. We develop some novel ordinary differential equation models of the cell cycle. For these models, and for random and stochastic perturbations, we give both rigorous proofs and simulations showing that both positive and negative growth rate feedback within the cell cycle are possible agents that can cause clustering of populations within the cell cycle. It occurs for a variety of models and for a broad selection of parameter values. These results suggest that the clustering phenomenon is robust and is likely to be observed in nature. Since there are necessarily an integer number of clusters, clustering would lead to periodic-like behaviour with periods that are nearly integer divisors of the period of the cell cycle. Related experiments have shown conclusively that cell cycle clustering occurs in some oscillating yeast cultures.

  16. Not all stars form in clusters - measuring the kinematics of OB associations with Gaia

    NASA Astrophysics Data System (ADS)

    Ward, Jacob L.; Kruijssen, J. M. Diederik

    2018-04-01

    It is often stated that star clusters are the fundamental units of star formation and that most (if not all) stars form in dense stellar clusters. In this monolithic formation scenario, low-density OB associations are formed from the expansion of gravitationally bound clusters following gas expulsion due to stellar feedback. N-body simulations of this process show that OB associations formed this way retain signs of expansion and elevated radial anisotropy over tens of Myr. However, recent theoretical and observational studies suggest that star formation is a hierarchical process, following the fractal nature of natal molecular clouds and allowing the formation of large-scale associations in situ. We distinguish between these two scenarios by characterizing the kinematics of OB associations using the Tycho-Gaia Astrometric Solution catalogue. To this end, we quantify four key kinematic diagnostics: the number ratio of stars with positive radial velocities to those with negative radial velocities, the median radial velocity, the median radial velocity normalized by the tangential velocity, and the radial anisotropy parameter. Each quantity presents a useful diagnostic of whether the association was more compact in the past. We compare these diagnostics to models representing random motion and the expanding products of monolithic cluster formation. None of these diagnostics show evidence of expansion, either from a single cluster or multiple clusters, and the observed kinematics are better represented by a random velocity distribution. This result favours the hierarchical star formation model in which a minority of stars forms in bound clusters and large-scale, hierarchically structured associations are formed in situ.

  17. Binomial outcomes in dataset with some clusters of size two: can the dependence of twins be accounted for? A simulation study comparing the reliability of statistical methods based on a dataset of preterm infants.

    PubMed

    Sauzet, Odile; Peacock, Janet L

    2017-07-20

    The analysis of perinatal outcomes often involves datasets with some multiple births. These are datasets mostly formed of independent observations and a limited number of clusters of size two (twins) and maybe of size three or more. This non-independence needs to be accounted for in the statistical analysis. Using simulated data based on a dataset of preterm infants we have previously investigated the performance of several approaches to the analysis of continuous outcomes in the presence of some clusters of size two. Mixed models have been developed for binomial outcomes but very little is known about their reliability when only a limited number of small clusters are present. Using simulated data based on a dataset of preterm infants we investigated the performance of several approaches to the analysis of binomial outcomes in the presence of some clusters of size two. Logistic models, several methods of estimation for the logistic random intercept models and generalised estimating equations were compared. The presence of even a small percentage of twins means that a logistic regression model will underestimate all parameters but a logistic random intercept model fails to estimate the correlation between siblings if the percentage of twins is too small and will provide similar estimates to logistic regression. The method which seems to provide the best balance between estimation of the standard error and the parameter for any percentage of twins is the generalised estimating equations. This study has shown that the number of covariates or the level two variance do not necessarily affect the performance of the various methods used to analyse datasets containing twins but when the percentage of small clusters is too small, mixed models cannot capture the dependence between siblings.

  18. Cluster-size entropy in the Axelrod model of social influence: Small-world networks and mass media

    NASA Astrophysics Data System (ADS)

    Gandica, Y.; Charmell, A.; Villegas-Febres, J.; Bonalde, I.

    2011-10-01

    We study the Axelrod's cultural adaptation model using the concept of cluster-size entropy Sc, which gives information on the variability of the cultural cluster size present in the system. Using networks of different topologies, from regular to random, we find that the critical point of the well-known nonequilibrium monocultural-multicultural (order-disorder) transition of the Axelrod model is given by the maximum of the Sc(q) distributions. The width of the cluster entropy distributions can be used to qualitatively determine whether the transition is first or second order. By scaling the cluster entropy distributions we were able to obtain a relationship between the critical cultural trait qc and the number F of cultural features in two-dimensional regular networks. We also analyze the effect of the mass media (external field) on social systems within the Axelrod model in a square network. We find a partially ordered phase whose largest cultural cluster is not aligned with the external field, in contrast with a recent suggestion that this type of phase cannot be formed in regular networks. We draw a q-B phase diagram for the Axelrod model in regular networks.

  19. Cluster-size entropy in the Axelrod model of social influence: small-world networks and mass media.

    PubMed

    Gandica, Y; Charmell, A; Villegas-Febres, J; Bonalde, I

    2011-10-01

    We study the Axelrod's cultural adaptation model using the concept of cluster-size entropy S(c), which gives information on the variability of the cultural cluster size present in the system. Using networks of different topologies, from regular to random, we find that the critical point of the well-known nonequilibrium monocultural-multicultural (order-disorder) transition of the Axelrod model is given by the maximum of the S(c)(q) distributions. The width of the cluster entropy distributions can be used to qualitatively determine whether the transition is first or second order. By scaling the cluster entropy distributions we were able to obtain a relationship between the critical cultural trait q(c) and the number F of cultural features in two-dimensional regular networks. We also analyze the effect of the mass media (external field) on social systems within the Axelrod model in a square network. We find a partially ordered phase whose largest cultural cluster is not aligned with the external field, in contrast with a recent suggestion that this type of phase cannot be formed in regular networks. We draw a q-B phase diagram for the Axelrod model in regular networks.

  20. Cluster structure in the correlation coefficient matrix can be characterized by abnormal eigenvalues

    NASA Astrophysics Data System (ADS)

    Nie, Chun-Xiao

    2018-02-01

    In a large number of previous studies, the researchers found that some of the eigenvalues of the financial correlation matrix were greater than the predicted values of the random matrix theory (RMT). Here, we call these eigenvalues as abnormal eigenvalues. In order to reveal the hidden meaning of these abnormal eigenvalues, we study the toy model with cluster structure and find that these eigenvalues are related to the cluster structure of the correlation coefficient matrix. In this paper, model-based experiments show that in most cases, the number of abnormal eigenvalues of the correlation matrix is equal to the number of clusters. In addition, empirical studies show that the sum of the abnormal eigenvalues is related to the clarity of the cluster structure and is negatively correlated with the correlation dimension.

  1. Outcomes of a Pilot Hand Hygiene Randomized Cluster Trial to Reduce Communicable Infections Among US Office-Based Employees

    PubMed Central

    DuBois, Cathy L.Z.; Grey, Scott F.; Kingsbury, Diana M.; Shakya, Sunita; Scofield, Jennifer; Slenkovich, Ken

    2015-01-01

    Objective: To determine the effectiveness of an office-based multimodal hand hygiene improvement intervention in reducing self-reported communicable infections and work-related absence. Methods: A randomized cluster trial including an electronic training video, hand sanitizer, and educational posters (n = 131, intervention; n = 193, control). Primary outcomes include (1) self-reported acute respiratory infections (ARIs)/influenza-like illness (ILI) and/or gastrointestinal (GI) infections during the prior 30 days; and (2) related lost work days. Incidence rate ratios calculated using generalized linear mixed models with a Poisson distribution, adjusted for confounders and random cluster effects. Results: A 31% relative reduction in self-reported combined ARI-ILI/GI infections (incidence rate ratio: 0.69; 95% confidence interval, 0.49 to 0.98). A 21% nonsignificant relative reduction in lost work days. Conclusions: An office-based multimodal hand hygiene improvement intervention demonstrated a substantive reduction in self-reported combined ARI-ILI/GI infections. PMID:25719534

  2. Hospital recruitment for a pragmatic cluster-randomized clinical trial: Lessons learned from the COMPASS study.

    PubMed

    Johnson, Anna M; Jones, Sara B; Duncan, Pamela W; Bushnell, Cheryl D; Coleman, Sylvia W; Mettam, Laurie H; Kucharska-Newton, Anna M; Sissine, Mysha E; Rosamond, Wayne D

    2018-01-26

    Pragmatic randomized clinical trials are essential to determine the effectiveness of interventions in "real-world" clinical practice. These trials frequently use a cluster-randomized methodology, with randomization at the site level. Despite policymakers' increased interest in supporting pragmatic randomized clinical trials, no studies to date have reported on the unique recruitment challenges faced by cluster-randomized pragmatic trials. We investigated key challenges and successful strategies for hospital recruitment in the Comprehensive Post-Acute Stroke Services (COMPASS) study. The COMPASS study is designed to compare the effectiveness of the COMPASS model versus usual care in improving functional outcomes, reducing the numbers of hospital readmissions, and reducing caregiver strain for patients discharged home after stroke or transient ischemic attack. This model integrates early supported discharge planning with transitional care management, including nurse-led follow-up phone calls after 2, 30, and 60 days and an in-person clinic visit at 7-14 days involving a functional assessment and neurological examination. We present descriptive statistics of the characteristics of successfully recruited hospitals compared with all eligible hospitals, reasons for non-participation, and effective recruitment strategies. We successfully recruited 41 (43%) of 95 eligible North Carolina hospitals. Leading, non-exclusive reasons for non-participation included: insufficient staff or financial resources (n = 33, 61%), lack of health system support (n = 16, 30%), and lack of support of individual decision-makers (n = 11, 20%). Successful recruitment strategies included: building and nurturing relationships, engaging team members and community partners with a diverse skill mix, identifying gatekeepers, finding mutually beneficial solutions, having a central institutional review board, sharing published pilot data, and integrating contracts and review board administrators. Although we incorporated strategies based on the best available evidence at the outset of the study, hospital recruitment required three times as much time and considerably more staff than anticipated. To reach our goal, we tailored strategies to individuals, hospitals, and health systems. Successful recruitment of a sufficient number and representative mix of hospitals requires considerable preparation, planning, and flexibility. Strategies presented here may assist future trial organizers in implementing cluster-randomized pragmatic trials. Clinicaltrials.gov, NCT02588664 . Registered on 23 October 2015.

  3. Effect Sizes in Cluster-Randomized Designs

    ERIC Educational Resources Information Center

    Hedges, Larry V.

    2007-01-01

    Multisite research designs involving cluster randomization are becoming increasingly important in educational and behavioral research. Researchers would like to compute effect size indexes based on the standardized mean difference to compare the results of cluster-randomized studies (and corresponding quasi-experiments) with other studies and to…

  4. Higher-order clustering in networks

    NASA Astrophysics Data System (ADS)

    Yin, Hao; Benson, Austin R.; Leskovec, Jure

    2018-05-01

    A fundamental property of complex networks is the tendency for edges to cluster. The extent of the clustering is typically quantified by the clustering coefficient, which is the probability that a length-2 path is closed, i.e., induces a triangle in the network. However, higher-order cliques beyond triangles are crucial to understanding complex networks, and the clustering behavior with respect to such higher-order network structures is not well understood. Here we introduce higher-order clustering coefficients that measure the closure probability of higher-order network cliques and provide a more comprehensive view of how the edges of complex networks cluster. Our higher-order clustering coefficients are a natural generalization of the traditional clustering coefficient. We derive several properties about higher-order clustering coefficients and analyze them under common random graph models. Finally, we use higher-order clustering coefficients to gain new insights into the structure of real-world networks from several domains.

  5. Droplet localization in the random XXZ model and its manifestations

    NASA Astrophysics Data System (ADS)

    Elgart, A.; Klein, A.; Stolz, G.

    2018-01-01

    We examine many-body localization properties for the eigenstates that lie in the droplet sector of the random-field spin- \\frac 1 2 XXZ chain. These states satisfy a basic single cluster localization property (SCLP), derived in Elgart et al (2018 J. Funct. Anal. (in press)). This leads to many consequences, including dynamical exponential clustering, non-spreading of information under the time evolution, and a zero velocity Lieb-Robinson bound. Since SCLP is only applicable to the droplet sector, our definitions and proofs do not rely on knowledge of the spectral and dynamical characteristics of the model outside this regime. Rather, to allow for a possible mobility transition, we adapt the notion of restricting the Hamiltonian to an energy window from the single particle setting to the many body context.

  6. A Cluster Randomized Trial of Tailored Breastfeeding Support for Women with Gestational Diabetes.

    PubMed

    Stuebe, Alison M; Bonuck, Karen; Adatorwovor, Reuben; Schwartz, Todd A; Berry, Diane C

    2016-12-01

    Women with gestational diabetes mellitus (GDM) and their infants are at increased risk of developing metabolic disease; however, longer breastfeeding is associated with a reduction in these risks. We tested an intervention to increase breastfeeding duration among women with GDM. We conducted a cluster randomized trial to determine the efficacy of a breastfeeding education and support program for women with GDM. Women were enrolled between 22 and 36 weeks of pregnancy and cluster randomized to an experimental lifestyle intervention or wait-list control group. Breastfeeding duration and intensity were prespecified secondary outcomes of the trial. Duration of exclusive and any breastfeeding was assessed at 6 weeks and at 4, 7, and 10 months postpartum. We quantified differences in breastfeeding rates using Kaplan-Meier estimates, log-rank tests, and Cox regression models. We enrolled 100 women, of whom 52% were African American, 31% non-Hispanic white, 11% Hispanic, 9% American Indian or Alaskan Native, 2% Asian, 2% other, and 4% more than one race. In models accounting for within-cluster correlation and adjusted for study site, breastfeeding intention, and African American race, women allocated to the intervention group were less likely to stop breastfeeding (adjusted hazard ratio [HR] 0.40, 95% confidence interval [CI] 0.21-0.74) or to introduce formula (adjusted HR 0.50, 95% CI 0.34-0.72). Our results suggest that targeted breastfeeding education for women with GDM is feasible and efficacious. http://clinicaltrials.gov/ct2/show/NCT01809431.

  7. Cluster-randomized Studies in Educational Research: Principles and Methodological Aspects.

    PubMed

    Dreyhaupt, Jens; Mayer, Benjamin; Keis, Oliver; Öchsner, Wolfgang; Muche, Rainer

    2017-01-01

    An increasing number of studies are being performed in educational research to evaluate new teaching methods and approaches. These studies could be performed more efficiently and deliver more convincing results if they more strictly applied and complied with recognized standards of scientific studies. Such an approach could substantially increase the quality in particular of prospective, two-arm (intervention) studies that aim to compare two different teaching methods. A key standard in such studies is randomization, which can minimize systematic bias in study findings; such bias may result if the two study arms are not structurally equivalent. If possible, educational research studies should also achieve this standard, although this is not yet generally the case. Some difficulties and concerns exist, particularly regarding organizational and methodological aspects. An important point to consider in educational research studies is that usually individuals cannot be randomized, because of the teaching situation, and instead whole groups have to be randomized (so-called "cluster randomization"). Compared with studies with individual randomization, studies with cluster randomization normally require (significantly) larger sample sizes and more complex methods for calculating sample size. Furthermore, cluster-randomized studies require more complex methods for statistical analysis. The consequence of the above is that a competent expert with respective special knowledge needs to be involved in all phases of cluster-randomized studies. Studies to evaluate new teaching methods need to make greater use of randomization in order to achieve scientifically convincing results. Therefore, in this article we describe the general principles of cluster randomization and how to implement these principles, and we also outline practical aspects of using cluster randomization in prospective, two-arm comparative educational research studies.

  8. Rumor Diffusion in an Interests-Based Dynamic Social Network

    PubMed Central

    Mao, Xinjun; Guessoum, Zahia; Zhou, Huiping

    2013-01-01

    To research rumor diffusion in social friend network, based on interests, a dynamic friend network is proposed, which has the characteristics of clustering and community, and a diffusion model is also proposed. With this friend network and rumor diffusion model, based on the zombie-city model, some simulation experiments to analyze the characteristics of rumor diffusion in social friend networks have been conducted. The results show some interesting observations: (1) positive information may evolve to become a rumor through the diffusion process that people may modify the information by word of mouth; (2) with the same average degree, a random social network has a smaller clustering coefficient and is more beneficial for rumor diffusion than the dynamic friend network; (3) a rumor is spread more widely in a social network with a smaller global clustering coefficient than in a social network with a larger global clustering coefficient; and (4) a network with a smaller clustering coefficient has a larger efficiency. PMID:24453911

  9. Rumor diffusion in an interests-based dynamic social network.

    PubMed

    Tang, Mingsheng; Mao, Xinjun; Guessoum, Zahia; Zhou, Huiping

    2013-01-01

    To research rumor diffusion in social friend network, based on interests, a dynamic friend network is proposed, which has the characteristics of clustering and community, and a diffusion model is also proposed. With this friend network and rumor diffusion model, based on the zombie-city model, some simulation experiments to analyze the characteristics of rumor diffusion in social friend networks have been conducted. The results show some interesting observations: (1) positive information may evolve to become a rumor through the diffusion process that people may modify the information by word of mouth; (2) with the same average degree, a random social network has a smaller clustering coefficient and is more beneficial for rumor diffusion than the dynamic friend network; (3) a rumor is spread more widely in a social network with a smaller global clustering coefficient than in a social network with a larger global clustering coefficient; and (4) a network with a smaller clustering coefficient has a larger efficiency.

  10. Effect of village-wide use of long-lasting insecticidal nets on visceral Leishmaniasis vectors in India and Nepal: a cluster randomized trial.

    PubMed

    Picado, Albert; Das, Murari L; Kumar, Vijay; Kesari, Shreekant; Dinesh, Diwakar S; Roy, Lalita; Rijal, Suman; Das, Pradeep; Rowland, Mark; Sundar, Shyam; Coosemans, Marc; Boelaert, Marleen; Davies, Clive R

    2010-01-26

    Visceral leishmaniasis (VL) control in the Indian subcontinent is currently based on case detection and treatment, and on vector control using indoor residual spraying (IRS). The use of long-lasting insecticidal nets (LN) has been postulated as an alternative or complement to IRS. Here we tested the impact of comprehensive distribution of LN on the density of Phlebotomus argentipes in VL-endemic villages. A cluster-randomized controlled trial with household P. argentipes density as outcome was designed. Twelve clusters from an ongoing LN clinical trial--three intervention and three control clusters in both India and Nepal--were selected on the basis of accessibility and VL incidence. Ten houses per cluster selected on the basis of high pre-intervention P. argentipes density were monitored monthly for 12 months after distribution of LN using CDC light traps (LT) and mouth aspiration methods. Ten cattle sheds per cluster were also monitored by aspiration. A random effect linear regression model showed that the cluster-wide distribution of LNs significantly reduced the P. argentipes density/house by 24.9% (95% CI 1.80%-42.5%) as measured by means of LTs. The ongoing clinical trial, designed to measure the impact of LNs on VL incidence, will confirm whether LNs should be adopted as a control strategy in the regional VL elimination programs. The entomological evidence described here provides some evidence that LNs could be usefully deployed as part of the VL control program. ClinicalTrials.gov CT-2005-015374.

  11. Modeling fractal cities using the correlated percolation model.

    NASA Astrophysics Data System (ADS)

    Makse, Hernán A.; Havlin, Shlomo; Stanley, H. Eugene

    1996-03-01

    Cities grow in a way that might be expected to resemble the growth of two-dimensional aggregates of particles, and this has led to recent attempts to model urban growth using ideas from the statistical physics of clusters. In particular, the model of diffusion limited aggregation (DLA) has been invoked to rationalize the apparently fractal nature of urban morphologies(M. Batty and P. Longley, Fractal Cities) (Academic, San Diego, 1994). The DLA model predicts that there should exist only one large fractal cluster, which is almost perfectly screened from incoming 'development units' (representing, for example, people, capital or resources), so that almost all of the cluster growth takes place at the tips of the cluster's branches. We show that an alternative model(H. A. Makse, S. Havlin, H. E. Stanley, Nature 377), 608 (1995), in which development units are correlated rather than being added to the cluster at random, is better able to reproduce the observed morphology of cities and the area distribution of sub-clusters ('towns') in an urban system, and can also describe urban growth dynamics. Our physical model, which corresponds to the correlated percolation model in the presence of a density gradient, is motivated by the fact that in urban areas development attracts further development. The model offers the possibility of predicting the global properties (such as scaling behavior) of urban morphologies.

  12. Mechanisms contributing to cluster formation in the inferior olivary nucleus in brainstem slices from postnatal mice

    PubMed Central

    Kølvraa, Mathias; Müller, Felix C; Jahnsen, Henrik; Rekling, Jens C

    2014-01-01

    Abstract The inferior olivary nucleus (IO) in in vitro slices from postnatal mice (P5.5–P15.5) spontaneously generates clusters of neurons with synchronous calcium transients, and intracellular recordings from IO neurons suggest that electrical coupling between neighbouring IO neurons may serve as a synchronizing mechanism. Here, we studied the cluster-forming mechanism and find that clusters overlap extensively with an overlap distribution that resembles the distribution for a random overlap model. The average somatodendritic field size of single curly IO neurons was ∼6400 μm2, which is slightly smaller than the average IO cluster size. Eighty-seven neurons with overlapping dendrites were estimated to be contained in the principal olive mean cluster size, and about six non-overlapping curly IO neurons could be contained within the largest clusters. Clusters could also be induced by iontophoresis with glutamate. Induced clusters were inhibited by tetrodotoxin, carbenoxelone and 18β-glycyrrhetinic acid, suggesting that sodium action potentials and electrical coupling are involved in glutamate-induced cluster formation, which could also be induced by activation of N-methyl-d-aspartate and α-amino-3-hydroxy-5-methyl-4-isoxazolepropionic acid receptors. Spikelets and a small transient depolarizing response were observed during glutamate-induced cluster formation. Calcium transients spread with decreasing velocity during cluster formation, and somatic action potentials and cluster formation are accompanied by large dendritic calcium transients. In conclusion, cluster formation depends on gap junctions, sodium action potentials and spontaneous clusters occur randomly throughout the IO. The relative slow signal spread during cluster formation, combined with a strong dendritic influx of calcium, may signify that active dendritic properties contribute to cluster formation. PMID:24042500

  13. On the limiting characteristics of quantum random number generators at various clusterings of photocounts

    NASA Astrophysics Data System (ADS)

    Molotkov, S. N.

    2017-03-01

    Various methods for the clustering of photocounts constituting a sequence of random numbers are considered. It is shown that the clustering of photocounts resulting in the Fermi-Dirac distribution makes it possible to achieve the theoretical limit of the random number generation rate.

  14. Cluster randomization and political philosophy.

    PubMed

    Chwang, Eric

    2012-11-01

    In this paper, I will argue that, while the ethical issues raised by cluster randomization can be challenging, they are not new. My thesis divides neatly into two parts. In the first, easier part I argue that many of the ethical challenges posed by cluster randomized human subjects research are clearly present in other types of human subjects research, and so are not novel. In the second, more difficult part I discuss the thorniest ethical challenge for cluster randomized research--cases where consent is genuinely impractical to obtain. I argue that once again these cases require no new analytic insight; instead, we should look to political philosophy for guidance. In other words, the most serious ethical problem that arises in cluster randomized research also arises in political philosophy. © 2011 Blackwell Publishing Ltd.

  15. Critical behavior of the contact process on small-world networks

    NASA Astrophysics Data System (ADS)

    Ferreira, Ronan S.; Ferreira, Silvio C.

    2013-11-01

    We investigate the role of clustering on the critical behavior of the contact process (CP) on small-world networks using the Watts-Strogatz (WS) network model with an edge rewiring probability p. The critical point is well predicted by a homogeneous cluster-approximation for the limit of vanishing clustering ( p → 1). The critical exponents and dimensionless moment ratios of the CP are in agreement with those predicted by the mean-field theory for any p > 0. This independence on the network clustering shows that the small-world property is a sufficient condition for the mean-field theory to correctly predict the universality of the model. Moreover, we compare the CP dynamics on WS networks with rewiring probability p = 1 and random regular networks and show that the weak heterogeneity of the WS network slightly changes the critical point but does not alter other critical quantities of the model.

  16. Comparison of cluster-based and source-attribution methods for estimating transmission risk using large HIV sequence databases.

    PubMed

    Le Vu, Stéphane; Ratmann, Oliver; Delpech, Valerie; Brown, Alison E; Gill, O Noel; Tostevin, Anna; Fraser, Christophe; Volz, Erik M

    2018-06-01

    Phylogenetic clustering of HIV sequences from a random sample of patients can reveal epidemiological transmission patterns, but interpretation is hampered by limited theoretical support and statistical properties of clustering analysis remain poorly understood. Alternatively, source attribution methods allow fitting of HIV transmission models and thereby quantify aspects of disease transmission. A simulation study was conducted to assess error rates of clustering methods for detecting transmission risk factors. We modeled HIV epidemics among men having sex with men and generated phylogenies comparable to those that can be obtained from HIV surveillance data in the UK. Clustering and source attribution approaches were applied to evaluate their ability to identify patient attributes as transmission risk factors. We find that commonly used methods show a misleading association between cluster size or odds of clustering and covariates that are correlated with time since infection, regardless of their influence on transmission. Clustering methods usually have higher error rates and lower sensitivity than source attribution method for identifying transmission risk factors. But neither methods provide robust estimates of transmission risk ratios. Source attribution method can alleviate drawbacks from phylogenetic clustering but formal population genetic modeling may be required to estimate quantitative transmission risk factors. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.

  17. The effect of clustering on lot quality assurance sampling: a probabilistic model to calculate sample sizes for quality assessments

    PubMed Central

    2013-01-01

    Background Traditional Lot Quality Assurance Sampling (LQAS) designs assume observations are collected using simple random sampling. Alternatively, randomly sampling clusters of observations and then individuals within clusters reduces costs but decreases the precision of the classifications. In this paper, we develop a general framework for designing the cluster(C)-LQAS system and illustrate the method with the design of data quality assessments for the community health worker program in Rwanda. Results To determine sample size and decision rules for C-LQAS, we use the beta-binomial distribution to account for inflated risk of errors introduced by sampling clusters at the first stage. We present general theory and code for sample size calculations. The C-LQAS sample sizes provided in this paper constrain misclassification risks below user-specified limits. Multiple C-LQAS systems meet the specified risk requirements, but numerous considerations, including per-cluster versus per-individual sampling costs, help identify optimal systems for distinct applications. Conclusions We show the utility of C-LQAS for data quality assessments, but the method generalizes to numerous applications. This paper provides the necessary technical detail and supplemental code to support the design of C-LQAS for specific programs. PMID:24160725

  18. The effect of clustering on lot quality assurance sampling: a probabilistic model to calculate sample sizes for quality assessments.

    PubMed

    Hedt-Gauthier, Bethany L; Mitsunaga, Tisha; Hund, Lauren; Olives, Casey; Pagano, Marcello

    2013-10-26

    Traditional Lot Quality Assurance Sampling (LQAS) designs assume observations are collected using simple random sampling. Alternatively, randomly sampling clusters of observations and then individuals within clusters reduces costs but decreases the precision of the classifications. In this paper, we develop a general framework for designing the cluster(C)-LQAS system and illustrate the method with the design of data quality assessments for the community health worker program in Rwanda. To determine sample size and decision rules for C-LQAS, we use the beta-binomial distribution to account for inflated risk of errors introduced by sampling clusters at the first stage. We present general theory and code for sample size calculations.The C-LQAS sample sizes provided in this paper constrain misclassification risks below user-specified limits. Multiple C-LQAS systems meet the specified risk requirements, but numerous considerations, including per-cluster versus per-individual sampling costs, help identify optimal systems for distinct applications. We show the utility of C-LQAS for data quality assessments, but the method generalizes to numerous applications. This paper provides the necessary technical detail and supplemental code to support the design of C-LQAS for specific programs.

  19. E-Rehabilitation - an Internet and mobile phone based tailored intervention to enhance self-management of cardiovascular disease: study protocol for a randomized controlled trial.

    PubMed

    Antypas, Konstantinos; Wangberg, Silje C

    2012-07-09

    Cardiac rehabilitation is very important for the recovery and the secondary prevention of cardiovascular disease, and one of its main strategies is to increase the level of physical activity. Internet and mobile phone based interventions have been successfully used to help people to achieve this. One of the components that are related to the efficacy of these interventions is tailoring of content to the individual. This trial is studying the effect of a longitudinally tailored Internet and mobile phone based intervention that is based on models of health behaviour, on the level of physical activity and the adherence to the intervention, as an extension of a face-to-face cardiac rehabilitation stay. A parallel group, cluster randomized controlled trial. The study population is adult participants of a cardiac rehabilitation programme in Norway with home Internet access and mobile phone, who in monthly clusters are randomized to the control or the intervention condition. Participants have access to a website with information regarding cardiac rehabilitation, an online discussion forum and an online activity calendar. Those randomized to the intervention condition, receive in addition tailored content based on models of health behaviour, through the website and mobile text messages. The objective is to assess the effect of the intervention on maintenance of self-management behaviours after the rehabilitation stay. Main outcome is the level of physical activity one month, three months and one year after the end of the cardiac rehabilitation programme. The randomization of clusters is based on a true random number online service, and participants, investigators and outcome assessor are blinded to the condition of the clusters. The study suggests a theory-based intervention that combines models of health behaviour in an innovative way, in order to tailor the delivered content. The users have been actively involved in its design, and because of the use of Open-Source software, the intervention can easily and at low-cost be reproduced and expanded by others. Challenges are the recruitment in the elderly population and the possible underrepresentation of women in the study sample. Funding by Northern Norway Regional Health Authority. Trial registry http://www.clinicaltrials.gov: NCT01223170.

  20. Manifestations of Dynamical Localization in the Disordered XXZ Spin Chain

    NASA Astrophysics Data System (ADS)

    Elgart, Alexander; Klein, Abel; Stolz, Günter

    2018-04-01

    We study disordered XXZ spin chains in the Ising phase exhibiting droplet localization, a single cluster localization property we previously proved for random XXZ spin chains. It holds in an energy interval I near the bottom of the spectrum, known as the droplet spectrum. We establish dynamical manifestations of localization in the energy window I, including non-spreading of information, zero-velocity Lieb-Robinson bounds, and general dynamical clustering. Our results do not rely on knowledge of the dynamical characteristics of the model outside the droplet spectrum. A byproduct of our analysis is that for random XXZ spin chains this droplet localization can happen only inside the droplet spectrum.

  1. Cluster-randomized Studies in Educational Research: Principles and Methodological Aspects

    PubMed Central

    Dreyhaupt, Jens; Mayer, Benjamin; Keis, Oliver; Öchsner, Wolfgang; Muche, Rainer

    2017-01-01

    An increasing number of studies are being performed in educational research to evaluate new teaching methods and approaches. These studies could be performed more efficiently and deliver more convincing results if they more strictly applied and complied with recognized standards of scientific studies. Such an approach could substantially increase the quality in particular of prospective, two-arm (intervention) studies that aim to compare two different teaching methods. A key standard in such studies is randomization, which can minimize systematic bias in study findings; such bias may result if the two study arms are not structurally equivalent. If possible, educational research studies should also achieve this standard, although this is not yet generally the case. Some difficulties and concerns exist, particularly regarding organizational and methodological aspects. An important point to consider in educational research studies is that usually individuals cannot be randomized, because of the teaching situation, and instead whole groups have to be randomized (so-called “cluster randomization”). Compared with studies with individual randomization, studies with cluster randomization normally require (significantly) larger sample sizes and more complex methods for calculating sample size. Furthermore, cluster-randomized studies require more complex methods for statistical analysis. The consequence of the above is that a competent expert with respective special knowledge needs to be involved in all phases of cluster-randomized studies. Studies to evaluate new teaching methods need to make greater use of randomization in order to achieve scientifically convincing results. Therefore, in this article we describe the general principles of cluster randomization and how to implement these principles, and we also outline practical aspects of using cluster randomization in prospective, two-arm comparative educational research studies. PMID:28584874

  2. The Method of Randomization for Cluster-Randomized Trials: Challenges of Including Patients with Multiple Chronic Conditions

    PubMed Central

    Esserman, Denise; Allore, Heather G.; Travison, Thomas G.

    2016-01-01

    Cluster-randomized clinical trials (CRT) are trials in which the unit of randomization is not a participant but a group (e.g. healthcare systems or community centers). They are suitable when the intervention applies naturally to the cluster (e.g. healthcare policy); when lack of independence among participants may occur (e.g. nursing home hygiene); or when it is most ethical to apply an intervention to all within a group (e.g. school-level immunization). Because participants in the same cluster receive the same intervention, CRT may approximate clinical practice, and may produce generalizable findings. However, when not properly designed or interpreted, CRT may induce biased results. CRT designs have features that add complexity to statistical estimation and inference. Chief among these is the cluster-level correlation in response measurements induced by the randomization. A critical consideration is the experimental unit of inference; often it is desirable to consider intervention effects at the level of the individual rather than the cluster. Finally, given that the number of clusters available may be limited, simple forms of randomization may not achieve balance between intervention and control arms at either the cluster- or participant-level. In non-clustered clinical trials, balance of key factors may be easier to achieve because the sample can be homogenous by exclusion of participants with multiple chronic conditions (MCC). CRTs, which are often pragmatic, may eschew such restrictions. Failure to account for imbalance may induce bias and reducing validity. This article focuses on the complexities of randomization in the design of CRTs, such as the inclusion of patients with MCC, and imbalances in covariate factors across clusters. PMID:27478520

  3. How mutation affects evolutionary games on graphs

    PubMed Central

    Allen, Benjamin; Traulsen, Arne; Tarnita, Corina E.; Nowak, Martin A.

    2011-01-01

    Evolutionary dynamics are affected by population structure, mutation rates and update rules. Spatial or network structure facilitates the clustering of strategies, which represents a mechanism for the evolution of cooperation. Mutation dilutes this effect. Here we analyze how mutation influences evolutionary clustering on graphs. We introduce new mathematical methods to evolutionary game theory, specifically the analysis of coalescing random walks via generating functions. These techniques allow us to derive exact identity-by-descent (IBD) probabilities, which characterize spatial assortment on lattices and Cayley trees. From these IBD probabilities we obtain exact conditions for the evolution of cooperation and other game strategies, showing the dual effects of graph topology and mutation rate. High mutation rates diminish the clustering of cooperators, hindering their evolutionary success. Our model can represent either genetic evolution with mutation, or social imitation processes with random strategy exploration. PMID:21473871

  4. A multifaceted intervention to narrow the evidence-based gap in the treatment of acute coronary syndromes: rationale and design of the Brazilian Intervention to Increase Evidence Usage in Acute Coronary Syndromes (BRIDGE-ACS) cluster-randomized trial.

    PubMed

    Berwanger, Otávio; Guimarães, Hélio P; Laranjeira, Ligia N; Cavalcanti, Alexandre B; Kodama, Alessandra; Zazula, Ana Denise; Santucci, Eliana; Victor, Elivane; Flato, Uri A; Tenuta, Marcos; Carvalho, Vitor; Mira, Vera Lucia; Pieper, Karen S; Mota, Luiz Henrique; Peterson, Eric D; Lopes, Renato D

    2012-03-01

    Translating evidence into clinical practice in the management of acute coronary syndromes (ACS) is challenging. Few ACS quality improvement interventions have been rigorously evaluated to determine their impact on patient care and clinical outcomes. We designed a pragmatic, 2-arm, cluster-randomized trial involving 34 clusters (Brazilian public hospitals). Clusters were randomized to receive a multifaceted quality improvement intervention (experimental group) or routine practice (control group). The 6-month educational intervention included reminders, care algorithms, a case manager, and distribution of educational materials to health care providers. The primary end point was a composite of evidence-based post-ACS therapies within 24 hours of admission, with the secondary measure of major cardiovascular clinical events (death, nonfatal myocardial infarction, nonfatal cardiac arrest, and nonfatal stroke). Prescription of evidence-based therapies at hospital discharge were also evaluated as part of the secondary outcomes. All analyses were performed by the intention-to-treat principle and took the cluster design into account using individual-level regression modeling (generalized estimating equations). If proven effective, this multifaceted intervention would have wide use as a means of promoting optimal use of evidence-based interventions for the management of ACS. Copyright © 2012 Mosby, Inc. All rights reserved.

  5. Percolation and epidemics in random clustered networks

    NASA Astrophysics Data System (ADS)

    Miller, Joel C.

    2009-08-01

    The social networks that infectious diseases spread along are typically clustered. Because of the close relation between percolation and epidemic spread, the behavior of percolation in such networks gives insight into infectious disease dynamics. A number of authors have studied percolation or epidemics in clustered networks, but the networks often contain preferential contacts in high degree nodes. We introduce a class of random clustered networks and a class of random unclustered networks with the same preferential mixing. Percolation in the clustered networks reduces the component sizes and increases the epidemic threshold compared to the unclustered networks.

  6. Using Cluster Bootstrapping to Analyze Nested Data with a Few Clusters

    ERIC Educational Resources Information Center

    Huang, Francis L.

    2018-01-01

    Cluster randomized trials involving participants nested within intact treatment and control groups are commonly performed in various educational, psychological, and biomedical studies. However, recruiting and retaining intact groups present various practical, financial, and logistical challenges to evaluators and often, cluster randomized trials…

  7. Intraclass Correlations for Three-Level Multi-Site Cluster-Randomized Trials of Science Achievement

    ERIC Educational Resources Information Center

    Westine, Carl D.

    2015-01-01

    A cluster-randomized trial (CRT) relies on random assignment of intact clusters to treatment conditions, such as classrooms or schools (Raudenbush & Bryk, 2002). One specific type of CRT, a multi-site CRT (MSCRT), is commonly employed in educational research and evaluation studies (Spybrook & Raudenbush, 2009; Spybrook, 2014; Bloom,…

  8. Sydney Playground Project: A Cluster-Randomized Trial to Increase Physical Activity, Play, and Social Skills

    ERIC Educational Resources Information Center

    Bundy, Anita; Engelen, Lina; Wyver, Shirley; Tranter, Paul; Ragen, Jo; Bauman, Adrian; Baur, Louise; Schiller, Wendy; Simpson, Judy M.; Niehues, Anita N.; Perry, Gabrielle; Jessup, Glenda; Naughton, Geraldine

    2017-01-01

    Background: We assessed the effectiveness of a simple intervention for increasing children's physical activity, play, perceived competence/social acceptance, and social skills. Methods: A cluster-randomized controlled trial was conducted, in which schools were the clusters. Twelve Sydney (Australia) primary schools were randomly allocated to…

  9. Percolation on fitness landscapes: effects of correlation, phenotype, and incompatibilities

    PubMed Central

    Gravner, Janko; Pitman, Damien; Gavrilets, Sergey

    2009-01-01

    We study how correlations in the random fitness assignment may affect the structure of fitness landscapes, in three classes of fitness models. The first is a phenotype space in which individuals are characterized by a large number n of continuously varying traits. In a simple model of random fitness assignment, viable phenotypes are likely to form a giant connected cluster percolating throughout the phenotype space provided the viability probability is larger than 1/2n. The second model explicitly describes genotype-to-phenotype and phenotype-to-fitness maps, allows for neutrality at both phenotype and fitness levels, and results in a fitness landscape with tunable correlation length. Here, phenotypic neutrality and correlation between fitnesses can reduce the percolation threshold, and correlations at the point of phase transition between local and global are most conducive to the formation of the giant cluster. In the third class of models, particular combinations of alleles or values of phenotypic characters are “incompatible” in the sense that the resulting genotypes or phenotypes have zero fitness. This setting can be viewed as a generalization of the canonical Bateson-Dobzhansky-Muller model of speciation and is related to K- SAT problems, prominent in computer science. We analyze the conditions for the existence of viable genotypes, their number, as well as the structure and the number of connected clusters of viable genotypes. We show that analysis based on expected values can easily lead to wrong conclusions, especially when fitness correlations are strong. We focus on pairwise incompatibilities between diallelic loci, but we also address multiple alleles, complex incompatibilities, and continuous phenotype spaces. In the case of diallelic loci, the number of clusters is stochastically bounded and each cluster contains a very large sub-cube. Finally, we demonstrate that the discrete NK model shares some signature properties of models with high correlations. PMID:17692873

  10. A Simple Model for the Earthquake Cycle Combining Self-Organized Criticality with Critical Point Behavior

    NASA Astrophysics Data System (ADS)

    Newman, W. I.; Turcotte, D. L.

    2002-12-01

    We have studied a hybrid model combining the forest-fire model with the site-percolation model in order to better understand the earthquake cycle. We consider a square array of sites. At each time step, a "tree" is dropped on a randomly chosen site and is planted if the site is unoccupied. When a cluster of "trees" spans the site (a percolating cluster), all the trees in the cluster are removed ("burned") in a "fire." The removal of the cluster is analogous to a characteristic earthquake and planting "trees" is analogous to increasing the regional stress. The clusters are analogous to the metastable regions of a fault over which an earthquake rupture can propagate once triggered. We find that the frequency-area statistics of the metastable regions are power-law with a negative exponent of two (as in the forest-fire model). This is analogous to the Gutenberg-Richter distribution of seismicity. This "self-organized critical behavior" can be explained in terms of an inverse cascade of clusters. Individual trees move from small to larger clusters until they are destroyed. This inverse cascade of clusters is self-similar and the power-law distribution of cluster sizes has been shown to have an exponent of two. We have quantified the forecasting of the spanning fires using error diagrams. The assumption that "fires" (earthquakes) are quasi-periodic has moderate predictability. The density of trees gives an improved degree of predictability, while the size of the largest cluster of trees provides a substantial improvement in forecasting a "fire."

  11. Cluster growth mechanisms in Lennard-Jones fluids: A comparison between molecular dynamics and Brownian dynamics simulations

    NASA Astrophysics Data System (ADS)

    Jung, Jiyun; Lee, Jumin; Kim, Jun Soo

    2015-03-01

    We present a simulation study on the mechanisms of a phase separation in dilute fluids of Lennard-Jones (LJ) particles as a model of self-interacting molecules. Molecular dynamics (MD) and Brownian dynamics (BD) simulations of the LJ fluids are employed to model the condensation of a liquid droplet in the vapor phase and the mesoscopic aggregation in the solution phase, respectively. With emphasis on the cluster growth at late times well beyond the nucleation stage, we find that the growth mechanisms can be qualitatively different: cluster diffusion and coalescence in the MD simulations and Ostwald ripening in the BD simulations. We also show that the rates of the cluster growth have distinct scaling behaviors during cluster growth. This work suggests that in the solution phase the random Brownian nature of the solute dynamics may lead to the Ostwald ripening that is qualitatively different from the cluster coalescence in the vapor phase.

  12. Impacts of clustering on interacting epidemics.

    PubMed

    Wang, Bing; Cao, Lang; Suzuki, Hideyuki; Aihara, Kazuyuki

    2012-07-07

    Since community structures in real networks play a major role for the epidemic spread, we therefore explore two interacting diseases spreading in networks with community structures. As a network model with community structures, we propose a random clique network model composed of different orders of cliques. We further assume that each disease spreads only through one type of cliques; this assumption corresponds to the issue that two diseases spread inside communities and outside them. Considering the relationship between the susceptible-infected-recovered (SIR) model and the bond percolation theory, we apply this theory to clique random networks under the assumption that the occupation probability is clique-type dependent, which is consistent with the observation that infection rates inside a community and outside it are different, and obtain a number of statistical properties for this model. Two interacting diseases that compete the same hosts are also investigated, which leads to a natural generalization of analyzing an arbitrary number of infectious diseases. For two-disease dynamics, the clustering effect is hypersensitive to the cohesiveness and concentration of cliques; this illustrates the impacts of clustering and the composition of subgraphs in networks on epidemic behavior. The analysis of coexistence/bistability regions provides significant insight into the relationship between the network structure and the potential epidemic prevalence. Copyright © 2012 Elsevier Ltd. All rights reserved.

  13. Implementation of Structured Inquiry Based Model Learning toward Students' Understanding of Geometry

    ERIC Educational Resources Information Center

    Salim, Kalbin; Tiawa, Dayang Hjh

    2015-01-01

    The purpose of this study is implementation of a structured inquiry learning model in instruction of geometry. The model used is a model with a quasi-experimental study amounted to two classes of samples selected from the population of the ten classes with cluster random sampling technique. Data collection tool consists of a test item…

  14. Spatial-temporal clustering of tornadoes

    NASA Astrophysics Data System (ADS)

    Malamud, Bruce D.; Turcotte, Donald L.; Brooks, Harold E.

    2016-12-01

    The standard measure of the intensity of a tornado is the Enhanced Fujita scale, which is based qualitatively on the damage caused by a tornado. An alternative measure of tornado intensity is the tornado path length, L. Here we examine the spatial-temporal clustering of severe tornadoes, which we define as having path lengths L ≥ 10 km. Of particular concern are tornado outbreaks, when a large number of severe tornadoes occur in a day in a restricted region. We apply a spatial-temporal clustering analysis developed for earthquakes. We take all pairs of severe tornadoes in observed and modelled outbreaks, and for each pair plot the spatial lag (distance between touchdown points) against the temporal lag (time between touchdown points). We apply our spatial-temporal lag methodology to the intense tornado outbreaks in the central United States on 26 and 27 April 2011, which resulted in over 300 fatalities and produced 109 severe (L ≥ 10 km) tornadoes. The patterns of spatial-temporal lag correlations that we obtain for the 2 days are strikingly different. On 26 April 2011, there were 45 severe tornadoes and our clustering analysis is dominated by a complex sequence of linear features. We associate the linear patterns with the tornadoes generated in either a single cell thunderstorm or a closely spaced cluster of single cell thunderstorms moving at a near-constant velocity. Our study of a derecho tornado outbreak of six severe tornadoes on 4 April 2011 along with modelled outbreak scenarios confirms this association. On 27 April 2011, there were 64 severe tornadoes and our clustering analysis is predominantly random with virtually no embedded linear patterns. We associate this pattern with a large number of interacting supercell thunderstorms generating tornadoes randomly in space and time. In order to better understand these associations, we also applied our approach to the Great Plains tornado outbreak of 3 May 1999. Careful studies by others have associated individual tornadoes with specified supercell thunderstorms. Our analysis of the 3 May 1999 tornado outbreak directly associated linear features in the largely random spatial-temporal analysis with several supercell thunderstorms, which we then confirmed using model scenarios of synthetic tornado outbreaks. We suggest that it may be possible to develop a semi-automated modelling of tornado touchdowns to match the type of observations made on the 3 May 1999 outbreak.

  15. Spatial-Temporal Clustering of Tornadoes

    NASA Astrophysics Data System (ADS)

    Malamud, Bruce D.; Turcotte, Donald L.; Brooks, Harold E.

    2017-04-01

    The standard measure of the intensity of a tornado is the Enhanced Fujita scale, which is based qualitatively on the damage caused by a tornado. An alternative measure of tornado intensity is the tornado path length, L. Here we examine the spatial-temporal clustering of severe tornadoes, which we define as having path lengths L ≥ 10 km. Of particular concern are tornado outbreaks, when a large number of severe tornadoes occur in a day in a restricted region. We apply a spatial-temporal clustering analysis developed for earthquakes. We take all pairs of severe tornadoes in observed and modelled outbreaks, and for each pair plot the spatial lag (distance between touchdown points) against the temporal lag (time between touchdown points). We apply our spatial-temporal lag methodology to the intense tornado outbreaks in the central United States on 26 and 27 April 2011, which resulted in over 300 fatalities and produced 109 severe (L ≥ 10 km) tornadoes. The patterns of spatial-temporal lag correlations that we obtain for the 2 days are strikingly different. On 26 April 2011, there were 45 severe tornadoes and our clustering analysis is dominated by a complex sequence of linear features. We associate the linear patterns with the tornadoes generated in either a single cell thunderstorm or a closely spaced cluster of single cell thunderstorms moving at a near-constant velocity. Our study of a derecho tornado outbreak of six severe tornadoes on 4 April 2011 along with modelled outbreak scenarios confirms this association. On 27 April 2011, there were 64 severe tornadoes and our clustering analysis is predominantly random with virtually no embedded linear patterns. We associate this pattern with a large number of interacting supercell thunderstorms generating tornadoes randomly in space and time. In order to better understand these associations, we also applied our approach to the Great Plains tornado outbreak of 3 May 1999. Careful studies by others have associated individual tornadoes with specified supercell thunderstorms. Our analysis of the 3 May 1999 tornado outbreak directly associated linear features in the largely random spatial-temporal analysis with several supercell thunderstorms, which we then confirmed using model scenarios of synthetic tornado outbreaks. We suggest that it may be possible to develop a semi-automated modelling of tornado touchdowns to match the type of observations made on the 3 May 1999 outbreak.

  16. Avoiding Boundary Estimates in Hierarchical Linear Models through Weakly Informative Priors

    ERIC Educational Resources Information Center

    Chung, Yeojin; Rabe-Hesketh, Sophia; Gelman, Andrew; Dorie, Vincent; Liu, Jinchen

    2012-01-01

    Hierarchical or multilevel linear models are widely used for longitudinal or cross-sectional data on students nested in classes and schools, and are particularly important for estimating treatment effects in cluster-randomized trials, multi-site trials, and meta-analyses. The models can allow for variation in treatment effects, as well as…

  17. Spectra of random networks in the weak clustering regime

    NASA Astrophysics Data System (ADS)

    Peron, Thomas K. DM.; Ji, Peng; Kurths, Jürgen; Rodrigues, Francisco A.

    2018-03-01

    The asymptotic behavior of dynamical processes in networks can be expressed as a function of spectral properties of the corresponding adjacency and Laplacian matrices. Although many theoretical results are known for the spectra of traditional configuration models, networks generated through these models fail to describe many topological features of real-world networks, in particular non-null values of the clustering coefficient. Here we study effects of cycles of order three (triangles) in network spectra. By using recent advances in random matrix theory, we determine the spectral distribution of the network adjacency matrix as a function of the average number of triangles attached to each node for networks without modular structure and degree-degree correlations. Implications to network dynamics are discussed. Our findings can shed light in the study of how particular kinds of subgraphs influence network dynamics.

  18. Modeling stock price dynamics by continuum percolation system and relevant complex systems analysis

    NASA Astrophysics Data System (ADS)

    Xiao, Di; Wang, Jun

    2012-10-01

    The continuum percolation system is developed to model a random stock price process in this work. Recent empirical research has demonstrated various statistical features of stock price changes, the financial model aiming at understanding price fluctuations needs to define a mechanism for the formation of the price, in an attempt to reproduce and explain this set of empirical facts. The continuum percolation model is usually referred to as a random coverage process or a Boolean model, the local interaction or influence among traders is constructed by the continuum percolation, and a cluster of continuum percolation is applied to define the cluster of traders sharing the same opinion about the market. We investigate and analyze the statistical behaviors of normalized returns of the price model by some analysis methods, including power-law tail distribution analysis, chaotic behavior analysis and Zipf analysis. Moreover, we consider the daily returns of Shanghai Stock Exchange Composite Index from January 1997 to July 2011, and the comparisons of return behaviors between the actual data and the simulation data are exhibited.

  19. Clustered multistate models with observation level random effects, mover-stayer effects and dynamic covariates: modelling transition intensities and sojourn times in a study of psoriatic arthritis.

    PubMed

    Yiu, Sean; Farewell, Vernon T; Tom, Brian D M

    2018-02-01

    In psoriatic arthritis, it is important to understand the joint activity (represented by swelling and pain) and damage processes because both are related to severe physical disability. The paper aims to provide a comprehensive investigation into both processes occurring over time, in particular their relationship, by specifying a joint multistate model at the individual hand joint level, which also accounts for many of their important features. As there are multiple hand joints, such an analysis will be based on the use of clustered multistate models. Here we consider an observation level random-effects structure with dynamic covariates and allow for the possibility that a subpopulation of patients is at minimal risk of damage. Such an analysis is found to provide further understanding of the activity-damage relationship beyond that provided by previous analyses. Consideration is also given to the modelling of mean sojourn times and jump probabilities. In particular, a novel model parameterization which allows easily interpretable covariate effects to act on these quantities is proposed.

  20. Measures of clustering and heterogeneity in multilevel Poisson regression analyses of rates/count data

    PubMed Central

    Austin, Peter C.; Stryhn, Henrik; Leckie, George; Merlo, Juan

    2017-01-01

    Multilevel data occur frequently in many research areas like health services research and epidemiology. A suitable way to analyze such data is through the use of multilevel regression models. These models incorporate cluster‐specific random effects that allow one to partition the total variation in the outcome into between‐cluster variation and between‐individual variation. The magnitude of the effect of clustering provides a measure of the general contextual effect. When outcomes are binary or time‐to‐event in nature, the general contextual effect can be quantified by measures of heterogeneity like the median odds ratio or the median hazard ratio, respectively, which can be calculated from a multilevel regression model. Outcomes that are integer counts denoting the number of times that an event occurred are common in epidemiological and medical research. The median (incidence) rate ratio in multilevel Poisson regression for counts that corresponds to the median odds ratio or median hazard ratio for binary or time‐to‐event outcomes respectively is relatively unknown and is rarely used. The median rate ratio is the median relative change in the rate of the occurrence of the event when comparing identical subjects from 2 randomly selected different clusters that are ordered by rate. We also describe how the variance partition coefficient, which denotes the proportion of the variation in the outcome that is attributable to between‐cluster differences, can be computed with count outcomes. We illustrate the application and interpretation of these measures in a case study analyzing the rate of hospital readmission in patients discharged from hospital with a diagnosis of heart failure. PMID:29114926

  1. Under What Circumstances Does External Knowledge about the Correlation Structure Improve Power in Cluster Randomized Designs?

    ERIC Educational Resources Information Center

    Rhoads, Christopher

    2014-01-01

    Recent publications have drawn attention to the idea of utilizing prior information about the correlation structure to improve statistical power in cluster randomized experiments. Because power in cluster randomized designs is a function of many different parameters, it has been difficult for applied researchers to discern a simple rule explaining…

  2. Evaluation of Model Specification, Variable Selection, and Adjustment Methods in Relation to Propensity Scores and Prognostic Scores in Multilevel Data

    ERIC Educational Resources Information Center

    Yu, Bing; Hong, Guanglei

    2012-01-01

    This study uses simulation examples representing three types of treatment assignment mechanisms in data generation (the random intercept and slopes setting, the random intercept setting, and a third setting with a cluster-level treatment and an individual-level outcome) in order to determine optimal procedures for reducing bias and improving…

  3. Existence of the Harmonic Measure for Random Walks on Graphs and in Random Environments

    NASA Astrophysics Data System (ADS)

    Boivin, Daniel; Rau, Clément

    2013-01-01

    We give a sufficient condition for the existence of the harmonic measure from infinity of transient random walks on weighted graphs. In particular, this condition is verified by the random conductance model on ℤ d , d≥3, when the conductances are i.i.d. and the bonds with positive conductance percolate. The harmonic measure from infinity also exists for random walks on supercritical clusters of ℤ2. This is proved using results of Barlow (Ann. Probab. 32:3024-3084, 2004) and Barlow and Hambly (Electron. J. Probab. 14(1):1-27, 2009).

  4. Clustering of galaxies near damped Lyman-alpha systems with (z) = 2.6

    NASA Technical Reports Server (NTRS)

    Wolfe, A. M

    1993-01-01

    The galaxy two-point correlation function, xi, at (z) = 2.6 is determined by comparing the number of Ly-alpha-emitting galaxies in narrowband CCD fields selected for the presence of damped L-alpha absorption to their number in randomly selected control fields. Comparisons between the presented determination of (xi), a density-weighted volume average of xi, and model predictions for (xi) at large redshifts show that models in which the clustering pattern is fixed in proper coordinates are highly unlikely, while better agreement is obtained if the clustering pattern is fixed in comoving coordinates. Therefore, clustering of Ly-alpha-emitting galaxies around damped Ly-alpha systems at large redshifts is strong. It is concluded that the faint blue galaxies are drawn from a parent population different from normal galaxies, the presumed offspring of damped Ly-alpha systems.

  5. Tidal disruption of open clusters in their parent molecular clouds

    NASA Technical Reports Server (NTRS)

    Long, Kevin

    1989-01-01

    A simple model of tidal encounters has been applied to the problem of an open cluster in a clumpy molecular cloud. The parameters of the clumps are taken from the Blitz, Stark, and Long (1988) catalog of clumps in the Rosette molecular cloud. Encounters are modeled as impulsive, rectilinear collisions between Plummer spheres, but the tidal approximation is not invoked. Mass and binding energy changes during an encounter are computed by considering the velocity impulses given to individual stars in a random realization of a Plummer sphere. Mean rates of mass and binding energy loss are then computed by integrating over many encounters. Self-similar evolutionary calculations using these rates indicate that the disruption process is most sensitive to the cluster radius and relatively insensitive to cluster mass. The calculations indicate that clusters which are born in a cloud similar to the Rosette with a cluster radius greater than about 2.5 pc will not survive long enough to leave the cloud. The majority of clusters, however, have smaller radii and will survive the passage through their parent cloud.

  6. Hierarchical Velocity Structure in the Core of Abell 2597

    NASA Technical Reports Server (NTRS)

    Still, Martin; Mushotzky, Richard

    2004-01-01

    We present XMM-Newton RGS and EPIC data of the putative cooling flow cluster Abell 2597. Velocities of the low-ionization emission lines in the spectrum are blue shifted with respect to the high-ionization lines by 1320 (sup +660) (sub -210) kilometers per second, which is consistent with the difference in the two peaks of the galaxy velocity distribution and may be the signature of bulk turbulence, infall, rotation or damped oscillation in the cluster. A hierarchical velocity structure such as this could be the direct result of galaxy mergers in the cluster core, or the injection of power into the cluster gas from a central engine. The uniform X-ray morphology of the cluster, the absence of fine scale temperature structure and the random distribution of the the galaxy positions, independent of velocity, suggests that our line of sight is close to the direction of motion. These results have strong implications for cooling flow models of the cluster Abell 2597. They give impetus to those models which account for the observed temperature structure of some clusters using mergers instead of cooling flows.

  7. A Cluster Randomized Trial of Tailored Breastfeeding Support for Women with Gestational Diabetes

    PubMed Central

    Bonuck, Karen; Adatorwovor, Reuben; Schwartz, Todd A.; Berry, Diane C.

    2016-01-01

    Abstract Background: Women with gestational diabetes mellitus (GDM) and their infants are at increased risk of developing metabolic disease; however, longer breastfeeding is associated with a reduction in these risks. We tested an intervention to increase breastfeeding duration among women with GDM. Materials and Methods: We conducted a cluster randomized trial to determine the efficacy of a breastfeeding education and support program for women with GDM. Women were enrolled between 22 and 36 weeks of pregnancy and cluster randomized to an experimental lifestyle intervention or wait-list control group. Breastfeeding duration and intensity were prespecified secondary outcomes of the trial. Duration of exclusive and any breastfeeding was assessed at 6 weeks and at 4, 7, and 10 months postpartum. We quantified differences in breastfeeding rates using Kaplan–Meier estimates, log-rank tests, and Cox regression models. Results: We enrolled 100 women, of whom 52% were African American, 31% non-Hispanic white, 11% Hispanic, 9% American Indian or Alaskan Native, 2% Asian, 2% other, and 4% more than one race. In models accounting for within-cluster correlation and adjusted for study site, breastfeeding intention, and African American race, women allocated to the intervention group were less likely to stop breastfeeding (adjusted hazard ratio [HR] 0.40, 95% confidence interval [CI] 0.21–0.74) or to introduce formula (adjusted HR 0.50, 95% CI 0.34–0.72). Conclusion: Our results suggest that targeted breastfeeding education for women with GDM is feasible and efficacious. Clinical Trials Registration: http://clinicaltrials.gov/ct2/show/NCT01809431 PMID:27782758

  8. Effects of cluster location and cluster distribution on performance on the traveling salesman problem.

    PubMed

    MacGregor, James N

    2015-10-01

    Research on human performance in solving traveling salesman problems typically uses point sets as stimuli, and most models have proposed a processing stage at which stimulus dots are clustered. However, few empirical studies have investigated the effects of clustering on performance. In one recent study, researchers compared the effects of clustered, random, and regular stimuli, and concluded that clustering facilitates performance (Dry, Preiss, & Wagemans, 2012). Another study suggested that these results may have been influenced by the location rather than the degree of clustering (MacGregor, 2013). Two experiments are reported that mark an attempt to disentangle these factors. The first experiment tested several combinations of degree of clustering and cluster location, and revealed mixed evidence that clustering influences performance. In a second experiment, both factors were varied independently, showing that they interact. The results are discussed in terms of the importance of clustering effects, in particular, and perceptual factors, in general, during performance of the traveling salesman problem.

  9. K-Means Algorithm Performance Analysis With Determining The Value Of Starting Centroid With Random And KD-Tree Method

    NASA Astrophysics Data System (ADS)

    Sirait, Kamson; Tulus; Budhiarti Nababan, Erna

    2017-12-01

    Clustering methods that have high accuracy and time efficiency are necessary for the filtering process. One method that has been known and applied in clustering is K-Means Clustering. In its application, the determination of the begining value of the cluster center greatly affects the results of the K-Means algorithm. This research discusses the results of K-Means Clustering with starting centroid determination with a random and KD-Tree method. The initial determination of random centroid on the data set of 1000 student academic data to classify the potentially dropout has a sse value of 952972 for the quality variable and 232.48 for the GPA, whereas the initial centroid determination by KD-Tree has a sse value of 504302 for the quality variable and 214,37 for the GPA variable. The smaller sse values indicate that the result of K-Means Clustering with initial KD-Tree centroid selection have better accuracy than K-Means Clustering method with random initial centorid selection.

  10. A Framework for Designing Cluster Randomized Trials with Binary Outcomes

    ERIC Educational Resources Information Center

    Spybrook, Jessaca; Martinez, Andres

    2011-01-01

    The purpose of this paper is to provide a frame work for approaching a power analysis for a CRT (cluster randomized trial) with a binary outcome. The authors suggest a framework in the context of a simple CRT and then extend it to a blocked design, or a multi-site cluster randomized trial (MSCRT). The framework is based on proportions, an…

  11. Understanding Statistical Power in Cluster Randomized Trials: Challenges Posed by Differences in Notation and Terminology

    ERIC Educational Resources Information Center

    Spybrook, Jessaca; Hedges, Larry; Borenstein, Michael

    2014-01-01

    Research designs in which clusters are the unit of randomization are quite common in the social sciences. Given the multilevel nature of these studies, the power analyses for these studies are more complex than in a simple individually randomized trial. Tools are now available to help researchers conduct power analyses for cluster randomized…

  12. A Comparison of Single Sample and Bootstrap Methods to Assess Mediation in Cluster Randomized Trials

    ERIC Educational Resources Information Center

    Pituch, Keenan A.; Stapleton, Laura M.; Kang, Joo Youn

    2006-01-01

    A Monte Carlo study examined the statistical performance of single sample and bootstrap methods that can be used to test and form confidence interval estimates of indirect effects in two cluster randomized experimental designs. The designs were similar in that they featured random assignment of clusters to one of two treatment conditions and…

  13. Multiple filters affect tree species assembly in mid-latitude forest communities.

    PubMed

    Kubota, Y; Kusumoto, B; Shiono, T; Ulrich, W

    2018-05-01

    Species assembly patterns of local communities are shaped by the balance between multiple abiotic/biotic filters and dispersal that both select individuals from species pools at the regional scale. Knowledge regarding functional assembly can provide insight into the relative importance of the deterministic and stochastic processes that shape species assembly. We evaluated the hierarchical roles of the α niche and β niches by analyzing the influence of environmental filtering relative to functional traits on geographical patterns of tree species assembly in mid-latitude forests. Using forest plot datasets, we examined the α niche traits (leaf and wood traits) and β niche properties (cold/drought tolerance) of tree species, and tested non-randomness (clustering/over-dispersion) of trait assembly based on null models that assumed two types of species pools related to biogeographical regions. For most plots, species assembly patterns fell within the range of random expectation. However, particularly for cold/drought tolerance-related β niche properties, deviation from randomness was frequently found; non-random clustering was predominant in higher latitudes with harsh climates. Our findings demonstrate that both randomness and non-randomness in trait assembly emerged as a result of the α and β niches, although we suggest the potential role of dispersal processes and/or species equalization through trait similarities in generating the prevalence of randomness. Clustering of β niche traits along latitudinal climatic gradients provides clear evidence of species sorting by filtering particular traits. Our results reveal that multiple filters through functional niches and stochastic processes jointly shape geographical patterns of species assembly across mid-latitude forests.

  14. The correlation function for density perturbations in an expanding universe. III The three-point and predictions of the four-point and higher order correlation functions

    NASA Technical Reports Server (NTRS)

    Mcclelland, J.; Silk, J.

    1978-01-01

    Higher-order correlation functions for the large-scale distribution of galaxies in space are investigated. It is demonstrated that the three-point correlation function observed by Peebles and Groth (1975) is not consistent with a distribution of perturbations that at present are randomly distributed in space. The two-point correlation function is shown to be independent of how the perturbations are distributed spatially, and a model of clustered perturbations is developed which incorporates a nonuniform perturbation distribution and which explains the three-point correlation function. A model with hierarchical perturbations incorporating the same nonuniform distribution is also constructed; it is found that this model also explains the three-point correlation function, but predicts different results for the four-point and higher-order correlation functions than does the model with clustered perturbations. It is suggested that the model of hierarchical perturbations might be explained by the single assumption of having density fluctuations or discrete objects all of the same mass randomly placed at some initial epoch.

  15. Generating clustered scale-free networks using Poisson based localization of edges

    NASA Astrophysics Data System (ADS)

    Türker, İlker

    2018-05-01

    We introduce a variety of network models using a Poisson-based edge localization strategy, which result in clustered scale-free topologies. We first verify the success of our localization strategy by realizing a variant of the well-known Watts-Strogatz model with an inverse approach, implying a small-world regime of rewiring from a random network through a regular one. We then apply the rewiring strategy to a pure Barabasi-Albert model and successfully achieve a small-world regime, with a limited capacity of scale-free property. To imitate the high clustering property of scale-free networks with higher accuracy, we adapted the Poisson-based wiring strategy to a growing network with the ingredients of both preferential attachment and local connectivity. To achieve the collocation of these properties, we used a routine of flattening the edges array, sorting it, and applying a mixing procedure to assemble both global connections with preferential attachment and local clusters. As a result, we achieved clustered scale-free networks with a computational fashion, diverging from the recent studies by following a simple but efficient approach.

  16. Intra-class correlation estimates for assessment of vitamin A intake in children.

    PubMed

    Agarwal, Girdhar G; Awasthi, Shally; Walter, Stephen D

    2005-03-01

    In many community-based surveys, multi-level sampling is inherent in the design. In the design of these studies, especially to calculate the appropriate sample size, investigators need good estimates of intra-class correlation coefficient (ICC), along with the cluster size, to adjust for variation inflation due to clustering at each level. The present study used data on the assessment of clinical vitamin A deficiency and intake of vitamin A-rich food in children in a district in India. For the survey, 16 households were sampled from 200 villages nested within eight randomly-selected blocks of the district. ICCs and components of variances were estimated from a three-level hierarchical random effects analysis of variance model. Estimates of ICCs and variance components were obtained at village and block levels. Between-cluster variation was evident at each level of clustering. In these estimates, ICCs were inversely related to cluster size, but the design effect could be substantial for large clusters. At the block level, most ICC estimates were below 0.07. At the village level, many ICC estimates ranged from 0.014 to 0.45. These estimates may provide useful information for the design of epidemiological studies in which the sampled (or allocated) units range in size from households to large administrative zones.

  17. Using Design-Based Latent Growth Curve Modeling with Cluster-Level Predictor to Address Dependency

    ERIC Educational Resources Information Center

    Wu, Jiun-Yu; Kwok, Oi-Man; Willson, Victor L.

    2014-01-01

    The authors compared the effects of using the true Multilevel Latent Growth Curve Model (MLGCM) with single-level regular and design-based Latent Growth Curve Models (LGCM) with or without the higher-level predictor on various criterion variables for multilevel longitudinal data. They found that random effect estimates were biased when the…

  18. Diffusion maps, clustering and fuzzy Markov modeling in peptide folding transitions

    NASA Astrophysics Data System (ADS)

    Nedialkova, Lilia V.; Amat, Miguel A.; Kevrekidis, Ioannis G.; Hummer, Gerhard

    2014-09-01

    Using the helix-coil transitions of alanine pentapeptide as an illustrative example, we demonstrate the use of diffusion maps in the analysis of molecular dynamics simulation trajectories. Diffusion maps and other nonlinear data-mining techniques provide powerful tools to visualize the distribution of structures in conformation space. The resulting low-dimensional representations help in partitioning conformation space, and in constructing Markov state models that capture the conformational dynamics. In an initial step, we use diffusion maps to reduce the dimensionality of the conformational dynamics of Ala5. The resulting pretreated data are then used in a clustering step. The identified clusters show excellent overlap with clusters obtained previously by using the backbone dihedral angles as input, with small—but nontrivial—differences reflecting torsional degrees of freedom ignored in the earlier approach. We then construct a Markov state model describing the conformational dynamics in terms of a discrete-time random walk between the clusters. We show that by combining fuzzy C-means clustering with a transition-based assignment of states, we can construct robust Markov state models. This state-assignment procedure suppresses short-time memory effects that result from the non-Markovianity of the dynamics projected onto the space of clusters. In a comparison with previous work, we demonstrate how manifold learning techniques may complement and enhance informed intuition commonly used to construct reduced descriptions of the dynamics in molecular conformation space.

  19. Diffusion maps, clustering and fuzzy Markov modeling in peptide folding transitions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nedialkova, Lilia V.; Amat, Miguel A.; Kevrekidis, Ioannis G., E-mail: yannis@princeton.edu, E-mail: gerhard.hummer@biophys.mpg.de

    Using the helix-coil transitions of alanine pentapeptide as an illustrative example, we demonstrate the use of diffusion maps in the analysis of molecular dynamics simulation trajectories. Diffusion maps and other nonlinear data-mining techniques provide powerful tools to visualize the distribution of structures in conformation space. The resulting low-dimensional representations help in partitioning conformation space, and in constructing Markov state models that capture the conformational dynamics. In an initial step, we use diffusion maps to reduce the dimensionality of the conformational dynamics of Ala5. The resulting pretreated data are then used in a clustering step. The identified clusters show excellent overlapmore » with clusters obtained previously by using the backbone dihedral angles as input, with small—but nontrivial—differences reflecting torsional degrees of freedom ignored in the earlier approach. We then construct a Markov state model describing the conformational dynamics in terms of a discrete-time random walk between the clusters. We show that by combining fuzzy C-means clustering with a transition-based assignment of states, we can construct robust Markov state models. This state-assignment procedure suppresses short-time memory effects that result from the non-Markovianity of the dynamics projected onto the space of clusters. In a comparison with previous work, we demonstrate how manifold learning techniques may complement and enhance informed intuition commonly used to construct reduced descriptions of the dynamics in molecular conformation space.« less

  20. Diffusion maps, clustering and fuzzy Markov modeling in peptide folding transitions

    PubMed Central

    Nedialkova, Lilia V.; Amat, Miguel A.; Kevrekidis, Ioannis G.; Hummer, Gerhard

    2014-01-01

    Using the helix-coil transitions of alanine pentapeptide as an illustrative example, we demonstrate the use of diffusion maps in the analysis of molecular dynamics simulation trajectories. Diffusion maps and other nonlinear data-mining techniques provide powerful tools to visualize the distribution of structures in conformation space. The resulting low-dimensional representations help in partitioning conformation space, and in constructing Markov state models that capture the conformational dynamics. In an initial step, we use diffusion maps to reduce the dimensionality of the conformational dynamics of Ala5. The resulting pretreated data are then used in a clustering step. The identified clusters show excellent overlap with clusters obtained previously by using the backbone dihedral angles as input, with small—but nontrivial—differences reflecting torsional degrees of freedom ignored in the earlier approach. We then construct a Markov state model describing the conformational dynamics in terms of a discrete-time random walk between the clusters. We show that by combining fuzzy C-means clustering with a transition-based assignment of states, we can construct robust Markov state models. This state-assignment procedure suppresses short-time memory effects that result from the non-Markovianity of the dynamics projected onto the space of clusters. In a comparison with previous work, we demonstrate how manifold learning techniques may complement and enhance informed intuition commonly used to construct reduced descriptions of the dynamics in molecular conformation space. PMID:25240340

  1. New Estimates of Design Parameters for Clustered Randomization Studies: Findings from North Carolina and Florida. Working Paper 43

    ERIC Educational Resources Information Center

    Xu, Zeyu; Nichols, Austin

    2010-01-01

    The gold standard in making causal inference on program effects is a randomized trial. Most randomization designs in education randomize classrooms or schools rather than individual students. Such "clustered randomization" designs have one principal drawback: They tend to have limited statistical power or precision. This study aims to…

  2. When is informed consent required in cluster randomized trials in health research?

    PubMed Central

    2011-01-01

    This article is part of a series of papers examining ethical issues in cluster randomized trials (CRTs) in health research. In the introductory paper in this series, we set out six areas of inquiry that must be addressed if the cluster trial is to be set on a firm ethical foundation. This paper addresses the second of the questions posed, namely, from whom, when, and how must informed consent be obtained in CRTs in health research? The ethical principle of respect for persons implies that researchers are generally obligated to obtain the informed consent of research subjects. Aspects of CRT design, including cluster randomization, cluster level interventions, and cluster size, present challenges to obtaining informed consent. Here we address five questions related to consent and CRTs: How can a study proceed if informed consent is not possible? Is consent to randomization always required? What information must be disclosed to potential subjects if their cluster has already been randomized? Is passive consent a valid substitute for informed consent? Do health professionals have a moral obligation to participate as subjects in CRTs designed to improve professional practice? We set out a framework based on the moral foundations of informed consent and international regulatory provisions to address each of these questions. First, when informed consent is not possible, a study may proceed if a research ethics committee is satisfied that conditions for a waiver of consent are satisfied. Second, informed consent to randomization may not be required if it is not possible to approach subjects at the time of randomization. Third, when potential subjects are approached after cluster randomization, they must be provided with a detailed description of the interventions in the trial arm to which their cluster has been randomized; detailed information on interventions in other trial arms need not be provided. Fourth, while passive consent may serve a variety of practical ends, it is not a substitute for valid informed consent. Fifth, while health professionals may have a moral obligation to participate as subjects in research, this does not diminish the necessity of informed consent to study participation. PMID:21906277

  3. Propensity score to detect baseline imbalance in cluster randomized trials: the role of the c-statistic.

    PubMed

    Leyrat, Clémence; Caille, Agnès; Foucher, Yohann; Giraudeau, Bruno

    2016-01-22

    Despite randomization, baseline imbalance and confounding bias may occur in cluster randomized trials (CRTs). Covariate imbalance may jeopardize the validity of statistical inferences if they occur on prognostic factors. Thus, the diagnosis of a such imbalance is essential to adjust statistical analysis if required. We developed a tool based on the c-statistic of the propensity score (PS) model to detect global baseline covariate imbalance in CRTs and assess the risk of confounding bias. We performed a simulation study to assess the performance of the proposed tool and applied this method to analyze the data from 2 published CRTs. The proposed method had good performance for large sample sizes (n =500 per arm) and when the number of unbalanced covariates was not too small as compared with the total number of baseline covariates (≥40% of unbalanced covariates). We also provide a strategy for pre selection of the covariates needed to be included in the PS model to enhance imbalance detection. The proposed tool could be useful in deciding whether covariate adjustment is required before performing statistical analyses of CRTs.

  4. Promoting the Development of Preschool Children's Emergent Literacy Skills: A Randomized Evaluation of a Literacy-Focused Curriculum and Two Professional Development Models

    ERIC Educational Resources Information Center

    Lonigan, Christopher J.; Farver, JoAnn M.; Phillips, Beth M.; Clancy-Menchetti, Jeanine

    2011-01-01

    To date, there have been few causally interpretable evaluations of the impacts of preschool curricula on the skills of children at-risk for academic difficulties, and even fewer studies have demonstrated statistically significant or educationally meaningful effects. In this cluster-randomized study, we evaluated the impacts of a literacy-focused…

  5. Cancerous tumor: the high frequency of a rare event.

    PubMed

    Galam, S; Radomski, J P

    2001-05-01

    A simple model for cancer growth is presented using cellular automata. Cells diffuse randomly on a two-dimensional square lattice. Individual cells can turn cancerous at a very low rate. During each diffusive step, local fights may occur between healthy and cancerous cells. Associated outcomes depend on some biased local rules, which are independent of the overall cancerous cell density. The models unique ingredients are the frequency of local fights and the bias amplitude. While each isolated cancerous cell is eventually destroyed, an initial two-cell tumor cluster is found to have a nonzero probabilty to spread over the whole system. The associated phase diagram for survival or death is obtained as a function of both the rate of fight and the bias distribution. Within the model, although the occurrence of a killing cluster is a very rare event, it turns out to happen almost systematically over long periods of time, e.g., on the order of an adults life span. Thus, after some age, survival from tumorous cancer becomes random.

  6. Random growth lattice filling model of percolation: a crossover from continuous to discontinuous transition

    NASA Astrophysics Data System (ADS)

    Roy, Bappaditya; Santra, S. B.

    2018-05-01

    A random growth lattice filling model of percolation with a touch and stop growth rule is developed and studied numerically on a two dimensional square lattice. Nucleation centers are continuously added one at a time to the empty lattice sites and clusters are grown from these nucleation centers with a growth probability g. For a given g (), the system passes through a critical point during the growth process where the transition from a disconnected to a connected phase occurs. The model is found to exhibit second order continuous percolation transitions as ordinary percolation for whereas for it exhibits weak first order discontinuous percolation transitions. The continuous transitions are characterized by estimating the values of the critical exponents associated with the order parameter fluctuation and the fractal dimension of the spanning cluster over the whole range of g. The discontinuous transitions, however, are characterized by a compact spanning cluster, lattice size independent fluctuation of the order parameter per lattice, departure from power law scaling in the cluster size distribution and weak bimodal distribution of the order parameter. The nature of transitions are further confirmed by studying the Binder cumulant. Instead of a sharp tricritical point, a tricritical region is found to occur for 0.5  <  g  <  0.8 within which the values of the critical exponents change continuously until the crossover from continuous to discontinuous transition is completed.

  7. Learning Bayesian Networks from Correlated Data

    NASA Astrophysics Data System (ADS)

    Bae, Harold; Monti, Stefano; Montano, Monty; Steinberg, Martin H.; Perls, Thomas T.; Sebastiani, Paola

    2016-05-01

    Bayesian networks are probabilistic models that represent complex distributions in a modular way and have become very popular in many fields. There are many methods to build Bayesian networks from a random sample of independent and identically distributed observations. However, many observational studies are designed using some form of clustered sampling that introduces correlations between observations within the same cluster and ignoring this correlation typically inflates the rate of false positive associations. We describe a novel parameterization of Bayesian networks that uses random effects to model the correlation within sample units and can be used for structure and parameter learning from correlated data without inflating the Type I error rate. We compare different learning metrics using simulations and illustrate the method in two real examples: an analysis of genetic and non-genetic factors associated with human longevity from a family-based study, and an example of risk factors for complications of sickle cell anemia from a longitudinal study with repeated measures.

  8. Integrating association data and disease dynamics: an illustration using African Buffalo in Kruger National Park

    USGS Publications Warehouse

    Cross, Paul C.; James O, Lloyd-Smith; Bowers, Justin A.; Hay, Craig T.; Hofmeyr, Markus; Getz, Wayne M.

    2004-01-01

    Recognition is a prerequisite for non-random association amongst individuals. We explore how non-random association patterns (i.e. who spends time with whom) affect disease dynamics. We estimated the amount of time individuals spent together per month using radio-tracking data from African buffalo and incorporated these data into a dynamic social network model. The dynamic nature of the network has a strong influence on simulated disease dynamics particularly for diseases with shorter infectious periods. Cluster analyses of the association data demonstrated that buffalo herds were not as well defined as previously thought. Associations were more tightly clustered in 2002 than 2003, perhaps due to drier conditions in 2003. As a result, diseases may spread faster during drought conditions due to increased population mixing. Association data are often collected but this is the first use of empirical data in a network disease model in a wildlife population.

  9. Cluster-level statistical inference in fMRI datasets: The unexpected behavior of random fields in high dimensions.

    PubMed

    Bansal, Ravi; Peterson, Bradley S

    2018-06-01

    Identifying regional effects of interest in MRI datasets usually entails testing a priori hypotheses across many thousands of brain voxels, requiring control for false positive findings in these multiple hypotheses testing. Recent studies have suggested that parametric statistical methods may have incorrectly modeled functional MRI data, thereby leading to higher false positive rates than their nominal rates. Nonparametric methods for statistical inference when conducting multiple statistical tests, in contrast, are thought to produce false positives at the nominal rate, which has thus led to the suggestion that previously reported studies should reanalyze their fMRI data using nonparametric tools. To understand better why parametric methods may yield excessive false positives, we assessed their performance when applied both to simulated datasets of 1D, 2D, and 3D Gaussian Random Fields (GRFs) and to 710 real-world, resting-state fMRI datasets. We showed that both the simulated 2D and 3D GRFs and the real-world data contain a small percentage (<6%) of very large clusters (on average 60 times larger than the average cluster size), which were not present in 1D GRFs. These unexpectedly large clusters were deemed statistically significant using parametric methods, leading to empirical familywise error rates (FWERs) as high as 65%: the high empirical FWERs were not a consequence of parametric methods failing to model spatial smoothness accurately, but rather of these very large clusters that are inherently present in smooth, high-dimensional random fields. In fact, when discounting these very large clusters, the empirical FWER for parametric methods was 3.24%. Furthermore, even an empirical FWER of 65% would yield on average less than one of those very large clusters in each brain-wide analysis. Nonparametric methods, in contrast, estimated distributions from those large clusters, and therefore, by construct rejected the large clusters as false positives at the nominal FWERs. Those rejected clusters were outlying values in the distribution of cluster size but cannot be distinguished from true positive findings without further analyses, including assessing whether fMRI signal in those regions correlates with other clinical, behavioral, or cognitive measures. Rejecting the large clusters, however, significantly reduced the statistical power of nonparametric methods in detecting true findings compared with parametric methods, which would have detected most true findings that are essential for making valid biological inferences in MRI data. Parametric analyses, in contrast, detected most true findings while generating relatively few false positives: on average, less than one of those very large clusters would be deemed a true finding in each brain-wide analysis. We therefore recommend the continued use of parametric methods that model nonstationary smoothness for cluster-level, familywise control of false positives, particularly when using a Cluster Defining Threshold of 2.5 or higher, and subsequently assessing rigorously the biological plausibility of the findings, even for large clusters. Finally, because nonparametric methods yielded a large reduction in statistical power to detect true positive findings, we conclude that the modest reduction in false positive findings that nonparametric analyses afford does not warrant a re-analysis of previously published fMRI studies using nonparametric techniques. Copyright © 2018 Elsevier Inc. All rights reserved.

  10. Novel layered clustering-based approach for generating ensemble of classifiers.

    PubMed

    Rahman, Ashfaqur; Verma, Brijesh

    2011-05-01

    This paper introduces a novel concept for creating an ensemble of classifiers. The concept is based on generating an ensemble of classifiers through clustering of data at multiple layers. The ensemble classifier model generates a set of alternative clustering of a dataset at different layers by randomly initializing the clustering parameters and trains a set of base classifiers on the patterns at different clusters in different layers. A test pattern is classified by first finding the appropriate cluster at each layer and then using the corresponding base classifier. The decisions obtained at different layers are fused into a final verdict using majority voting. As the base classifiers are trained on overlapping patterns at different layers, the proposed approach achieves diversity among the individual classifiers. Identification of difficult-to-classify patterns through clustering as well as achievement of diversity through layering leads to better classification results as evidenced from the experimental results.

  11. A machine learning approach for ranking clusters of docked protein‐protein complexes by pairwise cluster comparison

    PubMed Central

    Pfeiffenberger, Erik; Chaleil, Raphael A.G.; Moal, Iain H.

    2017-01-01

    ABSTRACT Reliable identification of near‐native poses of docked protein–protein complexes is still an unsolved problem. The intrinsic heterogeneity of protein–protein interactions is challenging for traditional biophysical or knowledge based potentials and the identification of many false positive binding sites is not unusual. Often, ranking protocols are based on initial clustering of docked poses followed by the application of an energy function to rank each cluster according to its lowest energy member. Here, we present an approach of cluster ranking based not only on one molecular descriptor (e.g., an energy function) but also employing a large number of descriptors that are integrated in a machine learning model, whereby, an extremely randomized tree classifier based on 109 molecular descriptors is trained. The protocol is based on first locally enriching clusters with additional poses, the clusters are then characterized using features describing the distribution of molecular descriptors within the cluster, which are combined into a pairwise cluster comparison model to discriminate near‐native from incorrect clusters. The results show that our approach is able to identify clusters containing near‐native protein–protein complexes. In addition, we present an analysis of the descriptors with respect to their power to discriminate near native from incorrect clusters and how data transformations and recursive feature elimination can improve the ranking performance. Proteins 2017; 85:528–543. © 2016 Wiley Periodicals, Inc. PMID:27935158

  12. Pattern selection and super-patterns in the bounded confidence model

    DOE PAGES

    Ben-Naim, E.; Scheel, A.

    2015-10-26

    We study pattern formation in the bounded confidence model of opinion dynamics. In this random process, opinion is quantified by a single variable. Two agents may interact and reach a fair compromise, but only if their difference of opinion falls below a fixed threshold. Starting from a uniform distribution of opinions with compact support, a traveling wave forms and it propagates from the domain boundary into the unstable uniform state. Consequently, the system reaches a steady state with isolated clusters that are separated by distance larger than the interaction range. These clusters form a quasi-periodic pattern where the sizes ofmore » the clusters and the separations between them are nearly constant. We obtain analytically the average separation between clusters L. Interestingly, there are also very small quasi-periodic modulations in the size of the clusters. Furthermore, the spatial periods of these modulations are a series of integers that follow from the continued-fraction representation of the irrational average separation L.« less

  13. Pattern selection and super-patterns in the bounded confidence model

    NASA Astrophysics Data System (ADS)

    Ben-Naim, E.; Scheel, A.

    2015-10-01

    We study pattern formation in the bounded confidence model of opinion dynamics. In this random process, opinion is quantified by a single variable. Two agents may interact and reach a fair compromise, but only if their difference of opinion falls below a fixed threshold. Starting from a uniform distribution of opinions with compact support, a traveling wave forms and it propagates from the domain boundary into the unstable uniform state. Consequently, the system reaches a steady state with isolated clusters that are separated by distance larger than the interaction range. These clusters form a quasi-periodic pattern where the sizes of the clusters and the separations between them are nearly constant. We obtain analytically the average separation between clusters L. Interestingly, there are also very small quasi-periodic modulations in the size of the clusters. The spatial periods of these modulations are a series of integers that follow from the continued-fraction representation of the irrational average separation L.

  14. Methods for sample size determination in cluster randomized trials

    PubMed Central

    Rutterford, Clare; Copas, Andrew; Eldridge, Sandra

    2015-01-01

    Background: The use of cluster randomized trials (CRTs) is increasing, along with the variety in their design and analysis. The simplest approach for their sample size calculation is to calculate the sample size assuming individual randomization and inflate this by a design effect to account for randomization by cluster. The assumptions of a simple design effect may not always be met; alternative or more complicated approaches are required. Methods: We summarise a wide range of sample size methods available for cluster randomized trials. For those familiar with sample size calculations for individually randomized trials but with less experience in the clustered case, this manuscript provides formulae for a wide range of scenarios with associated explanation and recommendations. For those with more experience, comprehensive summaries are provided that allow quick identification of methods for a given design, outcome and analysis method. Results: We present first those methods applicable to the simplest two-arm, parallel group, completely randomized design followed by methods that incorporate deviations from this design such as: variability in cluster sizes; attrition; non-compliance; or the inclusion of baseline covariates or repeated measures. The paper concludes with methods for alternative designs. Conclusions: There is a large amount of methodology available for sample size calculations in CRTs. This paper gives the most comprehensive description of published methodology for sample size calculation and provides an important resource for those designing these trials. PMID:26174515

  15. Android Malware Classification Using K-Means Clustering Algorithm

    NASA Astrophysics Data System (ADS)

    Hamid, Isredza Rahmi A.; Syafiqah Khalid, Nur; Azma Abdullah, Nurul; Rahman, Nurul Hidayah Ab; Chai Wen, Chuah

    2017-08-01

    Malware was designed to gain access or damage a computer system without user notice. Besides, attacker exploits malware to commit crime or fraud. This paper proposed Android malware classification approach based on K-Means clustering algorithm. We evaluate the proposed model in terms of accuracy using machine learning algorithms. Two datasets were selected to demonstrate the practicing of K-Means clustering algorithms that are Virus Total and Malgenome dataset. We classify the Android malware into three clusters which are ransomware, scareware and goodware. Nine features were considered for each types of dataset such as Lock Detected, Text Detected, Text Score, Encryption Detected, Threat, Porn, Law, Copyright and Moneypak. We used IBM SPSS Statistic software for data classification and WEKA tools to evaluate the built cluster. The proposed K-Means clustering algorithm shows promising result with high accuracy when tested using Random Forest algorithm.

  16. Evaluation of a demand-creation intervention for couples' HIV testing services among married or cohabiting individuals in Rakai, Uganda: a cluster-randomized intervention trial.

    PubMed

    Matovu, Joseph K B; Todd, Jim; Wanyenze, Rhoda K; Kairania, Robert; Serwadda, David; Wabwire-Mangen, Fred

    2016-08-08

    Uptake of couples' HIV counseling and testing (couples' HCT) services remains largely low in most settings. We report the effect of a demand-creation intervention trial on couples' HCT uptake among married or cohabiting individuals who had never received couples' HCT. This was a cluster-randomized intervention trial implemented in three study regions with differing HIV prevalence levels (range: 9-43 %) in Rakai district, southwestern Uganda, between February and September 2014. We randomly assigned six clusters (1:1) to receive the intervention or serve as the comparison arm using computer-generated random numbers. In the intervention clusters, individuals attended small group, couple and male-focused interactive sessions, reinforced with testimonies from 'expert couples', and received invitation coupons to test together with their partners at designated health facilities. In the comparison clusters, participants attended general adult health education sessions but received no invitation coupons. The primary outcome was couples' HCT uptake, measured 12 months post-baseline. Baseline data were collected between November 2013 and February 2014 while follow-up data were collected between March and April 2015. We conducted intention-to-treat analysis using a mixed effects Poisson regression model to assess for differences in couples' HCT uptake between the intervention and comparison clusters. Data analysis was conducted using STATA statistical software, version 14.1. Of 2135 married or cohabiting individuals interviewed at baseline, 42 % (n = 846) had ever received couples' HCT. Of those who had never received couples' HCT (n = 1,174), 697 were interviewed in the intervention clusters while 477 were interviewed in the comparison clusters. 73.6 % (n = 513) of those interviewed in the intervention and 82.6 % (n = 394) of those interviewed in the comparison cluster were interviewed at follow-up. Of those interviewed, 72.3 % (n = 371) in the intervention and 65.2 % (n = 257) in the comparison clusters received HCT. Couples' HCT uptake was higher in the intervention than in the comparison clusters (20.3 % versus 13.7 %; adjusted prevalence ratio (aPR) = 1.43, 95 % CI: 1.02, 2.01, P = 0.04). Our findings show that a small group, couple and male-focused, demand-creation intervention reinforced with testimonies from 'expert couples', improved uptake of couples' HCT in this rural setting. ClinicalTrials.gov, NCT02492061 . Date of registration: June 14, 2015.

  17. Robustness and structure of complex networks

    NASA Astrophysics Data System (ADS)

    Shao, Shuai

    This dissertation covers the two major parts of my PhD research on statistical physics and complex networks: i) modeling a new type of attack -- localized attack, and investigating robustness of complex networks under this type of attack; ii) discovering the clustering structure in complex networks and its influence on the robustness of coupled networks. Complex networks appear in every aspect of our daily life and are widely studied in Physics, Mathematics, Biology, and Computer Science. One important property of complex networks is their robustness under attacks, which depends crucially on the nature of attacks and the structure of the networks themselves. Previous studies have focused on two types of attack: random attack and targeted attack, which, however, are insufficient to describe many real-world damages. Here we propose a new type of attack -- localized attack, and study the robustness of complex networks under this type of attack, both analytically and via simulation. On the other hand, we also study the clustering structure in the network, and its influence on the robustness of a complex network system. In the first part, we propose a theoretical framework to study the robustness of complex networks under localized attack based on percolation theory and generating function method. We investigate the percolation properties, including the critical threshold of the phase transition pc and the size of the giant component Pinfinity. We compare localized attack with random attack and find that while random regular (RR) networks are more robust against localized attack, Erdoḧs-Renyi (ER) networks are equally robust under both types of attacks. As for scale-free (SF) networks, their robustness depends crucially on the degree exponent lambda. The simulation results show perfect agreement with theoretical predictions. We also test our model on two real-world networks: a peer-to-peer computer network and an airline network, and find that the real-world networks are much more vulnerable to localized attack compared with random attack. In the second part, we extend the tree-like generating function method to incorporating clustering structure in complex networks. We study the robustness of a complex network system, especially a network of networks (NON) with clustering structure in each network. We find that the system becomes less robust as we increase the clustering coefficient of each network. For a partially dependent network system, we also find that the influence of the clustering coefficient on network robustness decreases as we decrease the coupling strength, and the critical coupling strength qc, at which the first-order phase transition changes to second-order, increases as we increase the clustering coefficient.

  18. Cluster Tails for Critical Power-Law Inhomogeneous Random Graphs

    NASA Astrophysics Data System (ADS)

    van der Hofstad, Remco; Kliem, Sandra; van Leeuwaarden, Johan S. H.

    2018-04-01

    Recently, the scaling limit of cluster sizes for critical inhomogeneous random graphs of rank-1 type having finite variance but infinite third moment degrees was obtained in Bhamidi et al. (Ann Probab 40:2299-2361, 2012). It was proved that when the degrees obey a power law with exponent τ \\in (3,4), the sequence of clusters ordered in decreasing size and multiplied through by n^{-(τ -2)/(τ -1)} converges as n→ ∞ to a sequence of decreasing non-degenerate random variables. Here, we study the tails of the limit of the rescaled largest cluster, i.e., the probability that the scaling limit of the largest cluster takes a large value u, as a function of u. This extends a related result of Pittel (J Combin Theory Ser B 82(2):237-269, 2001) for the Erdős-Rényi random graph to the setting of rank-1 inhomogeneous random graphs with infinite third moment degrees. We make use of delicate large deviations and weak convergence arguments.

  19. Interpreting semantic clustering effects in free recall.

    PubMed

    Manning, Jeremy R; Kahana, Michael J

    2012-07-01

    The order in which participants choose to recall words from a studied list of randomly selected words provides insights into how memories of the words are represented, organised, and retrieved. One pervasive finding is that when a pair of semantically related words (e.g., "cat" and "dog") is embedded in the studied list, the related words are often recalled successively. This tendency to successively recall semantically related words is termed semantic clustering (Bousfield, 1953; Bousfield & Sedgewick, 1944; Cofer, Bruce, & Reicher, 1966). Measuring semantic clustering effects requires making assumptions about which words participants consider to be similar in meaning. However, it is often difficult to gain insights into individual participants' internal semantic models, and for this reason researchers typically rely on standardised semantic similarity metrics. Here we use simulations to gain insights into the expected magnitudes of semantic clustering effects given systematic differences between participants' internal similarity models and the similarity metric used to quantify the degree of semantic clustering. Our results provide a number of useful insights into the interpretation of semantic clustering effects in free recall.

  20. Packing Fraction of a Two-dimensional Eden Model with Random-Sized Particles

    NASA Astrophysics Data System (ADS)

    Kobayashi, Naoki; Yamazaki, Hiroshi

    2018-01-01

    We have performed a numerical simulation of a two-dimensional Eden model with random-size particles. In the present model, the particle radii are generated from a Gaussian distribution with mean μ and standard deviation σ. First, we have examined the bulk packing fraction for the Eden cluster and investigated the effects of the standard deviation and the total number of particles NT. We show that the bulk packing fraction depends on the number of particles and the standard deviation. In particular, for the dependence on the standard deviation, we have determined the asymptotic value of the bulk packing fraction in the limit of the dimensionless standard deviation. This value is larger than the packing fraction obtained in a previous study of the Eden model with uniform-size particles. Secondly, we have investigated the packing fraction of the entire Eden cluster including the effect of the interface fluctuation. We find that the entire packing fraction depends on the number of particles while it is independent of the standard deviation, in contrast to the bulk packing fraction. In a similar way to the bulk packing fraction, we have obtained the asymptotic value of the entire packing fraction in the limit NT → ∞. The obtained value of the entire packing fraction is smaller than that of the bulk value. This fact suggests that the interface fluctuation of the Eden cluster influences the packing fraction.

  1. Anisotropy in Fracking: A Percolation Model for Observed Microseismicity

    NASA Astrophysics Data System (ADS)

    Norris, J. Quinn; Turcotte, Donald L.; Rundle, John B.

    2015-01-01

    Hydraulic fracturing (fracking), using high pressures and a low viscosity fluid, allow the extraction of large quantiles of oil and gas from very low permeability shale formations. The initial production of oil and gas at depth leads to high pressures and an extensive distribution of natural fractures which reduce the pressures. With time these fractures heal, sealing the remaining oil and gas in place. High volume fracking opens the healed fractures allowing the oil and gas to flow to horizontal production wells. We model the injection process using invasion percolation. We use a 2D square lattice of bonds to model the sealed natural fractures. The bonds are assigned random strengths and the fluid, injected at a point, opens the weakest bond adjacent to the growing cluster of opened bonds. Our model exhibits burst dynamics in which the clusters extend rapidly into regions with weak bonds. We associate these bursts with the microseismic activity generated by fracking injections. A principal object of this paper is to study the role of anisotropic stress distributions. Bonds in the y-direction are assigned higher random strengths than bonds in the x-direction. We illustrate the spatial distribution of clusters and the spatial distribution of bursts (small earthquakes) for several degrees of anisotropy. The results are compared with observed distributions of microseismicity in a fracking injection. Both our bursts and the observed microseismicity satisfy Gutenberg-Richter frequency-size statistics.

  2. Effective cluster model of dielectric enhancement in metal-insulator composites

    NASA Astrophysics Data System (ADS)

    Doyle, W. T.; Jacobs, I. S.

    1990-11-01

    The electrical permittivity of a suspension of conducting spheres at high volume loading exhibits a large enhancement above the value predicted by the Clausius-Mossotti approximation. The permittivity enhancement is a dielectric anomaly accompanying a metallization transition that occurs when conducting particles are close packed. In disordered suspensions, close encounters can cause a permittivity enhancement at any volume loading. We attribute the permittivity enhancements typically observed in monodisperse disordered suspensions of conducting spheres to local metallized regions of high density produced by density fluctuations. We model a disordered suspension as a mixture, or mesosuspension, of isolated spheres and random close-packed spherical clusters of arbitrary size. Multipole interactions within the clusters are treated exactly. External interactions between clusters and isolated spheres are treated in the dipole approximation. Model permittivities are compared with Guillien's experimental permittivity measurements [Ann. Phys. (Paris) Ser. 11, 16, 205 (1941)] on liquid suspensions of Hg droplets in oil and with Turner's conductivity measurements [Chem. Eng. Sci. 31, 487 (1976)] on fluidized bed suspensions of ion-exchange resin beads in aqueous solution. New permittivity measurements at 10 GHz on solid suspensions of monodisperse metal spheres in polyurethane are presented and compared with the model permittivities. The effective spherical cluster model is in excellent agreement with the experiments over the entire accessible range of volume loading.

  3. Person mobility in the design and analysis of cluster-randomized cohort prevention trials.

    PubMed

    Vuchinich, Sam; Flay, Brian R; Aber, Lawrence; Bickman, Leonard

    2012-06-01

    Person mobility is an inescapable fact of life for most cluster-randomized (e.g., schools, hospitals, clinic, cities, state) cohort prevention trials. Mobility rates are an important substantive consideration in estimating the effects of an intervention. In cluster-randomized trials, mobility rates are often correlated with ethnicity, poverty and other variables associated with disparity. This raises the possibility that estimated intervention effects may generalize to only the least mobile segments of a population and, thus, create a threat to external validity. Such mobility can also create threats to the internal validity of conclusions from randomized trials. Researchers must decide how to deal with persons who leave study clusters during a trial (dropouts), persons and clusters that do not comply with an assigned intervention, and persons who enter clusters during a trial (late entrants), in addition to the persons who remain for the duration of a trial (stayers). Statistical techniques alone cannot solve the key issues of internal and external validity raised by the phenomenon of person mobility. This commentary presents a systematic, Campbellian-type analysis of person mobility in cluster-randomized cohort prevention trials. It describes four approaches for dealing with dropouts, late entrants and stayers with respect to data collection, analysis and generalizability. The questions at issue are: 1) From whom should data be collected at each wave of data collection? 2) Which cases should be included in the analyses of an intervention effect? and 3) To what populations can trial results be generalized? The conclusions lead to recommendations for the design and analysis of future cluster-randomized cohort prevention trials.

  4. Effect Sizes in Three-Level Cluster-Randomized Experiments

    ERIC Educational Resources Information Center

    Hedges, Larry V.

    2011-01-01

    Research designs involving cluster randomization are becoming increasingly important in educational and behavioral research. Many of these designs involve two levels of clustering or nesting (students within classes and classes within schools). Researchers would like to compute effect size indexes based on the standardized mean difference to…

  5. Effect of Pendant Side-Chain Sterics and Dipole Forces on Short Range Ordering in Random Polyelectrolytes

    NASA Astrophysics Data System (ADS)

    Nwosu, Chinomso; Pandey, Tara; Herring, Andrew; Coughlin, Edward; University of Massachusetts, Amherst Collaboration; Colorado School of Mines Collaboration

    Backbone-to-backbone spacing in polymers is known to be dictated by the length of the pendant side-chains. Dipole forces in random polyelectrolytes lead to ionic clusters with a characteristic spacing that can be observed by SAXS. Repulsion due to side-chain sterics will compete with dipole forces driving cluster formation in random polyelectrolytes. A model study on short range order in anion exchange membranes (AEMs) of quaternized P4VP-ran-PI is presented. Quaternization of P4VP with alkyl bromides having different numbers of carbons, CnBr, introduces pendant side-chains as well as charges. X-ray scattering performed on PQ4VP-ran-PI(CnBr) show that when n <5 the dipole forces dominate leading to the formation of ionic clusters. However, when n >4, the chains remain separated due to sterics, forming a distinct backbone-to-backbone spacing morphology. For n=3, both dipole clustering and backbone spacing can coexist. Crosslinking of the isoprene units increased the coexistence window from n=3 to n=6. Impedance measurements show that a maximum conductivity of 110mS/cm was obtained for PQ4VP-ran-PI(C3Br). A discussion on short range order due to competition, or counter balancing, of steric repulsion and dipole forces will be presented. US Army MURI project (W911NF1010520).

  6. The Relationship of Dynamical Heterogeneity to the Adam-Gibbs and Random First-Order Transition Theories of Glass Formation

    NASA Astrophysics Data System (ADS)

    Starr, Francis; Douglas, Jack; Sastry, Srikanth

    2013-03-01

    We examine measures of dynamical heterogeneity for a bead-spring polymer melt and test how these scales compare with the scales hypothesized by the Adam and Gibbs (AG) and random first-order transition (RFOT) theories. We show that the time scale of the high-mobility clusters and strings is associated with a diffusive time scale, while the low-mobility particles' time scale relates to a structural relaxation time. The difference of the characteristic times naturally explains the decoupling of diffusion and structural relaxation time scales. We examine the appropriateness of identifying the size scales of mobile particle clusters or strings with the size of cooperatively rearranging regions (CRR) in the AG and RFOT theories. We find that the string size appears to be the most consistent measure of CRR for both the AG and RFOT models. Identifying strings or clusters with the``mosaic'' length of the RFOT model relaxes the conventional assumption that the``entropic droplet'' are compact. We also confirm the validity of the entropy formulation of the AG theory, constraining the exponent values of the RFOT theory. This constraint, together with the analysis of size scales, enables us to estimate the characteristic exponents of RFOT.

  7. The need of adequate information to achieve total compliance of mass drug administration in Pekalongan

    NASA Astrophysics Data System (ADS)

    Ginandjar, Praba; Saraswati, Lintang Dian; Taufik, Opik; Nurjazuli; Widjanarko, Bagoes

    2017-02-01

    World Health Organization (WHO) initiated The Global Program to Eliminate Lymphatic Filariasis (LF) through mass drug administration (MDA). Pekalongan started MDA in 2011. Yet the LF prevalence in 2015 remained exceed the threshold (1%). This study aimed to describe the inhibiting factors related to the compliance of MDA in community level. This was a rapid survey with cross sectional approach. A two-stages random sampling was used in this study. In the first stage, 25 clusters were randomly selected from 27 villages with proportionate to population size (PPS) methods (C-Survey). In the second stage, 10 subjects were randomly selected from each cluster. Subject consisted of 250 respondents from 25 selected clusters. Variables consisted of MDA coverage, practice of taking medication during MDA, enabling and inhibiting factors to MDA in community level. The results showed most respondents had poor knowledge on filariasis, which influence awareness of the disease. Health-illness perception, did not receive the drugs, lactation, side effect, and size of the drugs were dominant factors of non-compliance to MDA. MDA information and community empowerment were needed to improve MDA coverage. Further study to explore the appropriate model of socialization will support the success of MDA program

  8. The median hazard ratio: a useful measure of variance and general contextual effects in multilevel survival analysis.

    PubMed

    Austin, Peter C; Wagner, Philippe; Merlo, Juan

    2017-03-15

    Multilevel data occurs frequently in many research areas like health services research and epidemiology. A suitable way to analyze such data is through the use of multilevel regression models (MLRM). MLRM incorporate cluster-specific random effects which allow one to partition the total individual variance into between-cluster variation and between-individual variation. Statistically, MLRM account for the dependency of the data within clusters and provide correct estimates of uncertainty around regression coefficients. Substantively, the magnitude of the effect of clustering provides a measure of the General Contextual Effect (GCE). When outcomes are binary, the GCE can also be quantified by measures of heterogeneity like the Median Odds Ratio (MOR) calculated from a multilevel logistic regression model. Time-to-event outcomes within a multilevel structure occur commonly in epidemiological and medical research. However, the Median Hazard Ratio (MHR) that corresponds to the MOR in multilevel (i.e., 'frailty') Cox proportional hazards regression is rarely used. Analogously to the MOR, the MHR is the median relative change in the hazard of the occurrence of the outcome when comparing identical subjects from two randomly selected different clusters that are ordered by risk. We illustrate the application and interpretation of the MHR in a case study analyzing the hazard of mortality in patients hospitalized for acute myocardial infarction at hospitals in Ontario, Canada. We provide R code for computing the MHR. The MHR is a useful and intuitive measure for expressing cluster heterogeneity in the outcome and, thereby, estimating general contextual effects in multilevel survival analysis. © 2016 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd. © 2016 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

  9. The Design of Cluster Randomized Trials with Random Cross-Classifications

    ERIC Educational Resources Information Center

    Moerbeek, Mirjam; Safarkhani, Maryam

    2018-01-01

    Data from cluster randomized trials do not always have a pure hierarchical structure. For instance, students are nested within schools that may be crossed by neighborhoods, and soldiers are nested within army units that may be crossed by mental health-care professionals. It is important that the random cross-classification is taken into account…

  10. Information sharing and sorting in a community

    NASA Astrophysics Data System (ADS)

    Bhattacherjee, Biplab; Manna, S. S.; Mukherjee, Animesh

    2013-06-01

    We present the results of a detailed numerical study of a model for the sharing and sorting of information in a community consisting of a large number of agents. The information gathering takes place in a sequence of mutual bipartite interactions where randomly selected pairs of agents communicate with each other to enhance their knowledge and sort out the common information. Although our model is less restricted compared to the well-established naming game, the numerical results strongly indicate that the whole set of exponents characterizing this model are different from those of the naming game and they assume nontrivial values. Finally, it appears that in analogy to the emergence of clusters in the phenomenon of percolation, one can define clusters of agents here having the same information. We have studied in detail the growth of the largest cluster in this article and performed its finite-size scaling analysis.

  11. Reproductive pair correlations and the clustering of organisms.

    PubMed

    Young, W R; Roberts, A J; Stuhne, G

    2001-07-19

    Clustering of organisms can be a consequence of social behaviour, or of the response of individuals to chemical and physical cues. Environmental variability can also cause clustering: for example, marine turbulence transports plankton and produces chlorophyll concentration patterns in the upper ocean. Even in a homogeneous environment, nonlinear interactions between species can result in spontaneous pattern formation. Here we show that a population of independent, random-walking organisms ('brownian bugs'), reproducing by binary division and dying at constant rates, spontaneously aggregates. Using an individual-based model, we show that clusters form out of spatially homogeneous initial conditions without environmental variability, predator-prey interactions, kinesis or taxis. The clustering mechanism is reproductively driven-birth must always be adjacent to a living organism. This clustering can overwhelm diffusion and create non-poissonian correlations between pairs (parent and offspring) or organisms, leading to the emergence of patterns.

  12. Segmentation of Large Unstructured Point Clouds Using Octree-Based Region Growing and Conditional Random Fields

    NASA Astrophysics Data System (ADS)

    Bassier, M.; Bonduel, M.; Van Genechten, B.; Vergauwen, M.

    2017-11-01

    Point cloud segmentation is a crucial step in scene understanding and interpretation. The goal is to decompose the initial data into sets of workable clusters with similar properties. Additionally, it is a key aspect in the automated procedure from point cloud data to BIM. Current approaches typically only segment a single type of primitive such as planes or cylinders. Also, current algorithms suffer from oversegmenting the data and are often sensor or scene dependent. In this work, a method is presented to automatically segment large unstructured point clouds of buildings. More specifically, the segmentation is formulated as a graph optimisation problem. First, the data is oversegmented with a greedy octree-based region growing method. The growing is conditioned on the segmentation of planes as well as smooth surfaces. Next, the candidate clusters are represented by a Conditional Random Field after which the most likely configuration of candidate clusters is computed given a set of local and contextual features. The experiments prove that the used method is a fast and reliable framework for unstructured point cloud segmentation. Processing speeds up to 40,000 points per second are recorded for the region growing. Additionally, the recall and precision of the graph clustering is approximately 80%. Overall, nearly 22% of oversegmentation is reduced by clustering the data. These clusters will be classified and used as a basis for the reconstruction of BIM models.

  13. Monte Carlo investigation of the increased radiation deposition due to gold nanoparticles using kilovoltage and megavoltage photons in a 3D randomized cell model.

    PubMed

    Douglass, Michael; Bezak, Eva; Penfold, Scott

    2013-07-01

    Investigation of increased radiation dose deposition due to gold nanoparticles (GNPs) using a 3D computational cell model during x-ray radiotherapy. Two GNP simulation scenarios were set up in Geant4; a single 400 nm diameter gold cluster randomly positioned in the cytoplasm and a 300 nm gold layer around the nucleus of the cell. Using an 80 kVp photon beam, the effect of GNP on the dose deposition in five modeled regions of the cell including cytoplasm, membrane, and nucleus was simulated. Two Geant4 physics lists were tested: the default Livermore and custom built Livermore/DNA hybrid physics list. 10(6) particles were simulated at 840 cells in the simulation. Each cell was randomly placed with random orientation and a diameter varying between 9 and 13 μm. A mathematical algorithm was used to ensure that none of the 840 cells overlapped. The energy dependence of the GNP physical dose enhancement effect was calculated by simulating the dose deposition in the cells with two energy spectra of 80 kVp and 6 MV. The contribution from Auger electrons was investigated by comparing the two GNP simulation scenarios while activating and deactivating atomic de-excitation processes in Geant4. The physical dose enhancement ratio (DER) of GNP was calculated using the Monte Carlo model. The model has demonstrated that the DER depends on the amount of gold and the position of the gold cluster within the cell. Individual cell regions experienced statistically significant (p < 0.05) change in absorbed dose (DER between 1 and 10) depending on the type of gold geometry used. The DER resulting from gold clusters attached to the cell nucleus had the more significant effect of the two cases (DER ≈ 55). The DER value calculated at 6 MV was shown to be at least an order of magnitude smaller than the DER values calculated for the 80 kVp spectrum. Based on simulations, when 80 kVp photons are used, Auger electrons have a statistically insignificant (p < 0.05) effect on the overall dose increase in the cell. The low energy of the Auger electrons produced prevents them from propagating more than 250-500 nm from the gold cluster and, therefore, has a negligible effect on the overall dose increase due to GNP. The results presented in the current work show that the primary dose enhancement is due to the production of additional photoelectrons.

  14. Generic features of the primary relaxation in glass-forming materials (Review Article)

    NASA Astrophysics Data System (ADS)

    Kokshenev, Valery B.

    2017-08-01

    We discuss structural relaxation in molecular and polymeric supercooled liquids, metallic alloys and orientational glass crystals. The study stresses especially the relationships between observables raised from underlying constraints imposed on degrees of freedom of vitrification systems. A self-consistent parametrization of the α-timescale on macroscopic level results in the material-and-model independent universal equation, relating three fundamental temperatures, characteristic of the primary relaxation, that is numerically proven in all studied glass formers. During the primary relaxation, the corresponding small and large mesoscopic clusters modify their size and structure in a self-similar way, regardless of underlying microscopic realizations. We show that cluster-shape similarity, instead of cluster-size fictive divergence, gives rise to universal features observed in primary relaxation. In all glass formers with structural disorder, including orientational-glass materials (with the exception of plastic crystals), structural relaxation is shown to be driven by local random fields. Within the dynamic stochastic approach, the universal subdiffusive dynamics corresponds to random walks on small and large fractals.

  15. Leveraging microfinance to impact HIV and financial behaviors among adolescents and their mothers in West Bengal: a cluster randomized trial.

    PubMed

    Spielberg, Freya; Crookston, Benjamin T; Chanani, Sheila; Kim, Jaewhan; Kline, Sean; Gray, Bobbi L

    2013-01-01

    Microfinance can be used to reach women and adolescent girls with HIV prevention education. We report findings from a cluster-randomized control trial among 55 villages in West Bengal to determine the impact of non-formal education on knowledge, attitudes and behaviors for HIV prevention and savings. Multilevel regression models were used to evaluate differences between groups for key outcomes while adjusting for cluster correlation and differences in baseline characteristics. Women and girls who received HIV education showed significant gains in HIV knowledge, awareness that condoms can prevent HIV, self-efficacy for HIV prevention, and confirmed use of clean needles, as compared to the control group. Condom use was rare and did not improve for women. While HIV testing was uncommon, knowledge of HIV-testing resources significantly increased among girls, and trended in the positive direction among women in intervention groups. Conversely, the savings education showed no impact on financial knowledge or behavior change.

  16. Sensitivity Analysis of Multiple Informant Models When Data Are Not Missing at Random

    ERIC Educational Resources Information Center

    Blozis, Shelley A.; Ge, Xiaojia; Xu, Shu; Natsuaki, Misaki N.; Shaw, Daniel S.; Neiderhiser, Jenae M.; Scaramella, Laura V.; Leve, Leslie D.; Reiss, David

    2013-01-01

    Missing data are common in studies that rely on multiple informant data to evaluate relationships among variables for distinguishable individuals clustered within groups. Estimation of structural equation models using raw data allows for incomplete data, and so all groups can be retained for analysis even if only 1 member of a group contributes…

  17. A Statistical Model for Misreported Binary Outcomes in Clustered RCTs of Education Interventions

    ERIC Educational Resources Information Center

    Schochet, Peter Z.

    2013-01-01

    In education randomized control trials (RCTs), the misreporting of student outcome data could lead to biased estimates of average treatment effects (ATEs) and their standard errors. This article discusses a statistical model that adjusts for misreported binary outcomes for two-level, school-based RCTs, where it is assumed that misreporting could…

  18. Three-Level Models for Indirect Effects in School- and Class-Randomized Experiments in Education

    ERIC Educational Resources Information Center

    Pituch, Keenan A.; Murphy, Daniel L.; Tate, Richard L.

    2009-01-01

    Due to the clustered nature of field data, multi-level modeling has become commonly used to analyze data arising from educational field experiments. While recent methodological literature has focused on multi-level mediation analysis, relatively little attention has been devoted to mediation analysis when three levels (e.g., student, class,…

  19. Efficient sampling of complex network with modified random walk strategies

    NASA Astrophysics Data System (ADS)

    Xie, Yunya; Chang, Shuhua; Zhang, Zhipeng; Zhang, Mi; Yang, Lei

    2018-02-01

    We present two novel random walk strategies, choosing seed node (CSN) random walk and no-retracing (NR) random walk. Different from the classical random walk sampling, the CSN and NR strategies focus on the influences of the seed node choice and path overlap, respectively. Three random walk samplings are applied in the Erdös-Rényi (ER), Barabási-Albert (BA), Watts-Strogatz (WS), and the weighted USAir networks, respectively. Then, the major properties of sampled subnets, such as sampling efficiency, degree distributions, average degree and average clustering coefficient, are studied. The similar conclusions can be reached with these three random walk strategies. Firstly, the networks with small scales and simple structures are conducive to the sampling. Secondly, the average degree and the average clustering coefficient of the sampled subnet tend to the corresponding values of original networks with limited steps. And thirdly, all the degree distributions of the subnets are slightly biased to the high degree side. However, the NR strategy performs better for the average clustering coefficient of the subnet. In the real weighted USAir networks, some obvious characters like the larger clustering coefficient and the fluctuation of degree distribution are reproduced well by these random walk strategies.

  20. Input variable selection and calibration data selection for storm water quality regression models.

    PubMed

    Sun, Siao; Bertrand-Krajewski, Jean-Luc

    2013-01-01

    Storm water quality models are useful tools in storm water management. Interest has been growing in analyzing existing data for developing models for urban storm water quality evaluations. It is important to select appropriate model inputs when many candidate explanatory variables are available. Model calibration and verification are essential steps in any storm water quality modeling. This study investigates input variable selection and calibration data selection in storm water quality regression models. The two selection problems are mutually interacted. A procedure is developed in order to fulfil the two selection tasks in order. The procedure firstly selects model input variables using a cross validation method. An appropriate number of variables are identified as model inputs to ensure that a model is neither overfitted nor underfitted. Based on the model input selection results, calibration data selection is studied. Uncertainty of model performances due to calibration data selection is investigated with a random selection method. An approach using the cluster method is applied in order to enhance model calibration practice based on the principle of selecting representative data for calibration. The comparison between results from the cluster selection method and random selection shows that the former can significantly improve performances of calibrated models. It is found that the information content in calibration data is important in addition to the size of calibration data.

  1. Cluster Randomized Test-Negative Design (CR-TND) Trials: A Novel and Efficient Method to Assess the Efficacy of Community Level Dengue Interventions.

    PubMed

    Anders, Katherine L; Cutcher, Zoe; Kleinschmidt, Immo; Donnelly, Christl A; Ferguson, Neil M; Indriani, Citra; O'Neill, Scott L; Jewell, Nicholas P; Simmons, Cameron P

    2018-05-07

    Cluster randomized trials are the gold standard for assessing efficacy of community-level interventions, such as vector control strategies against dengue. We describe a novel cluster randomized trial methodology with a test-negative design, which offers advantages over traditional approaches. It utilizes outcome-based sampling of patients presenting with a syndrome consistent with the disease of interest, who are subsequently classified as test-positive cases or test-negative controls on the basis of diagnostic testing. We use simulations of a cluster trial to demonstrate validity of efficacy estimates under the test-negative approach. This demonstrates that, provided study arms are balanced for both test-negative and test-positive illness at baseline and that other test-negative design assumptions are met, the efficacy estimates closely match true efficacy. We also briefly discuss analytical considerations for an odds ratio-based effect estimate arising from clustered data, and outline potential approaches to analysis. We conclude that application of the test-negative design to certain cluster randomized trials could increase their efficiency and ease of implementation.

  2. Hybrid Percolation Transition in Cluster Merging Processes: Continuously Varying Exponents

    NASA Astrophysics Data System (ADS)

    Cho, Y. S.; Lee, J. S.; Herrmann, H. J.; Kahng, B.

    2016-01-01

    Consider growing a network, in which every new connection is made between two disconnected nodes. At least one node is chosen randomly from a subset consisting of g fraction of the entire population in the smallest clusters. Here we show that this simple strategy for improving connection exhibits a more unusual phase transition, namely a hybrid percolation transition exhibiting the properties of both first-order and second-order phase transitions. The cluster size distribution of finite clusters at a transition point exhibits power-law behavior with a continuously varying exponent τ in the range 2 <τ (g )≤2.5 . This pattern reveals a necessary condition for a hybrid transition in cluster aggregation processes, which is comparable to the power-law behavior of the avalanche size distribution arising in models with link-deleting processes in interdependent networks.

  3. Jammed systems of oriented needles always percolate on square lattices

    NASA Astrophysics Data System (ADS)

    Kondrat, Grzegorz; Koza, Zbigniew; Brzeski, Piotr

    2017-08-01

    Random sequential adsorption (RSA) is a standard method of modeling adsorption of large molecules at the liquid-solid interface. Several studies have recently conjectured that in the RSA of rectangular needles, or k -mers, on a square lattice, percolation is impossible if the needles are sufficiently long (k of order of several thousand). We refute these claims and present rigorous proof that in any jammed configuration of nonoverlapping, fixed-length, horizontal, or vertical needles on a square lattice, all clusters are percolating clusters.

  4. Space-time clusters for early detection of grizzly bear predation.

    PubMed

    Kermish-Wells, Joseph; Massolo, Alessandro; Stenhouse, Gordon B; Larsen, Terrence A; Musiani, Marco

    2018-01-01

    Accurate detection and classification of predation events is important to determine predation and consumption rates by predators. However, obtaining this information for large predators is constrained by the speed at which carcasses disappear and the cost of field data collection. To accurately detect predation events, researchers have used GPS collar technology combined with targeted site visits. However, kill sites are often investigated well after the predation event due to limited data retrieval options on GPS collars (VHF or UHF downloading) and to ensure crew safety when working with large predators. This can lead to missing information from small-prey (including young ungulates) kill sites due to scavenging and general site deterioration (e.g., vegetation growth). We used a space-time permutation scan statistic (STPSS) clustering method (SaTScan) to detect predation events of grizzly bears ( Ursus arctos ) fitted with satellite transmitting GPS collars. We used generalized linear mixed models to verify predation events and the size of carcasses using spatiotemporal characteristics as predictors. STPSS uses a probability model to compare expected cluster size (space and time) with the observed size. We applied this method retrospectively to data from 2006 to 2007 to compare our method to random GPS site selection. In 2013-2014, we applied our detection method to visit sites one week after their occupation. Both datasets were collected in the same study area. Our approach detected 23 of 27 predation sites verified by visiting 464 random grizzly bear locations in 2006-2007, 187 of which were within space-time clusters and 277 outside. Predation site detection increased by 2.75 times (54 predation events of 335 visited clusters) using 2013-2014 data. Our GLMMs showed that cluster size and duration predicted predation events and carcass size with high sensitivity (0.72 and 0.94, respectively). Coupling GPS satellite technology with clusters using a program based on space-time probability models allows for prompt visits to predation sites. This enables accurate identification of the carcass size and increases fieldwork efficiency in predation studies.

  5. Hierarchical Bayesian modelling of gene expression time series across irregularly sampled replicates and clusters.

    PubMed

    Hensman, James; Lawrence, Neil D; Rattray, Magnus

    2013-08-20

    Time course data from microarrays and high-throughput sequencing experiments require simple, computationally efficient and powerful statistical models to extract meaningful biological signal, and for tasks such as data fusion and clustering. Existing methodologies fail to capture either the temporal or replicated nature of the experiments, and often impose constraints on the data collection process, such as regularly spaced samples, or similar sampling schema across replications. We propose hierarchical Gaussian processes as a general model of gene expression time-series, with application to a variety of problems. In particular, we illustrate the method's capacity for missing data imputation, data fusion and clustering.The method can impute data which is missing both systematically and at random: in a hold-out test on real data, performance is significantly better than commonly used imputation methods. The method's ability to model inter- and intra-cluster variance leads to more biologically meaningful clusters. The approach removes the necessity for evenly spaced samples, an advantage illustrated on a developmental Drosophila dataset with irregular replications. The hierarchical Gaussian process model provides an excellent statistical basis for several gene-expression time-series tasks. It has only a few additional parameters over a regular GP, has negligible additional complexity, is easily implemented and can be integrated into several existing algorithms. Our experiments were implemented in python, and are available from the authors' website: http://staffwww.dcs.shef.ac.uk/people/J.Hensman/.

  6. Random covering of the circle: the configuration-space of the free deposition process

    NASA Astrophysics Data System (ADS)

    Huillet, Thierry

    2003-12-01

    Consider a circle of circumference 1. Throw at random n points, sequentially, on this circle and append clockwise an arc (or rod) of length s to each such point. The resulting random set (the free gas of rods) is a collection of a random number of clusters with random sizes. It models a free deposition process on a 1D substrate. For such processes, we shall consider the occurrence times (number of rods) and probabilities, as n grows, of the following configurations: those avoiding rod overlap (the hard-rod gas), those for which the largest gap is smaller than rod length s (the packing gas), those (parking configurations) for which hard rod and packing constraints are both fulfilled and covering configurations. Special attention is paid to the statistical properties of each such (rare) configuration in the asymptotic density domain when ns = rgr, for some finite density rgr of points. Using results from spacings in the random division of the circle, explicit large deviation rate functions can be computed in each case from state equations. Lastly, a process consisting in selecting at random one of these specific equilibrium configurations (called the observable) can be modelled. When particularized to the parking model, this system produces parking configurations differently from Rényi's random sequential adsorption model.

  7. Applying the zero-inflated Poisson model with random effects to detect abnormal rises in school absenteeism indicating infectious diseases outbreak.

    PubMed

    Song, X X; Zhao, Q; Tao, T; Zhou, C M; Diwan, V K; Xu, B

    2018-05-30

    Records of absenteeism from primary schools are valuable data for infectious diseases surveillance. However, the analysis of the absenteeism is complicated by the data features of clustering at zero, non-independence and overdispersion. This study aimed to generate an appropriate model to handle the absenteeism data collected in a European Commission granted project for infectious disease surveillance in rural China and to evaluate the validity and timeliness of the resulting model for early warnings of infectious disease outbreak. Four steps were taken: (1) building a 'well-fitting' model by the zero-inflated Poisson model with random effects (ZIP-RE) using the absenteeism data from the first implementation year; (2) applying the resulting model to predict the 'expected' number of absenteeism events in the second implementation year; (3) computing the differences between the observations and the expected values (O-E values) to generate an alternative series of data; (4) evaluating the early warning validity and timeliness of the observational data and model-based O-E values via the EARS-3C algorithms with regard to the detection of real cluster events. The results indicate that ZIP-RE and its corresponding O-E values could improve the detection of aberrations, reduce the false-positive signals and are applicable to the zero-inflated data.

  8. Statistical analysis and handling of missing data in cluster randomized trials: a systematic review.

    PubMed

    Fiero, Mallorie H; Huang, Shuang; Oren, Eyal; Bell, Melanie L

    2016-02-09

    Cluster randomized trials (CRTs) randomize participants in groups, rather than as individuals and are key tools used to assess interventions in health research where treatment contamination is likely or if individual randomization is not feasible. Two potential major pitfalls exist regarding CRTs, namely handling missing data and not accounting for clustering in the primary analysis. The aim of this review was to evaluate approaches for handling missing data and statistical analysis with respect to the primary outcome in CRTs. We systematically searched for CRTs published between August 2013 and July 2014 using PubMed, Web of Science, and PsycINFO. For each trial, two independent reviewers assessed the extent of the missing data and method(s) used for handling missing data in the primary and sensitivity analyses. We evaluated the primary analysis and determined whether it was at the cluster or individual level. Of the 86 included CRTs, 80 (93%) trials reported some missing outcome data. Of those reporting missing data, the median percent of individuals with a missing outcome was 19% (range 0.5 to 90%). The most common way to handle missing data in the primary analysis was complete case analysis (44, 55%), whereas 18 (22%) used mixed models, six (8%) used single imputation, four (5%) used unweighted generalized estimating equations, and two (2%) used multiple imputation. Fourteen (16%) trials reported a sensitivity analysis for missing data, but most assumed the same missing data mechanism as in the primary analysis. Overall, 67 (78%) trials accounted for clustering in the primary analysis. High rates of missing outcome data are present in the majority of CRTs, yet handling missing data in practice remains suboptimal. Researchers and applied statisticians should carry out appropriate missing data methods, which are valid under plausible assumptions in order to increase statistical power in trials and reduce the possibility of bias. Sensitivity analysis should be performed, with weakened assumptions regarding the missing data mechanism to explore the robustness of results reported in the primary analysis.

  9. Exploring the transparency mechanism and evaluating the effect of public reporting on prescription: a protocol for a cluster randomized controlled trial.

    PubMed

    Du, Xin; Wang, Dan; Wang, Xuan; Yang, Shiru; Zhang, Xinping

    2015-03-21

    The public reporting of health outcomes has become one of the most popular topics and is accepted as a quality improvement method in the healthcare field. However, little research has been conducted on the transparency mechanism, and results are mixed with regard to the evaluation of the effect of public reporting on quality improvement. The objectives of this trial are to investigate the transparency mechanism and to evaluate the effect of public reporting on prescription at the level of individual participants. This study involves a cluster randomized controlled trial conducted in 20 primary-care facilities (clusters). Eligible clusters are those facilities with excellent hospital information systems and that have agreed to participate in the trial. The 20 clusters are matched into 10 pairs according to Technique for Order Preference by Similarity to Ideal Solution score. As the unit of randomization, each pair of facilities is assigned at random to a control or an intervention group through coin flipping. Prescribed ranking information is publicly reported in the intervention group. The public materials include the posters of individuals and of facilities, the ranking lists of general practitioners, and brochures of patients, which are updated monthly. The intervention began on 13th November 2013 and lasted for one year. Specifically, participants are surveyed at five points in time (baseline, quarterly following the intervention) through questionnaires, interviews, and observations. These participants include an average of 600 patients, 300 general practitioners, 15 directors, and 6 health bureau administrators. The primary outcomes are the transparency mechanism model and the changes in medicine-prescribe. Subsequently, the modifications in the transparency mechanism constructs are evaluated. The outcomes are measured at the individual participant level, and the professional who analyzes the data is blind to the randomization status. This study protocol outlines a design that aims to examine the transparency mechanism and to evaluate the effect of public reporting on prescription. The research design is significant in the field of public policy. Furthermore, this study intends to fill the gap of the investigation of the transparency mechanism and the evaluation of public reporting on prescription.

  10. A quantitative approach to the topology of large-scale structure. [for galactic clustering computation

    NASA Technical Reports Server (NTRS)

    Gott, J. Richard, III; Weinberg, David H.; Melott, Adrian L.

    1987-01-01

    A quantitative measure of the topology of large-scale structure: the genus of density contours in a smoothed density distribution, is described and applied. For random phase (Gaussian) density fields, the mean genus per unit volume exhibits a universal dependence on threshold density, with a normalizing factor that can be calculated from the power spectrum. If large-scale structure formed from the gravitational instability of small-amplitude density fluctuations, the topology observed today on suitable scales should follow the topology in the initial conditions. The technique is illustrated by applying it to simulations of galaxy clustering in a flat universe dominated by cold dark matter. The technique is also applied to a volume-limited sample of the CfA redshift survey and to a model in which galaxies reside on the surfaces of polyhedral 'bubbles'. The topology of the evolved mass distribution and 'biased' galaxy distribution in the cold dark matter models closely matches the topology of the density fluctuations in the initial conditions. The topology of the observational sample is consistent with the random phase, cold dark matter model.

  11. Clustering, randomness, and regularity in cloud fields. 4. Stratocumulus cloud fields

    NASA Astrophysics Data System (ADS)

    Lee, J.; Chou, J.; Weger, R. C.; Welch, R. M.

    1994-07-01

    To complete the analysis of the spatial distribution of boundary layer cloudiness, the present study focuses on nine stratocumulus Landsat scenes. The results indicate many similarities between stratocumulus and cumulus spatial distributions. Most notably, at full spatial resolution all scenes exhibit a decidedly clustered distribution. The strength of the clustering signal decreases with increasing cloud size; the clusters themselves consist of a few clouds (less than 10), occupy a small percentage of the cloud field area (less than 5%), contain between 20% and 60% of the cloud field population, and are randomly located within the scene. In contrast, stratocumulus in almost every respect are more strongly clustered than are cumulus cloud fields. For instance, stratocumulus clusters contain more clouds per cluster, occupy a larger percentage of the total area, and have a larger percentage of clouds participating in clusters than the corresponding cumulus examples. To investigate clustering at intermediate spatial scales, the local dimensionality statistic is introduced. Results obtained from this statistic provide the first direct evidence for regularity among large (>900 m in diameter) clouds in stratocumulus and cumulus cloud fields, in support of the inhibition hypothesis of Ramirez and Bras (1990). Also, the size compensated point-to-cloud cumulative distribution function statistic is found to be necessary to obtain a consistent description of stratocumulus cloud distributions. A hypothesis regarding the underlying physical mechanisms responsible for cloud clustering is presented. It is suggested that cloud clusters often arise from 4 to 10 triggering events localized within regions less than 2 km in diameter and randomly distributed within the cloud field. As the size of the cloud surpasses the scale of the triggering region, the clustering signal weakens and the larger cloud locations become more random.

  12. Clustering, randomness, and regularity in cloud fields. 4: Stratocumulus cloud fields

    NASA Technical Reports Server (NTRS)

    Lee, J.; Chou, J.; Weger, R. C.; Welch, R. M.

    1994-01-01

    To complete the analysis of the spatial distribution of boundary layer cloudiness, the present study focuses on nine stratocumulus Landsat scenes. The results indicate many similarities between stratocumulus and cumulus spatial distributions. Most notably, at full spatial resolution all scenes exhibit a decidedly clustered distribution. The strength of the clustering signal decreases with increasing cloud size; the clusters themselves consist of a few clouds (less than 10), occupy a small percentage of the cloud field area (less than 5%), contain between 20% and 60% of the cloud field population, and are randomly located within the scene. In contrast, stratocumulus in almost every respect are more strongly clustered than are cumulus cloud fields. For instance, stratocumulus clusters contain more clouds per cluster, occupy a larger percentage of the total area, and have a larger percentage of clouds participating in clusters than the corresponding cumulus examples. To investigate clustering at intermediate spatial scales, the local dimensionality statistic is introduced. Results obtained from this statistic provide the first direct evidence for regularity among large (more than 900 m in diameter) clouds in stratocumulus and cumulus cloud fields, in support of the inhibition hypothesis of Ramirez and Bras (1990). Also, the size compensated point-to-cloud cumulative distribution function statistic is found to be necessary to obtain a consistent description of stratocumulus cloud distributions. A hypothesis regarding the underlying physical mechanisms responsible for cloud clustering is presented. It is suggested that cloud clusters often arise from 4 to 10 triggering events localized within regions less than 2 km in diameter and randomly distributed within the cloud field. As the size of the cloud surpasses the scale of the triggering region, the clustering signal weakens and the larger cloud locations become more random.

  13. Point process statistics in atom probe tomography.

    PubMed

    Philippe, T; Duguay, S; Grancher, G; Blavette, D

    2013-09-01

    We present a review of spatial point processes as statistical models that we have designed for the analysis and treatment of atom probe tomography (APT) data. As a major advantage, these methods do not require sampling. The mean distance to nearest neighbour is an attractive approach to exhibit a non-random atomic distribution. A χ(2) test based on distance distributions to nearest neighbour has been developed to detect deviation from randomness. Best-fit methods based on first nearest neighbour distance (1 NN method) and pair correlation function are presented and compared to assess the chemical composition of tiny clusters. Delaunay tessellation for cluster selection has been also illustrated. These statistical tools have been applied to APT experiments on microelectronics materials. Copyright © 2012 Elsevier B.V. All rights reserved.

  14. Kinematical evolution of tidally limited star clusters: rotational properties

    NASA Astrophysics Data System (ADS)

    Tiongco, Maria A.; Vesperini, Enrico; Varri, Anna Lisa

    2017-07-01

    We present the results of a set of N-body simulations following the long-term evolution of the rotational properties of star cluster models evolving in the external tidal field of their host galaxy, after an initial phase of violent relaxation. The effects of two-body relaxation and escape of stars lead to a redistribution of the ordered kinetic energy from the inner to the outer regions, ultimately determining a progressive general loss of angular momentum; these effects are reflected in the overall decline of the rotation curve as the cluster evolves and loses stars. We show that all of our models share the same dependence of the remaining fraction of the initial rotation on the fraction of the initial mass lost. As the cluster evolves and loses part of its initial angular momentum, it becomes increasingly dominated by random motions, but even after several tens of relaxation times, and losing a significant fraction of its initial mass, a cluster can still be characterized by a non-negligible ratio of the rotational velocity to the velocity dispersion. This result is in qualitative agreement with the recently observed kinematical complexity that characterizes several Galactic globular clusters.

  15. Cooperative epidemics on multiplex networks.

    PubMed

    Azimi-Tafreshi, N

    2016-04-01

    The spread of one disease, in some cases, can stimulate the spreading of another infectious disease. Here, we treat analytically a symmetric coinfection model for spreading of two diseases on a two-layer multiplex network. We allow layer overlapping, but we assume that each layer is random and locally loopless. Infection with one of the diseases increases the probability of getting infected with the other. Using the generating function method, we calculate exactly the fraction of individuals infected with both diseases (so-called coinfected clusters) in the stationary state, as well as the epidemic spreading thresholds and the phase diagram of the model. With increasing cooperation, we observe a tricritical point and the type of transition changes from continuous to hybrid. Finally, we compare the coinfected clusters in the case of cooperating diseases with the so-called "viable" clusters in networks with dependencies.

  16. Cooperative epidemics on multiplex networks

    NASA Astrophysics Data System (ADS)

    Azimi-Tafreshi, N.

    2016-04-01

    The spread of one disease, in some cases, can stimulate the spreading of another infectious disease. Here, we treat analytically a symmetric coinfection model for spreading of two diseases on a two-layer multiplex network. We allow layer overlapping, but we assume that each layer is random and locally loopless. Infection with one of the diseases increases the probability of getting infected with the other. Using the generating function method, we calculate exactly the fraction of individuals infected with both diseases (so-called coinfected clusters) in the stationary state, as well as the epidemic spreading thresholds and the phase diagram of the model. With increasing cooperation, we observe a tricritical point and the type of transition changes from continuous to hybrid. Finally, we compare the coinfected clusters in the case of cooperating diseases with the so-called "viable" clusters in networks with dependencies.

  17. Nonlocality and particle-clustering effects on the optical response of composite materials with metallic nanoparticles

    NASA Astrophysics Data System (ADS)

    Chen, C. W.; Chung, H. Y.; Chiang, H.-P.; Lu, J. Y.; Chang, R.; Tsai, D. P.; Leung, P. T.

    2010-10-01

    The optical properties of composites with metallic nanoparticles are studied, taking into account the effects due to the nonlocal dielectric response of the metal and the coalescing of the particles to form clusters. An approach based on various effective medium theories is followed, and the modeling results are compared with those from the cases with local response and particles randomly distributed through the host medium. Possible observations of our modeling results are illustrated via a calculation of the transmission of light through a thin film made of these materials. It is found that the nonlocal effects are particularly significant when the particles coalesce, leading to blue-shifted resonances and slightly lower values in the dielectric functions. The dependence of these effects on the volume fraction and fractal dimension of the metal clusters is studied in detail.

  18. Randomized controlled trial to test the RHANI Wives HIV intervention for women in India at risk for HIV from husbands.

    PubMed

    Raj, Anita; Saggurti, Niranjan; Battala, Madhusudana; Nair, Saritha; Dasgupta, Anindita; Naik, D D; Abramovitz, Daniela; Silverman, Jay G; Balaiah, Donta

    2013-11-01

    This study involved evaluation of the short-term impact of the RHANI Wives HIV intervention among wives at risk for HIV from husbands in Mumbai, India. A two-armed cluster RCT was conducted with 220 women surveyed on marital sex at baseline and 4-5 month follow-up. RHANI Wives was a multisession intervention focused on safer sex, marital communication, gender inequities and violence; control participants received basic HIV prevention education. Generalized linear mixed models were conducted to assess program impact, with cluster as a random effect and with time, treatment group, and the time by treatment interaction as fixed effects. A significant time by treatment effect on proportion of unprotected sex with husband (p = 0.01) was observed, and the rate of unprotected sex for intervention participants was lower than that of control participants at follow-up (RR = 0.83, 95 % CI = 0.75, 0.93). RHANI Wives is a promising model for women at risk for HIV from husbands.

  19. Improving Language Comprehension in Preschool Children with Language Difficulties: A Cluster Randomized Trial

    ERIC Educational Resources Information Center

    Hagen, Åste M.; Melby-Lervåg, Monica; Lervåg, Arne

    2017-01-01

    Background: Children with language comprehension difficulties are at risk of educational and social problems, which in turn impede employment prospects in adulthood. However, few randomized trials have examined how such problems can be ameliorated during the preschool years. Methods: We conducted a cluster randomized trial in 148 preschool…

  20. The Walking School Bus and children's physical activity: A pilot cluster randomized controlled trial

    USDA-ARS?s Scientific Manuscript database

    To evaluate the impact of a "walking school bus" program on children's rates of active commuting to school and physical activity. We conducted a pilot cluster randomized controlled trial among 4th-graders from 8 schools in Houston, Texas (N = 149). Random allocation to treatment or control condition...

  1. Coma cluster ultradiffuse galaxies are not standard radio galaxies

    NASA Astrophysics Data System (ADS)

    Struble, Mitchell F.

    2018-02-01

    Matching members in the Coma cluster catalogue of ultradiffuse galaxies (UDGs) from SUBARU imaging with a very deep radio continuum survey source catalogue of the cluster using the Karl G. Jansky Very Large Array (VLA) within a rectangular region of ∼1.19 deg2 centred on the cluster core reveals matches consistent with random. An overlapping set of 470 UDGs and 696 VLA radio sources in this rectangular area finds 33 matches within a separation of 25 arcsec; dividing the sample into bins with separations bounded by 5, 10, 20 and 25 arcsec finds 1, 4, 17 and 11 matches. An analytical model estimate, based on the Poisson probability distribution, of the number of randomly expected matches within these same separation bounds is 1.7, 4.9, 19.4 and 14.2, each, respectively, consistent with the 95 per cent Poisson confidence intervals of the observed values. Dividing the data into five clustercentric annuli of 0.1° and into the four separation bins, finds the same result. This random match of UDGs with VLA sources implies that UDGs are not radio galaxies by the standard definition. Those VLA sources having integrated flux >1 mJy at 1.4 GHz in Miller, Hornschemeier and Mobasher without SDSS galaxy matches are consistent with the known surface density of background radio sources. We briefly explore the possibility that some unresolved VLA sources near UDGs could be young, compact, bright, supernova remnants of Type Ia events, possibly in the intracluster volume.

  2. GPI-anchored proteins are confined in subdiffraction clusters at the apical surface of polarized epithelial cells

    PubMed Central

    Paladino, Simona; Lebreton, Stéphanie; Lelek, Mickaël; Riccio, Patrizia; De Nicola, Sergio; Zimmer, Christophe

    2017-01-01

    Spatio-temporal compartmentalization of membrane proteins is critical for the regulation of diverse vital functions in eukaryotic cells. It was previously shown that, at the apical surface of polarized MDCK cells, glycosylphosphatidylinositol (GPI)-anchored proteins (GPI-APs) are organized in small cholesterol-independent clusters of single GPI-AP species (homoclusters), which are required for the formation of larger cholesterol-dependent clusters formed by multiple GPI-AP species (heteroclusters). This clustered organization is crucial for the biological activities of GPI-APs; hence, understanding the spatio-temporal properties of their membrane organization is of fundamental importance. Here, by using direct stochastic optical reconstruction microscopy coupled to pair correlation analysis (pc-STORM), we were able to visualize and measure the size of these clusters. Specifically, we show that they are non-randomly distributed and have an average size of 67 nm. We also demonstrated that polarized MDCK and non-polarized CHO cells have similar cluster distribution and size, but different sensitivity to cholesterol depletion. Finally, we derived a model that allowed a quantitative characterization of the cluster organization of GPI-APs at the apical surface of polarized MDCK cells for the first time. Experimental FRET (fluorescence resonance energy transfer)/FLIM (fluorescence-lifetime imaging microscopy) data were correlated to the theoretical predictions of the model. PMID:29046391

  3. Clearing out a maze: A model of chemotactic motion in porous media

    NASA Astrophysics Data System (ADS)

    Schilling, Tanja; Voigtmann, Thomas

    2017-12-01

    We study the anomalous dynamics of a biased "hungry" (or "greedy") random walk on a percolating cluster. The model mimics chemotaxis in a porous medium: In close resemblance to the 1980s arcade game PAC-MA N ®, the hungry random walker consumes food, which is initially distributed in the maze, and biases its movement towards food-filled sites. We observe that the mean-squared displacement of the process follows a power law with an exponent that is different from previously known exponents describing passive or active microswimmer dynamics. The change in dynamics is well described by a dynamical exponent that depends continuously on the propensity to move towards food. It results in slower differential growth when compared to the unbiased random walk.

  4. A Separable Two-Dimensional Random Field Model of Binary Response Data from Multi-Day Behavioral Experiments.

    PubMed

    Malem-Shinitski, Noa; Zhang, Yingzhuo; Gray, Daniel T; Burke, Sara N; Smith, Anne C; Barnes, Carol A; Ba, Demba

    2018-04-18

    The study of learning in populations of subjects can provide insights into the changes that occur in the brain with aging, drug intervention, and psychiatric disease. We introduce a separable two-dimensional (2D) random field (RF) model for analyzing binary response data acquired during the learning of object-reward associations across multiple days. The method can quantify the variability of performance within a day and across days, and can capture abrupt changes in learning. We apply the method to data from young and aged macaque monkeys performing a reversal-learning task. The method provides an estimate of performance within a day for each age group, and a learning rate across days for each monkey. We find that, as a group, the older monkeys require more trials to learn the object discriminations than do the young monkeys, and that the cognitive flexibility of the younger group is higher. We also use the model estimates of performance as features for clustering the monkeys into two groups. The clustering results in two groups that, for the most part, coincide with those formed by the age groups. Simulation studies suggest that clustering captures inter-individual differences in performance levels. In comparison with generalized linear models, this method is better able to capture the inherent two-dimensional nature of the data and find between group differences. Applied to binary response data from groups of individuals performing multi-day behavioral experiments, the model discriminates between-group differences and identifies subgroups. Copyright © 2018. Published by Elsevier B.V.

  5. Applying the Dynamic Social Systems Model to HIV Prevention in a Rural African Context: The Maasai and the "Esoto" Dance

    ERIC Educational Resources Information Center

    Siegler, Aaron J.; Mbwambo, Jessie K.; DiClemente, Ralph J.

    2013-01-01

    This study applied the Dynamic Social Systems Model (DSSM) to the issue of HIV risk among the Maasai tribe of Tanzania, using data from a cross-sectional, cluster survey among 370 randomly selected participants from Ngorongoro and Siha Districts. A culturally appropriate survey instrument was developed to explore traditions reportedly coadunate…

  6. A pilot cluster randomized controlled trial of structured goal-setting following stroke.

    PubMed

    Taylor, William J; Brown, Melanie; William, Levack; McPherson, Kathryn M; Reed, Kirk; Dean, Sarah G; Weatherall, Mark

    2012-04-01

    To determine the feasibility, the cluster design effect and the variance and minimal clinical importance difference in the primary outcome in a pilot study of a structured approach to goal-setting. A cluster randomized controlled trial. Inpatient rehabilitation facilities. People who were admitted to inpatient rehabilitation following stroke who had sufficient cognition to engage in structured goal-setting and complete the primary outcome measure. Structured goal elicitation using the Canadian Occupational Performance Measure. Quality of life at 12 weeks using the Schedule for Individualised Quality of Life (SEIQOL-DW), Functional Independence Measure, Short Form 36 and Patient Perception of Rehabilitation (measuring satisfaction with rehabilitation). Assessors were blinded to the intervention. Four rehabilitation services and 41 patients were randomized. We found high values of the intraclass correlation for the outcome measures (ranging from 0.03 to 0.40) and high variance of the SEIQOL-DW (SD 19.6) in relation to the minimally importance difference of 2.1, leading to impractically large sample size requirements for a cluster randomized design. A cluster randomized design is not a practical means of avoiding contamination effects in studies of inpatient rehabilitation goal-setting. Other techniques for coping with contamination effects are necessary.

  7. Phase Transition Behavior in a Neutral Evolution Model

    NASA Astrophysics Data System (ADS)

    King, Dawn; Scott, Adam; Maric, Nevena; Bahar, Sonya

    2014-03-01

    The complexity of interactions among individuals and between individuals and the environment make agent based modeling ideal for studying emergent speciation. This is a dynamically complex problem that can be characterized via the critical behavior of a continuous phase transition. Concomitant with the main tenets of natural selection, we allow organisms to reproduce, mutate, and die within a neutral phenotype space. Previous work has shown phase transition behavior in an assortative mating model with variable fitness landscapes as the maximum mutation size (μ) was varied (Dees and Bahar, 2010). Similarly, this behavior was recently presented in the work of Scott et al. (2013), even on a completely neutral landscape, for bacterial-like fission as well as for assortative mating. Here we present another neutral model to investigate the `critical' phase transition behavior of three mating types - assortative, bacterial, and random - in a phenotype space as a function of the percentage of random death. Results show two types of phase transitions occurring for the parameters of the population size and the number of clusters (an analogue of species), indicating different evolutionary dynamics for system survival and clustering. This research was supported by funding from: University of Missouri Research Board and James S. McDonnell Foundation.

  8. Intercenter Differences in Bronchopulmonary Dysplasia or Death Among Very Low Birth Weight Infants

    PubMed Central

    Walsh, Michele; Bobashev, Georgiy; Das, Abhik; Levine, Burton; Carlo, Waldemar A.; Higgins, Rosemary D.

    2011-01-01

    OBJECTIVES: To determine (1) the magnitude of clustering of bronchopulmonary dysplasia (36 weeks) or death (the outcome) across centers of the Eunice Kennedy Shriver National Institute of Child and Human Development National Research Network, (2) the infant-level variables associated with the outcome and estimate their clustering, and (3) the center-specific practices associated with the differences and build predictive models. METHODS: Data on neonates with a birth weight of <1250 g from the cluster-randomized benchmarking trial were used to determine the magnitude of clustering of the outcome according to alternating logistic regression by using pairwise odds ratio and predictive modeling. Clinical variables associated with the outcome were identified by using multivariate analysis. The magnitude of clustering was then evaluated after correction for infant-level variables. Predictive models were developed by using center-specific and infant-level variables for data from 2001 2004 and projected to 2006. RESULTS: In 2001–2004, clustering of bronchopulmonary dysplasia/death was significant (pairwise odds ratio: 1.3; P < .001) and increased in 2006 (pairwise odds ratio: 1.6; overall incidence: 52%; range across centers: 32%–74%); center rates were relatively stable over time. Variables that varied according to center and were associated with increased risk of outcome included lower body temperature at NICU admission, use of prophylactic indomethacin, specific drug therapy on day 1, and lack of endotracheal intubation. Center differences remained significant even after correction for clustered variables. CONCLUSION: Bronchopulmonary dysplasia/death rates demonstrated moderate clustering according to center. Clinical variables associated with the outcome were also clustered. Center differences after correction of clustered variables indicate presence of as-yet unmeasured center variables. PMID:21149431

  9. A Bayesian cluster analysis method for single-molecule localization microscopy data.

    PubMed

    Griffié, Juliette; Shannon, Michael; Bromley, Claire L; Boelen, Lies; Burn, Garth L; Williamson, David J; Heard, Nicholas A; Cope, Andrew P; Owen, Dylan M; Rubin-Delanchy, Patrick

    2016-12-01

    Cell function is regulated by the spatiotemporal organization of the signaling machinery, and a key facet of this is molecular clustering. Here, we present a protocol for the analysis of clustering in data generated by 2D single-molecule localization microscopy (SMLM)-for example, photoactivated localization microscopy (PALM) or stochastic optical reconstruction microscopy (STORM). Three features of such data can cause standard cluster analysis approaches to be ineffective: (i) the data take the form of a list of points rather than a pixel array; (ii) there is a non-negligible unclustered background density of points that must be accounted for; and (iii) each localization has an associated uncertainty in regard to its position. These issues are overcome using a Bayesian, model-based approach. Many possible cluster configurations are proposed and scored against a generative model, which assumes Gaussian clusters overlaid on a completely spatially random (CSR) background, before every point is scrambled by its localization precision. We present the process of generating simulated and experimental data that are suitable to our algorithm, the analysis itself, and the extraction and interpretation of key cluster descriptors such as the number of clusters, cluster radii and the number of localizations per cluster. Variations in these descriptors can be interpreted as arising from changes in the organization of the cellular nanoarchitecture. The protocol requires no specific programming ability, and the processing time for one data set, typically containing 30 regions of interest, is ∼18 h; user input takes ∼1 h.

  10. Regularity of beating of small clusters of embryonic chick ventricular heart-cells: experiment vs. stochastic single-channel population model

    NASA Astrophysics Data System (ADS)

    Krogh-Madsen, Trine; Kold Taylor, Louise; Skriver, Anne D.; Schaffer, Peter; Guevara, Michael R.

    2017-09-01

    The transmembrane potential is recorded from small isopotential clusters of 2-4 embryonic chick ventricular cells spontaneously generating action potentials. We analyze the cycle-to-cycle fluctuations in the time between successive action potentials (the interbeat interval or IBI). We also convert an existing model of electrical activity in the cluster, which is formulated as a Hodgkin-Huxley-like deterministic system of nonlinear ordinary differential equations describing five individual ionic currents, into a stochastic model consisting of a population of ˜20 000 independently and randomly gating ionic channels, with the randomness being set by a real physical stochastic process (radio static). This stochastic model, implemented using the Clay-DeFelice algorithm, reproduces the fluctuations seen experimentally: e.g., the coefficient of variation (standard deviation/mean) of IBI is 4.3% in the model vs. the 3.9% average value of the 17 clusters studied. The model also replicates all but one of several other quantitative measures of the experimental results, including the power spectrum and correlation integral of the voltage, as well as the histogram, Poincaré plot, serial correlation coefficients, power spectrum, detrended fluctuation analysis, approximate entropy, and sample entropy of IBI. The channel noise from one particular ionic current (IKs), which has channel kinetics that are relatively slow compared to that of the other currents, makes the major contribution to the fluctuations in IBI. Reproduction of the experimental coefficient of variation of IBI by adding a Gaussian white noise-current into the deterministic model necessitates using an unrealistically high noise-current amplitude. Indeed, a major implication of the modelling results is that, given the wide range of time-scales over which the various species of channels open and close, only a cell-specific stochastic model that is formulated taking into consideration the widely different ranges in the frequency content of the channel-noise produced by the opening and closing of several different types of channels will be able to reproduce precisely the various effects due to membrane noise seen in a particular electrophysiological preparation.

  11. Re-analysis of health and educational impacts of a school-based deworming programme in western Kenya: a statistical replication of a cluster quasi-randomized stepped-wedge trial

    PubMed Central

    Davey, Calum; Aiken, Alexander M; Hayes, Richard J; Hargreaves, James R

    2015-01-01

    Introduction: Helminth (worm) infections cause morbidity among poor communities worldwide. An influential study conducted in Kenya in 1998–99 reported that a school-based drug-and-educational intervention had benefits for worm infections and school attendance. Methods: In this statistical replication, we re-analysed data from this cluster quasi-randomized stepped-wedge trial, specifying two co-primary outcomes: school attendance and examination performance. We estimated intention-to-treat effects using year-stratified cluster-summary analysis and observation-level random-effects regression, and combined both years with a random-effects model accounting for year. The participants were not blinded to allocation status, and other interventions were concurrently conducted in a sub-set of schools. A protocol guiding outcome data collection was not available. Results: Quasi-randomization resulted in three similar groups of 25 schools. There was a substantial amount of missing data. In year-stratified cluster-summary analysis, there was no clear evidence for improvement in either school attendance or examination performance. In year-stratified regression models, there was some evidence of improvement in school attendance [adjusted odds ratios (aOR): year 1: 1.48, 95% confidence interval (CI) 0.88–2.52, P = 0.147; year 2: 1.23, 95% CI 1.01–1.51, P = 0.044], but not examination performance (adjusted differences: year 1: −0.135, 95% CI −0.323–0.054, P = 0.161; year 2: −0.017, 95% CI −0.201–0.166, P = 0.854). When both years were combined, there was strong evidence of an effect on attendance (aOR 1.82, 95% CI 1.74–1.91, P < 0.001), but not examination performance (adjusted difference −0.121, 95% CI −0.293–0.052, P = 0.169). Conclusions: The evidence supporting an improvement in school attendance differed by analysis method. This, and various other important limitations of the data, caution against over-interpretation of the results. We find that the study provides some evidence, but with high risk of bias, that a school-based drug-treatment and health-education intervention improved school attendance and no evidence of effect on examination performance. PMID:26203171

  12. Medium-induced change of the optical response of metal clusters in rare-gas matrices

    NASA Astrophysics Data System (ADS)

    Xuan, Fengyuan; Guet, Claude

    2017-10-01

    Interaction with the surrounding medium modifies the optical response of embedded metal clusters. For clusters from about ten to a few hundreds of silver atoms, embedded in rare-gas matrices, we study the environment effect within the matrix random phase approximation with exact exchange (RPAE) quantum approach, which has proved successful for free silver clusters. The polarizable surrounding medium screens the residual two-body RPAE interaction, adds a polarization term to the one-body potential, and shifts the vacuum energy of the active delocalized valence electrons. Within this model, we calculate the dipole oscillator strength distribution for Ag clusters embedded in helium droplets, neon, argon, krypton, and xenon matrices. The main contribution to the dipole surface plasmon red shift originates from the rare-gas polarization screening of the two-body interaction. The large size limit of the dipole surface plasmon agrees well with the classical prediction.

  13. Pattern Selection and Super-Patterns in Opinion Dynamics

    NASA Astrophysics Data System (ADS)

    Ben-Naim, Eli; Scheel, Arnd

    We study pattern formation in the bounded confidence model of opinion dynamics. In this random process, opinion is quantified by a single variable. Two agents may interact and reach a fair compromise, but only if their difference of opinion falls below a fixed threshold. Starting from a uniform distribution of opinions with compact support, a traveling wave forms and it propagates from the domain boundary into the unstable uniform state. Consequently, the system reaches a steady state with isolated clusters that are separated by distance larger than the interaction range. These clusters form a quasi-periodic pattern where the sizes of the clusters and the separations between them are nearly constant. We obtain analytically the average separation between clusters L. Interestingly, there are also very small quasi-periodic modulations in the size of the clusters. The spatial periods of these modulations are a series of integers that follow from the continued-fraction representation of the irrational average separation L.

  14. The effect of clustering on perceived quantity in humans (Homo sapiens) and in chicks (Gallus gallus).

    PubMed

    Bertamini, Marco; Guest, Martin; Vallortigara, Giorgio; Rugani, Rosa; Regolin, Lucia

    2018-04-30

    Animals can perceive the numerosity of sets of visual elements. Qualitative and quantitative similarities in different species suggest the existence of a shared system (approximate number system). Biases associated with sensory properties are informative about the underlying mechanisms. In humans, regular spacing increases perceived numerosity (regular-random numerosity illusion). This has led to a model that predicts numerosity based on occupancy (a measure that decreases when elements are close together). We used a procedure in which observers selected one of two stimuli and were given feedback with respect to whether the choice was correct. One configuration had 20 elements and the other 40, randomly placed inside a circular region. Participants had to discover the rule based on feedback. Because density and clustering covaried with numerosity, different dimensions could be used. After reaching a criterion, test trials presented two types of configurations with 30 elements. One type had a larger interelement distance than the other (high or low clustering). If observers had adopted a numerosity strategy, they would choose low clustering (if reinforced with 40) and high clustering (if reinforced with 20). A clustering or density strategy predicts the opposite. Human adults used a numerosity strategy. Chicks were tested using a similar procedure. There were two behavioral measures: first approach response and final circumnavigation (walking behind the screen). The prediction based on numerosity was confirmed by the first approach data. For chicks, one clear pattern from both responses was a preference for the configurations with higher clustering. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  15. Sample size calculation in cost-effectiveness cluster randomized trials: optimal and maximin approaches.

    PubMed

    Manju, Md Abu; Candel, Math J J M; Berger, Martijn P F

    2014-07-10

    In this paper, the optimal sample sizes at the cluster and person levels for each of two treatment arms are obtained for cluster randomized trials where the cost-effectiveness of treatments on a continuous scale is studied. The optimal sample sizes maximize the efficiency or power for a given budget or minimize the budget for a given efficiency or power. Optimal sample sizes require information on the intra-cluster correlations (ICCs) for effects and costs, the correlations between costs and effects at individual and cluster levels, the ratio of the variance of effects translated into costs to the variance of the costs (the variance ratio), sampling and measuring costs, and the budget. When planning, a study information on the model parameters usually is not available. To overcome this local optimality problem, the current paper also presents maximin sample sizes. The maximin sample sizes turn out to be rather robust against misspecifying the correlation between costs and effects at the cluster and individual levels but may lose much efficiency when misspecifying the variance ratio. The robustness of the maximin sample sizes against misspecifying the ICCs depends on the variance ratio. The maximin sample sizes are robust under misspecification of the ICC for costs for realistic values of the variance ratio greater than one but not robust under misspecification of the ICC for effects. Finally, we show how to calculate optimal or maximin sample sizes that yield sufficient power for a test on the cost-effectiveness of an intervention.

  16. Clustered mixed nonhomogeneous Poisson process spline models for the analysis of recurrent event panel data.

    PubMed

    Nielsen, J D; Dean, C B

    2008-09-01

    A flexible semiparametric model for analyzing longitudinal panel count data arising from mixtures is presented. Panel count data refers here to count data on recurrent events collected as the number of events that have occurred within specific follow-up periods. The model assumes that the counts for each subject are generated by mixtures of nonhomogeneous Poisson processes with smooth intensity functions modeled with penalized splines. Time-dependent covariate effects are also incorporated into the process intensity using splines. Discrete mixtures of these nonhomogeneous Poisson process spline models extract functional information from underlying clusters representing hidden subpopulations. The motivating application is an experiment to test the effectiveness of pheromones in disrupting the mating pattern of the cherry bark tortrix moth. Mature moths arise from hidden, but distinct, subpopulations and monitoring the subpopulation responses was of interest. Within-cluster random effects are used to account for correlation structures and heterogeneity common to this type of data. An estimating equation approach to inference requiring only low moment assumptions is developed and the finite sample properties of the proposed estimating functions are investigated empirically by simulation.

  17. Direct construction of mesoscopic models from microscopic simulations

    NASA Astrophysics Data System (ADS)

    Lei, Huan; Caswell, Bruce; Karniadakis, George Em

    2010-02-01

    Starting from microscopic molecular-dynamics (MD) simulations of constrained Lennard-Jones (LJ) clusters (with constant radius of gyration Rg ), we construct two mesoscopic models [Langevin dynamics and dissipative particle dynamics (DPD)] by coarse graining the LJ clusters into single particles. Both static and dynamic properties of the coarse-grained models are investigated and compared with the MD results. The effective mean force field is computed as a function of the intercluster distance, and the corresponding potential scales linearly with the number of particles per cluster and the temperature. We verify that the mean force field can reproduce the equation of state of the atomistic systems within a wide density range but the radial distribution function only within the dilute and the semidilute regime. The friction force coefficients for both models are computed directly from the time-correlation function of the random force field of the microscopic system. For high density or a large cluster size the friction force is overestimated and the diffusivity underestimated due to the omission of many-body effects as a result of the assumed pairwise form of the coarse-grained force field. When the many-body effect is not as pronounced (e.g., smaller Rg or semidilute system), the DPD model can reproduce the dynamic properties of the MD system.

  18. Detecting treatment-subgroup interactions in clustered data with generalized linear mixed-effects model trees.

    PubMed

    Fokkema, M; Smits, N; Zeileis, A; Hothorn, T; Kelderman, H

    2017-10-25

    Identification of subgroups of patients for whom treatment A is more effective than treatment B, and vice versa, is of key importance to the development of personalized medicine. Tree-based algorithms are helpful tools for the detection of such interactions, but none of the available algorithms allow for taking into account clustered or nested dataset structures, which are particularly common in psychological research. Therefore, we propose the generalized linear mixed-effects model tree (GLMM tree) algorithm, which allows for the detection of treatment-subgroup interactions, while accounting for the clustered structure of a dataset. The algorithm uses model-based recursive partitioning to detect treatment-subgroup interactions, and a GLMM to estimate the random-effects parameters. In a simulation study, GLMM trees show higher accuracy in recovering treatment-subgroup interactions, higher predictive accuracy, and lower type II error rates than linear-model-based recursive partitioning and mixed-effects regression trees. Also, GLMM trees show somewhat higher predictive accuracy than linear mixed-effects models with pre-specified interaction effects, on average. We illustrate the application of GLMM trees on an individual patient-level data meta-analysis on treatments for depression. We conclude that GLMM trees are a promising exploratory tool for the detection of treatment-subgroup interactions in clustered datasets.

  19. Determining the Number of Clusters in a Data Set Without Graphical Interpretation

    NASA Technical Reports Server (NTRS)

    Aguirre, Nathan S.; Davies, Misty D.

    2011-01-01

    Cluster analysis is a data mining technique that is meant ot simplify the process of classifying data points. The basic clustering process requires an input of data points and the number of clusters wanted. The clustering algorithm will then pick starting C points for the clusters, which can be either random spatial points or random data points. It then assigns each data point to the nearest C point where "nearest usually means Euclidean distance, but some algorithms use another criterion. The next step is determining whether the clustering arrangement this found is within a certain tolerance. If it falls within this tolerance, the process ends. Otherwise the C points are adjusted based on how many data points are in each cluster, and the steps repeat until the algorithm converges,

  20. A Cluster-Randomized Trial of Restorative Practices: An Illustration to Spur High-Quality Research and Evaluation

    ERIC Educational Resources Information Center

    Acosta, Joie D.; Chinman, Matthew; Ebener, Patricia; Phillips, Andrea; Xenakis, Lea; Malone, Patrick S.

    2016-01-01

    Restorative practices in schools lack rigorous evaluation studies. As an example of rigorous school-based research, this article describes the first randomized control trial of restorative practices to date, the Study of Restorative Practices. It is a 5-year, cluster-randomized controlled trial (RCT) of the Restorative Practices Intervention (RPI)…

  1. Unsupervised classification of multivariate geostatistical data: Two algorithms

    NASA Astrophysics Data System (ADS)

    Romary, Thomas; Ors, Fabien; Rivoirard, Jacques; Deraisme, Jacques

    2015-12-01

    With the increasing development of remote sensing platforms and the evolution of sampling facilities in mining and oil industry, spatial datasets are becoming increasingly large, inform a growing number of variables and cover wider and wider areas. Therefore, it is often necessary to split the domain of study to account for radically different behaviors of the natural phenomenon over the domain and to simplify the subsequent modeling step. The definition of these areas can be seen as a problem of unsupervised classification, or clustering, where we try to divide the domain into homogeneous domains with respect to the values taken by the variables in hand. The application of classical clustering methods, designed for independent observations, does not ensure the spatial coherence of the resulting classes. Image segmentation methods, based on e.g. Markov random fields, are not adapted to irregularly sampled data. Other existing approaches, based on mixtures of Gaussian random functions estimated via the expectation-maximization algorithm, are limited to reasonable sample sizes and a small number of variables. In this work, we propose two algorithms based on adaptations of classical algorithms to multivariate geostatistical data. Both algorithms are model free and can handle large volumes of multivariate, irregularly spaced data. The first one proceeds by agglomerative hierarchical clustering. The spatial coherence is ensured by a proximity condition imposed for two clusters to merge. This proximity condition relies on a graph organizing the data in the coordinates space. The hierarchical algorithm can then be seen as a graph-partitioning algorithm. Following this interpretation, a spatial version of the spectral clustering algorithm is also proposed. The performances of both algorithms are assessed on toy examples and a mining dataset.

  2. Cluster formation of network-modifier cations in cesium silicate glasses

    NASA Astrophysics Data System (ADS)

    Jardón-Álvarez, Daniel; Sanders, Kevin J.; Phyo, Pyae; Baltisberger, Jay H.; Grandinetti, Philip J.

    2018-03-01

    Natural abundance 29Si two-dimensional magic-angle flipping (2D MAF) NMR spectra were measured in a series of ten cesium silicate glass compositions xCs2O.(1 - x)SiO2, where x is 0.067, 0.113, 0.175, 0.179, 0.218, 0.234, 0.263, 0.298, 0.31, and 0.36. The Q3 shielding anisotropy decreases with increasing Cs content—interpreted as an increase in the non-bridging oxygen (NBO) bond length from increasing Cs coordination (clustering) around the NBO. The 29Si 2D MAF spectra for four glass compositions x = 0.218, 0.234, 0.263, 0.298 exhibit a second co-existing and distinctly smaller shielding anisotropy corresponding to a significantly longer Si-NBO length arising from a higher degree of Cs clustering around the NBO. This second Q3 site appears at a Cs2O mole fraction close to the critical mole fraction of x = 0.24 associated with the percolation threshold of non-bridging oxygen in random close packing of oxygen, thus suggesting that the longer Si-NBO length is associated with an infinite size spanning cluster while the sites with larger anisotropies are associated with shorter Si-NBO lengths and belong to finite size clusters. The equilibrium constant of the Q3 disproportionation reaction was determined as k3 = 0.005, indicating a Qn anionic species distribution close to a binary model as expected for a low field strength modifier such as cesium. It is also found that evolution of the isotropic Q4 and line shapes with increasing Cs content are consistent with a random connectivity model between Qn of differing number of bridging oxygen, n.

  3. The median hazard ratio: a useful measure of variance and general contextual effects in multilevel survival analysis

    PubMed Central

    Wagner, Philippe; Merlo, Juan

    2016-01-01

    Multilevel data occurs frequently in many research areas like health services research and epidemiology. A suitable way to analyze such data is through the use of multilevel regression models (MLRM). MLRM incorporate cluster‐specific random effects which allow one to partition the total individual variance into between‐cluster variation and between‐individual variation. Statistically, MLRM account for the dependency of the data within clusters and provide correct estimates of uncertainty around regression coefficients. Substantively, the magnitude of the effect of clustering provides a measure of the General Contextual Effect (GCE). When outcomes are binary, the GCE can also be quantified by measures of heterogeneity like the Median Odds Ratio (MOR) calculated from a multilevel logistic regression model. Time‐to‐event outcomes within a multilevel structure occur commonly in epidemiological and medical research. However, the Median Hazard Ratio (MHR) that corresponds to the MOR in multilevel (i.e., ‘frailty’) Cox proportional hazards regression is rarely used. Analogously to the MOR, the MHR is the median relative change in the hazard of the occurrence of the outcome when comparing identical subjects from two randomly selected different clusters that are ordered by risk. We illustrate the application and interpretation of the MHR in a case study analyzing the hazard of mortality in patients hospitalized for acute myocardial infarction at hospitals in Ontario, Canada. We provide R code for computing the MHR. The MHR is a useful and intuitive measure for expressing cluster heterogeneity in the outcome and, thereby, estimating general contextual effects in multilevel survival analysis. © 2016 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd. PMID:27885709

  4. Decentralized cooperative TOA/AOA target tracking for hierarchical wireless sensor networks.

    PubMed

    Chen, Ying-Chih; Wen, Chih-Yu

    2012-11-08

    This paper proposes a distributed method for cooperative target tracking in hierarchical wireless sensor networks. The concept of leader-based information processing is conducted to achieve object positioning, considering a cluster-based network topology. Random timers and local information are applied to adaptively select a sub-cluster for the localization task. The proposed energy-efficient tracking algorithm allows each sub-cluster member to locally estimate the target position with a Bayesian filtering framework and a neural networking model, and further performs estimation fusion in the leader node with the covariance intersection algorithm. This paper evaluates the merits and trade-offs of the protocol design towards developing more efficient and practical algorithms for object position estimation.

  5. A cluster randomized trial of alcohol prevention in small businesses: a cascade model of help seeking and risk reduction.

    PubMed

    Reynolds, G Shawn; Bennett, Joel B

    2015-01-01

    The current study adapted two workplace substance abuse prevention programs and tested a conceptual model of workplace training effects on help seeking and alcohol consumption. Questionnaires were collected 1 month before, 1 month after, and 6 months within a cluster randomized field experiment. Texas small businesses in construction, transportation, and service industries. A total of 1510 employees from 45 businesses were randomly assigned to receive no training or one of the interventions. The interventions were 4-hour on-the-job classroom trainings that encouraged healthy lifestyles and seeking professional help (e.g., from the Employee Assistance Program [EAP]). The Team Awareness Program focused on peer referral and team building. The Choices in Health Promotion Program delivered various health topics based on a needs assessment. Questionnaires measured help-seeking attitudes and behavior, frequency of drinking alcohol, and job-related incidents. Mixed-model repeated-measures analyses of covariance were computed. Relative to the control group, training was associated with significantly greater reductions in drinking frequency, willingness to seek help, and seeking help from the EAP. After including help-seeking attitudes as a covariate, the correlation between training and help seeking becomes nonsignificant. Help-seeking behavior was not correlated with drinking frequency. Training improved help-seeking attitudes and behaviors and decreased alcohol risks. The reductions in drinking alcohol were directly correlated with training and independent from help seeking.

  6. Intraclass Correlation Coefficients for Obesity Indicators and Energy Balance-Related Behaviors Among New York City Public Elementary Schools.

    PubMed

    Gray, Heewon Lee; Burgermaster, Marissa; Tipton, Elizabeth; Contento, Isobel R; Koch, Pamela A; Di Noia, Jennifer

    2016-04-01

    Sample size and statistical power calculation should consider clustering effects when schools are the unit of randomization in intervention studies. The objective of the current study was to investigate how student outcomes are clustered within schools in an obesity prevention trial. Baseline data from the Food, Health & Choices project were used. Participants were 9- to 13-year-old students enrolled in 20 New York City public schools (n= 1,387). Body mass index (BMI) was calculated based on measures of height and weight, and body fat percentage was measured with a Tanita® body composition analyzer (Model SC-331s). Energy balance-related behaviors were self-reported with a frequency questionnaire. To examine the cluster effects, intraclass correlation coefficients (ICCs) were calculated as school variance over total variance for outcome variables. School-level covariates, percentage students eligible for free and reduced-price lunch, percentage Black or Hispanic, and English language learners were added in the model to examine ICC changes. The ICCs for obesity indicators are: .026 for BMI-percentile, .031 for BMIz-score, .035 for percentage of overweight students, .037 for body fat percentage, and .041 for absolute BMI. The ICC range for the six energy balance-related behaviors are .008 to .044 for fruit and vegetables, .013 to .055 for physical activity, .031 to .052 for recreational screen time, .013 to .091 for sweetened beverages, .033 to .121 for processed packaged snacks, and .020 to .083 for fast food. When school-level covariates were included in the model, ICC changes varied from -95% to 85%. This is the first study reporting ICCs for obesity-related anthropometric and behavioral outcomes among New York City public schools. The results of the study may aid sample size estimation for future school-based cluster randomized controlled trials in similar urban setting and population. Additionally, identifying school-level covariates that can reduce cluster effects is important when analyzing data. © 2015 Society for Public Health Education.

  7. Bohman-Frieze-Wormald model on the lattice, yielding a discontinuous percolation transition

    NASA Astrophysics Data System (ADS)

    Schrenk, K. J.; Felder, A.; Deflorin, S.; Araújo, N. A. M.; D'Souza, R. M.; Herrmann, H. J.

    2012-03-01

    The BFW model introduced by Bohman, Frieze, and Wormald [Random Struct. Algorithms1042-983210.1002/rsa.20038, 25, 432 (2004)], and recently investigated in the framework of discontinuous percolation by Chen and D'Souza [Phys. Rev. Lett.PRLTAO0031-900710.1103/PhysRevLett.106.115701 106, 115701 (2011)], is studied on the square and simple-cubic lattices. In two and three dimensions, we find numerical evidence for a strongly discontinuous transition. In two dimensions, the clusters at the threshold are compact with a fractal surface of fractal dimension df=1.49±0.02. On the simple-cubic lattice, distinct jumps in the size of the largest cluster are observed. We proceed to analyze the tree-like version of the model, where only merging bonds are sampled, for dimension two to seven. The transition is again discontinuous in any considered dimension. Finally, the dependence of the cluster-size distribution at the threshold on the spatial dimension is also investigated.

  8. Assessing the feasibility of interrupting the transmission of soil-transmitted helminths through mass drug administration: The DeWorm3 cluster randomized trial protocol.

    PubMed

    Ásbjörnsdóttir, Kristjana Hrönn; Ajjampur, Sitara S Rao; Anderson, Roy M; Bailey, Robin; Gardiner, Iain; Halliday, Katherine E; Ibikounle, Moudachirou; Kalua, Khumbo; Kang, Gagandeep; Littlewood, D Timothy J; Luty, Adrian J F; Means, Arianna Rubin; Oswald, William; Pullan, Rachel L; Sarkar, Rajiv; Schär, Fabian; Szpiro, Adam; Truscott, James E; Werkman, Marleen; Yard, Elodie; Walson, Judd L

    2018-01-01

    Current control strategies for soil-transmitted helminths (STH) emphasize morbidity control through mass drug administration (MDA) targeting preschool- and school-age children, women of childbearing age and adults in certain high-risk occupations such as agricultural laborers or miners. This strategy is effective at reducing morbidity in those treated but, without massive economic development, it is unlikely it will interrupt transmission. MDA will therefore need to continue indefinitely to maintain benefit. Mathematical models suggest that transmission interruption may be achievable through MDA alone, provided that all age groups are targeted with high coverage. The DeWorm3 Project will test the feasibility of interrupting STH transmission using biannual MDA targeting all age groups. Study sites (population ≥80,000) have been identified in Benin, Malawi and India. Each site will be divided into 40 clusters, to be randomized 1:1 to three years of twice-annual community-wide MDA or standard-of-care MDA, typically annual school-based deworming. Community-wide MDA will be delivered door-to-door, while standard-of-care MDA will be delivered according to national guidelines. The primary outcome is transmission interruption of the STH species present at each site, defined as weighted cluster-level prevalence ≤2% by quantitative polymerase chain reaction (qPCR), 24 months after the final round of MDA. Secondary outcomes include the endline prevalence of STH, overall and by species, and the endline prevalence of STH among children under five as an indicator of incident infections. Secondary analyses will identify cluster-level factors associated with transmission interruption. Prevalence will be assessed using qPCR of stool samples collected from a random sample of cluster residents at baseline, six months after the final round of MDA and 24 months post-MDA. A smaller number of individuals in each cluster will be followed with annual sampling to monitor trends in prevalence and reinfection throughout the trial. ClinicalTrials.gov NCT03014167.

  9. Shapiro effect as a possible cause of the low-frequency pulsar timing noise in globular clusters

    NASA Astrophysics Data System (ADS)

    Larchenkova, T. I.; Kopeikin, S. M.

    2006-01-01

    A prolonged timing of millisecond pulsars has revealed low-frequency uncorrelated (infrared) noise, presumably of astrophysical origin, in the pulse arrival time (PAT) residuals for some of them. Currently available pulsar timing methods allow the statistical parameters of this noise to be reliably measured by decomposing the PAT residual function into orthogonal Fourier harmonics. In most cases, pulsars in globular clusters show a low-frequency modulation of their rotational phase and spin rate. The relativistic time delay of the pulsar signal in the curved spacetime of randomly distributed and moving globular cluster stars (the Shapiro effect) is suggested as a possible cause of this modulation. Extremely important (from an astrophysical point of view) information about the structure of the globular cluster core, which is inaccessible to study by other observational methods, could be obtained by analyzing the spectral parameters of the low-frequency noise caused by the Shapiro effect and attributable to the random passages of stars near the line of sight to the pulsar. Given the smallness of the aberration corrections that arise from the nonstationarity of the gravitational field of the randomly distributed ensemble of stars under consideration, a formula is derived for the Shapiro effect for a pulsar in a globular cluster. The derived formula is used to calculate the autocorrelation function of the low-frequency pulsar noise, the slope of its power spectrum, and the behavior of the σz statistic that characterizes the spectral properties of this noise in the form of a time function. The Shapiro effect under discussion is shown to manifest itself for large impact parameters as a low-frequency noise of the pulsar spin rate with a spectral index of n = -1.8 that depends weakly on the specific model distribution of stars in the globular cluster. For small impact parameters, the spectral index of the noise is n = -1.5.

  10. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ben-Naim, Eli; Krapivsky, Paul

    Here we generalize the ordinary aggregation process to allow for choice. In ordinary aggregation, two random clusters merge and form a larger aggregate. In our implementation of choice, a target cluster and two candidate clusters are randomly selected and the target cluster merges with the larger of the two candidate clusters.We study the long-time asymptotic behavior and find that as in ordinary aggregation, the size density adheres to the standard scaling form. However, aggregation with choice exhibits a number of different features. First, the density of the smallest clusters exhibits anomalous scaling. Second, both the small-size and the large-size tailsmore » of the density are overpopulated, at the expense of the density of moderate-size clusters. Finally, we also study the complementary case where the smaller candidate cluster participates in the aggregation process and find an abundance of moderate clusters at the expense of small and large clusters. Additionally, we investigate aggregation processes with choice among multiple candidate clusters and a symmetric implementation where the choice is between two pairs of clusters.« less

  11. NASA Radiation Track Image GUI for Assessing Space Radiation Biological Effects

    NASA Technical Reports Server (NTRS)

    Ponomarev, Artem L.; Cucinotta, Francis A.

    2006-01-01

    The high-charge high-energy (HZE) ion components of the galactic cosmic rays when compared to terrestrial forms of radiations present unique challenges to biological systems. In this paper we present a deoxyribonucleic acid (DNA) breakage model to visualize and analyze the impact of chromatin domains and DNA loops on clustering of DNA damage from X rays, protons, and HZE ions. Our model of DNA breakage is based on a stochastic process of DNA double-strand break (DSB) formulation that includes the amorphous model of the radiation track and a polymer model of DNA packed in the cell nucleus. Our model is a Monte-Carlo simulation based on a randomly located DSB cluster formulation that accomodates both high- and low-linear energy transfer radiations. We demonstrate that HZE ions have a strong impact on DSB clustering, both along the chromosome length and in the nucleus volume. The effects of chromosomal domains and DNA loops on the DSB fragment-size distribution and the spatial distribution of DSB in the nucleus were studied. We compare our model predictions with the spatial distribution of DSB obtained from experiments. The implications of our model predictions for radiation protection are discussed.

  12. Functional Principal Component Analysis and Randomized Sparse Clustering Algorithm for Medical Image Analysis

    PubMed Central

    Lin, Nan; Jiang, Junhai; Guo, Shicheng; Xiong, Momiao

    2015-01-01

    Due to the advancement in sensor technology, the growing large medical image data have the ability to visualize the anatomical changes in biological tissues. As a consequence, the medical images have the potential to enhance the diagnosis of disease, the prediction of clinical outcomes and the characterization of disease progression. But in the meantime, the growing data dimensions pose great methodological and computational challenges for the representation and selection of features in image cluster analysis. To address these challenges, we first extend the functional principal component analysis (FPCA) from one dimension to two dimensions to fully capture the space variation of image the signals. The image signals contain a large number of redundant features which provide no additional information for clustering analysis. The widely used methods for removing the irrelevant features are sparse clustering algorithms using a lasso-type penalty to select the features. However, the accuracy of clustering using a lasso-type penalty depends on the selection of the penalty parameters and the threshold value. In practice, they are difficult to determine. Recently, randomized algorithms have received a great deal of attentions in big data analysis. This paper presents a randomized algorithm for accurate feature selection in image clustering analysis. The proposed method is applied to both the liver and kidney cancer histology image data from the TCGA database. The results demonstrate that the randomized feature selection method coupled with functional principal component analysis substantially outperforms the current sparse clustering algorithms in image cluster analysis. PMID:26196383

  13. Assessment of economic status in trauma registries: A new algorithm for generating population-specific clustering-based models of economic status for time-constrained low-resource settings.

    PubMed

    Eyler, Lauren; Hubbard, Alan; Juillard, Catherine

    2016-10-01

    Low and middle-income countries (LMICs) and the world's poor bear a disproportionate share of the global burden of injury. Data regarding disparities in injury are vital to inform injury prevention and trauma systems strengthening interventions targeted towards vulnerable populations, but are limited in LMICs. We aim to facilitate injury disparities research by generating a standardized methodology for assessing economic status in resource-limited country trauma registries where complex metrics such as income, expenditures, and wealth index are infeasible to assess. To address this need, we developed a cluster analysis-based algorithm for generating simple population-specific metrics of economic status using nationally representative Demographic and Health Surveys (DHS) household assets data. For a limited number of variables, g, our algorithm performs weighted k-medoids clustering of the population using all combinations of g asset variables and selects the combination of variables and number of clusters that maximize average silhouette width (ASW). In simulated datasets containing both randomly distributed variables and "true" population clusters defined by correlated categorical variables, the algorithm selected the correct variable combination and appropriate cluster numbers unless variable correlation was very weak. When used with 2011 Cameroonian DHS data, our algorithm identified twenty economic clusters with ASW 0.80, indicating well-defined population clusters. This economic model for assessing health disparities will be used in the new Cameroonian six-hospital centralized trauma registry. By describing our standardized methodology and algorithm for generating economic clustering models, we aim to facilitate measurement of health disparities in other trauma registries in resource-limited countries. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  14. Mindfulness-Based Stress Reduction in Post-treatment Breast Cancer Patients: Immediate and Sustained Effects Across Multiple Symptom Clusters.

    PubMed

    Reich, Richard R; Lengacher, Cecile A; Alinat, Carissa B; Kip, Kevin E; Paterson, Carly; Ramesar, Sophia; Han, Heather S; Ismail-Khan, Roohi; Johnson-Mallard, Versie; Moscoso, Manolete; Budhrani-Shani, Pinky; Shivers, Steve; Cox, Charles E; Goodman, Matthew; Park, Jong

    2017-01-01

    Breast cancer survivors (BCS) face adverse physical and psychological symptoms, often co-occurring. Biologic and psychological factors may link symptoms within clusters, distinguishable by prevalence and/or severity. Few studies have examined the effects of behavioral interventions or treatment of symptom clusters. The aim of this study was to identify symptom clusters among post-treatment BCS and determine symptom cluster improvement following the Mindfulness-Based Stress Reduction for Breast Cancer (MBSR(BC)) program. Three hundred twenty-two Stage 0-III post-treatment BCS were randomly assigned to either a six-week MBSR(BC) program or usual care. Psychological (depression, anxiety, stress, and fear of recurrence), physical (fatigue, pain, sleep, and drowsiness), and cognitive symptoms and quality of life were assessed at baseline, six, and 12 weeks, along with demographic and clinical history data at baseline. A three-step analytic process included the error-accounting models of factor analysis and structural equation modeling. Four symptom clusters emerged at baseline: pain, psychological, fatigue, and cognitive. From baseline to six weeks, the model demonstrated evidence of MBSR(BC) effectiveness in both the psychological (anxiety, depression, perceived stress and QOL, emotional well-being) (P = 0.007) and fatigue (fatigue, sleep, and drowsiness) (P < 0.001) clusters. Results between six and 12 weeks showed sustained effects, but further improvement was not observed. Our results provide clinical effectiveness evidence that MBSR(BC) works to improve symptom clusters, particularly for psychological and fatigue symptom clusters, with the greatest improvement occurring during the six-week program with sustained effects for several weeks after MBSR(BC) training. Name and URL of Registry: ClinicalTrials.gov. Registration number: NCT01177124. Copyright © 2016. Published by Elsevier Inc.

  15. Effects of Blended Instructional Models on Math Performance

    ERIC Educational Resources Information Center

    Bottge, Brian A.; Ma, Xin; Gassaway, Linda; Toland, Michael D.; Butler, Mark; Cho, Sun-Joo

    2014-01-01

    A pretest-posttest cluster-randomized trial involving 31 middle schools and 335 students with disabilities tested the effects of combining explicit and anchored instruction on fraction computation and problem solving. Results of standardized and researcher-developed tests showed that students who were taught with the blended units outscored…

  16. Image texture segmentation using a neural network

    NASA Astrophysics Data System (ADS)

    Sayeh, Mohammed R.; Athinarayanan, Ragu; Dhali, Pushpuak

    1992-09-01

    In this paper we use a neural network called the Lyapunov associative memory (LYAM) system to segment image texture into different categories or clusters. The LYAM system is constructed by a set of ordinary differential equations which are simulated on a digital computer. The clustering can be achieved by using a single tuning parameter in the simplest model. Pattern classes are represented by the stable equilibrium states of the system. Design of the system is based on synthesizing two local energy functions, namely, the learning and recall energy functions. Before the implementation of the segmentation process, a Gauss-Markov random field (GMRF) model is applied to the raw image. This application suitably reduces the image data and prepares the texture information for the neural network process. We give a simple image example illustrating the capability of the technique. The GMRF-generated features are also used for a clustering, based on the Euclidean distance.

  17. Interaction of multiple biomimetic antimicrobial polymers with model bacterial membranes

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Baul, Upayan, E-mail: upayanb@imsc.res.in; Vemparala, Satyavani, E-mail: vani@imsc.res.in; Kuroda, Kenichi, E-mail: kkuroda@umich.edu

    Using atomistic molecular dynamics simulations, interaction of multiple synthetic random copolymers based on methacrylates on prototypical bacterial membranes is investigated. The simulations show that the cationic polymers form a micellar aggregate in water phase and the aggregate, when interacting with the bacterial membrane, induces clustering of oppositely charged anionic lipid molecules to form clusters and enhances ordering of lipid chains. The model bacterial membrane, consequently, develops lateral inhomogeneity in membrane thickness profile compared to polymer-free system. The individual polymers in the aggregate are released into the bacterial membrane in a phased manner and the simulations suggest that the most probablemore » location of the partitioned polymers is near the 1-palmitoyl-2-oleoyl-phosphatidylglycerol (POPG) clusters. The partitioned polymers preferentially adopt facially amphiphilic conformations at lipid-water interface, despite lacking intrinsic secondary structures such as α-helix or β-sheet found in naturally occurring antimicrobial peptides.« less

  18. Stochastic competitive learning in complex networks.

    PubMed

    Silva, Thiago Christiano; Zhao, Liang

    2012-03-01

    Competitive learning is an important machine learning approach which is widely employed in artificial neural networks. In this paper, we present a rigorous definition of a new type of competitive learning scheme realized on large-scale networks. The model consists of several particles walking within the network and competing with each other to occupy as many nodes as possible, while attempting to reject intruder particles. The particle's walking rule is composed of a stochastic combination of random and preferential movements. The model has been applied to solve community detection and data clustering problems. Computer simulations reveal that the proposed technique presents high precision of community and cluster detections, as well as low computational complexity. Moreover, we have developed an efficient method for estimating the most likely number of clusters by using an evaluator index that monitors the information generated by the competition process itself. We hope this paper will provide an alternative way to the study of competitive learning..

  19. Dissociation kinetics of metal clusters on multiple electronic states including electronic level statistics into the vibronic soup

    NASA Astrophysics Data System (ADS)

    Shvartsburg, Alexandre A.; Siu, K. W. Michael

    2001-06-01

    Modeling the delayed dissociation of clusters had been over the last decade a frontline development area in chemical physics. It is of fundamental interest how statistical kinetics methods previously validated for regular molecules and atomic nuclei may apply to clusters, as this would help to understand the transferability of statistical models for disintegration of complex systems across various classes of physical objects. From a practical perspective, accurate simulation of unimolecular decomposition is critical for the extraction of true thermochemical values from measurements on the decay of energized clusters. Metal clusters are particularly challenging because of the multitude of low-lying electronic states that are coupled to vibrations. This has previously been accounted for assuming the average electronic structure of a conducting cluster approximated by the levels of electron in a cavity. While this provides a reasonable time-averaged description, it ignores the distribution of instantaneous electronic structures in a "boiling" cluster around that average. Here we set up a new treatment that incorporates the statistical distribution of electronic levels around the average picture using random matrix theory. This approach faithfully reflects the completely chaotic "vibronic soup" nature of hot metal clusters. We found that the consideration of electronic level statistics significantly promotes electronic excitation and thus increases the magnitude of its effect. As this excitation always depresses the decay rates, the inclusion of level statistics results in slower dissociation of metal clusters.

  20. Novel schemes for measurement-based quantum computation.

    PubMed

    Gross, D; Eisert, J

    2007-06-01

    We establish a framework which allows one to construct novel schemes for measurement-based quantum computation. The technique develops tools from many-body physics-based on finitely correlated or projected entangled pair states-to go beyond the cluster-state based one-way computer. We identify resource states radically different from the cluster state, in that they exhibit nonvanishing correlations, can be prepared using nonmaximally entangling gates, or have very different local entanglement properties. In the computational models, randomness is compensated in a different manner. It is shown that there exist resource states which are locally arbitrarily close to a pure state. We comment on the possibility of tailoring computational models to specific physical systems.

  1. A Cluster-Randomized Trial of Insecticide-Treated Curtains for Dengue Vector Control in Thailand

    PubMed Central

    Lenhart, Audrey; Trongtokit, Yuwadee; Alexander, Neal; Apiwathnasorn, Chamnarn; Satimai, Wichai; Vanlerberghe, Veerle; Van der Stuyft, Patrick; McCall, Philip J.

    2013-01-01

    The efficacy of insecticide-treated window curtains (ITCs) for dengue vector control was evaluated in Thailand in a cluster-randomized controlled trial. A total of 2,037 houses in 26 clusters was randomized to receive the intervention or act as control (no treatment). Entomological surveys measured Aedes infestations (Breteau index, house index, container index, and pupae per person index) and oviposition indices (mean numbers of eggs laid in oviposition traps) immediately before and after intervention, and at 3-month intervals over 12 months. There were no consistent statistically significant differences in entomological indices between intervention and control clusters, although oviposition indices were lower (P < 0.01) in ITC clusters during the wet season. It is possible that the open housing structures in the study reduced the likelihood of mosquitoes making contact with ITCs. ITCs deployed in a region where this house design is common may be unsuitable for dengue vector control. PMID:23166195

  2. A Cluster Randomized Controlled Trial Testing the Effectiveness of Houvast: A Strengths-Based Intervention for Homeless Young Adults

    ERIC Educational Resources Information Center

    Krabbenborg, Manon A. M.; Boersma, Sandra N.; van der Veld, William M.; van Hulst, Bente; Vollebergh, Wilma A. M.; Wolf, Judith R. L. M.

    2017-01-01

    Objective: To test the effectiveness of Houvast: a strengths-based intervention for homeless young adults. Method: A cluster randomized controlled trial was conducted with 10 Dutch shelter facilities randomly allocated to an intervention and a control group. Homeless young adults were interviewed when entering the facility and when care ended.…

  3. Extension of mixture-of-experts networks for binary classification of hierarchical data.

    PubMed

    Ng, Shu-Kay; McLachlan, Geoffrey J

    2007-09-01

    For many applied problems in the context of medically relevant artificial intelligence, the data collected exhibit a hierarchical or clustered structure. Ignoring the interdependence between hierarchical data can result in misleading classification. In this paper, we extend the mechanism for mixture-of-experts (ME) networks for binary classification of hierarchical data. Another extension is to quantify cluster-specific information on data hierarchy by random effects via the generalized linear mixed-effects model (GLMM). The extension of ME networks is implemented by allowing for correlation in the hierarchical data in both the gating and expert networks via the GLMM. The proposed model is illustrated using a real thyroid disease data set. In our study, we consider 7652 thyroid diagnosis records from 1984 to early 1987 with complete information on 20 attribute values. We obtain 10 independent random splits of the data into a training set and a test set in the proportions 85% and 15%. The test sets are used to assess the generalization performance of the proposed model, based on the percentage of misclassifications. For comparison, the results obtained from the ME network with independence assumption are also included. With the thyroid disease data, the misclassification rate on test sets for the extended ME network is 8.9%, compared to 13.9% for the ME network. In addition, based on model selection methods described in Section 2, a network with two experts is selected. These two expert networks can be considered as modeling two groups of patients with high and low incidence rates. Significant variation among the predicted cluster-specific random effects is detected in the patient group with low incidence rate. It is shown that the extended ME network outperforms the ME network for binary classification of hierarchical data. With the thyroid disease data, useful information on the relative log odds of patients with diagnosed conditions at different periods can be evaluated. This information can be taken into consideration for the assessment of treatment planning of the disease. The proposed extended ME network thus facilitates a more general approach to incorporate data hierarchy mechanism in network modeling.

  4. Central Limit Theorem for Exponentially Quasi-local Statistics of Spin Models on Cayley Graphs

    NASA Astrophysics Data System (ADS)

    Reddy, Tulasi Ram; Vadlamani, Sreekar; Yogeshwaran, D.

    2018-04-01

    Central limit theorems for linear statistics of lattice random fields (including spin models) are usually proven under suitable mixing conditions or quasi-associativity. Many interesting examples of spin models do not satisfy mixing conditions, and on the other hand, it does not seem easy to show central limit theorem for local statistics via quasi-associativity. In this work, we prove general central limit theorems for local statistics and exponentially quasi-local statistics of spin models on discrete Cayley graphs with polynomial growth. Further, we supplement these results by proving similar central limit theorems for random fields on discrete Cayley graphs taking values in a countable space, but under the stronger assumptions of α -mixing (for local statistics) and exponential α -mixing (for exponentially quasi-local statistics). All our central limit theorems assume a suitable variance lower bound like many others in the literature. We illustrate our general central limit theorem with specific examples of lattice spin models and statistics arising in computational topology, statistical physics and random networks. Examples of clustering spin models include quasi-associated spin models with fast decaying covariances like the off-critical Ising model, level sets of Gaussian random fields with fast decaying covariances like the massive Gaussian free field and determinantal point processes with fast decaying kernels. Examples of local statistics include intrinsic volumes, face counts, component counts of random cubical complexes while exponentially quasi-local statistics include nearest neighbour distances in spin models and Betti numbers of sub-critical random cubical complexes.

  5. Design of trials for interrupting the transmission of endemic pathogens.

    PubMed

    Silkey, Mariabeth; Homan, Tobias; Maire, Nicolas; Hiscox, Alexandra; Mukabana, Richard; Takken, Willem; Smith, Thomas A

    2016-06-06

    Many interventions against infectious diseases have geographically diffuse effects. This leads to contamination between arms in cluster-randomized trials (CRTs). Pathogen elimination is the goal of many intervention programs against infectious agents, but contamination means that standard CRT designs and analyses do not provide inferences about the potential of interventions to interrupt pathogen transmission at maximum scale-up. A generic model of disease transmission was used to simulate infections in stepped wedge cluster-randomized trials (SWCRTs) of a transmission-reducing intervention, where the intervention has spatially diffuse effects. Simulations of such trials were then used to examine the potential of such designs for providing generalizable causal inferences about the impact of such interventions, including measurements of the contamination effects. The simulations were applied to the geography of Rusinga Island, Lake Victoria, Kenya, the site of the SolarMal trial on the use of odor-baited mosquito traps to eliminate Plasmodium falciparum malaria. These were used to compare variants in the proposed SWCRT designs for the SolarMal trial. Measures of contamination effects were found that could be assessed in the simulated trials. Inspired by analyses of trials of insecticide-treated nets against malaria when applied to the geography of the SolarMal trial, these measures were found to be robust to different variants of SWCRT design. Analyses of the likely extent of contamination effects supported the choice of cluster size for the trial. The SWCRT is an appropriate design for trials that assess the feasibility of local elimination of a pathogen. The effects of incomplete coverage can be estimated by analyzing the extent of contamination between arms in such trials, and the estimates also support inferences about causality. The SolarMal example illustrates how generic transmission models incorporating spatial smoothing can be used to simulate such trials for a power calculation and optimization of cluster size and randomization strategies. The approach is applicable to a range of infectious diseases transmitted via environmental reservoirs or via arthropod vectors.

  6. The colour-magnitude relation as a constraint on the formation of rich cluster galaxies

    NASA Astrophysics Data System (ADS)

    Bower, Richard G.; Kodama, Tadayuki; Terlevich, Ale

    1998-10-01

    The colours and magnitudes of early-type galaxies in galaxy clusters are strongly correlated. The existence of such a correlation has been used to infer that early-type galaxies must be old passively evolving systems. Given the dominance of early-type galaxies in the cores of rich clusters, this view sits uncomfortably with the increasing fraction of blue galaxies found in clusters at intermediate redshifts, and with the late formation of galaxies favoured by cold dark matter type cosmologies. In this paper, we make a detailed investigation of these issues and examine the role that the colour-magnitude relation can play in constraining the formation history of galaxies currently found in the cores of rich clusters. We start by considering the colour evolution of galaxies after star formation ceases. We show that the scatter of the colour-magnitude relation places a strong constraint on the spread in age that is allowed for the bulk of the stellar population. In the extreme case that the stars are formed in a single event, the spread in age cannot be more than 4 Gyr. Although the bulk of stars must be formed in a short period, continuing formation of stars in a fraction of the galaxies is not so strongly constrained. We examine a model in which star formation occurs over an extended period of time in most galaxies with star formation being truncated randomly. This model is consistent with the formation of stars in a few systems until look-back times of ~5Gyr. An extension of this type of star formation history allows us to reconcile the small present-day scatter of the colour-magnitude relation with the observed blue galaxy fractions of intermediate redshift galaxy clusters. In addition to setting a limit on the variations in luminosity-weighted age between the stellar populations of cluster galaxies, the colour-magnitude relation can also be used to constrain the degree of merging between pre-existing stellar systems. This test relies on the slope of the colour-magnitude relation: mergers between galaxies of unequal mass tend to reduce the slope of the relation and to increase its scatter. We show that random mergers between galaxies very rapidly remove any well-defined colour-magnitude correlation. This model is not physically motivated, however, and we prefer to examine the merger process using a self-consistent merger tree. In such a model there are two effects. First, massive galaxies preferentially merge with systems of similar mass. Secondly, the rate of mass growth is considerably smaller than for the random merger case. As a result of both of these effects, the colour-magnitude correlation persists through a larger number of merger steps. The passive evolution of galaxy colours and their averaging in dissipationless mergers provide opposing constraints on the formation of cluster galaxies in a hierarchical model. At the level of current constraints, a compromise solution appears possible. The bulk of the stellar population must have formed before z=1, but cannot have formed in mass units much less than about half the mass of a present-day L_* galaxy. In this case, the galaxies are on average old enough that stellar population evolution is weak, yet formed recently enough that mass growth resulting from mergers is small.

  7. Internal Cluster Validation on Earthquake Data in the Province of Bengkulu

    NASA Astrophysics Data System (ADS)

    Rini, D. S.; Novianti, P.; Fransiska, H.

    2018-04-01

    K-means method is an algorithm for cluster n object based on attribute to k partition, where k < n. There is a deficiency of algorithms that is before the algorithm is executed, k points are initialized randomly so that the resulting data clustering can be different. If the random value for initialization is not good, the clustering becomes less optimum. Cluster validation is a technique to determine the optimum cluster without knowing prior information from data. There are two types of cluster validation, which are internal cluster validation and external cluster validation. This study aims to examine and apply some internal cluster validation, including the Calinski-Harabasz (CH) Index, Sillhouette (S) Index, Davies-Bouldin (DB) Index, Dunn Index (D), and S-Dbw Index on earthquake data in the Bengkulu Province. The calculation result of optimum cluster based on internal cluster validation is CH index, S index, and S-Dbw index yield k = 2, DB Index with k = 6 and Index D with k = 15. Optimum cluster (k = 6) based on DB Index gives good results for clustering earthquake in the Bengkulu Province.

  8. Predictive modeling of EEG time series for evaluating surgery targets in epilepsy patients.

    PubMed

    Steimer, Andreas; Müller, Michael; Schindler, Kaspar

    2017-05-01

    During the last 20 years, predictive modeling in epilepsy research has largely been concerned with the prediction of seizure events, whereas the inference of effective brain targets for resective surgery has received surprisingly little attention. In this exploratory pilot study, we describe a distributional clustering framework for the modeling of multivariate time series and use it to predict the effects of brain surgery in epilepsy patients. By analyzing the intracranial EEG, we demonstrate how patients who became seizure free after surgery are clearly distinguished from those who did not. More specifically, for 5 out of 7 patients who obtained seizure freedom (= Engel class I) our method predicts the specific collection of brain areas that got actually resected during surgery to yield a markedly lower posterior probability for the seizure related clusters, when compared to the resection of random or empty collections. Conversely, for 4 out of 5 Engel class III/IV patients who still suffer from postsurgical seizures, performance of the actually resected collection is not significantly better than performances displayed by random or empty collections. As the number of possible collections ranges into billions and more, this is a substantial contribution to a problem that today is still solved by visual EEG inspection. Apart from epilepsy research, our clustering methodology is also of general interest for the analysis of multivariate time series and as a generative model for temporally evolving functional networks in the neurosciences and beyond. Hum Brain Mapp 38:2509-2531, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  9. Mapping Health Data: Improved Privacy Protection With Donut Method Geomasking

    PubMed Central

    Hampton, Kristen H.; Fitch, Molly K.; Allshouse, William B.; Doherty, Irene A.; Gesink, Dionne C.; Leone, Peter A.; Serre, Marc L.; Miller, William C.

    2010-01-01

    A major challenge in mapping health data is protecting patient privacy while maintaining the spatial resolution necessary for spatial surveillance and outbreak identification. A new adaptive geomasking technique, referred to as the donut method, extends current methods of random displacement by ensuring a user-defined minimum level of geoprivacy. In donut method geomasking, each geocoded address is relocated in a random direction by at least a minimum distance, but less than a maximum distance. The authors compared the donut method with current methods of random perturbation and aggregation regarding measures of privacy protection and cluster detection performance by masking multiple disease field simulations under a range of parameters. Both the donut method and random perturbation performed better than aggregation in cluster detection measures. The performance of the donut method in geoprivacy measures was at least 42.7% higher and in cluster detection measures was less than 4.8% lower than that of random perturbation. Results show that the donut method provides a consistently higher level of privacy protection with a minimal decrease in cluster detection performance, especially in areas where the risk to individual geoprivacy is greatest. PMID:20817785

  10. Mapping health data: improved privacy protection with donut method geomasking.

    PubMed

    Hampton, Kristen H; Fitch, Molly K; Allshouse, William B; Doherty, Irene A; Gesink, Dionne C; Leone, Peter A; Serre, Marc L; Miller, William C

    2010-11-01

    A major challenge in mapping health data is protecting patient privacy while maintaining the spatial resolution necessary for spatial surveillance and outbreak identification. A new adaptive geomasking technique, referred to as the donut method, extends current methods of random displacement by ensuring a user-defined minimum level of geoprivacy. In donut method geomasking, each geocoded address is relocated in a random direction by at least a minimum distance, but less than a maximum distance. The authors compared the donut method with current methods of random perturbation and aggregation regarding measures of privacy protection and cluster detection performance by masking multiple disease field simulations under a range of parameters. Both the donut method and random perturbation performed better than aggregation in cluster detection measures. The performance of the donut method in geoprivacy measures was at least 42.7% higher and in cluster detection measures was less than 4.8% lower than that of random perturbation. Results show that the donut method provides a consistently higher level of privacy protection with a minimal decrease in cluster detection performance, especially in areas where the risk to individual geoprivacy is greatest.

  11. Does rational selection of training and test sets improve the outcome of QSAR modeling?

    PubMed

    Martin, Todd M; Harten, Paul; Young, Douglas M; Muratov, Eugene N; Golbraikh, Alexander; Zhu, Hao; Tropsha, Alexander

    2012-10-22

    Prior to using a quantitative structure activity relationship (QSAR) model for external predictions, its predictive power should be established and validated. In the absence of a true external data set, the best way to validate the predictive ability of a model is to perform its statistical external validation. In statistical external validation, the overall data set is divided into training and test sets. Commonly, this splitting is performed using random division. Rational splitting methods can divide data sets into training and test sets in an intelligent fashion. The purpose of this study was to determine whether rational division methods lead to more predictive models compared to random division. A special data splitting procedure was used to facilitate the comparison between random and rational division methods. For each toxicity end point, the overall data set was divided into a modeling set (80% of the overall set) and an external evaluation set (20% of the overall set) using random division. The modeling set was then subdivided into a training set (80% of the modeling set) and a test set (20% of the modeling set) using rational division methods and by using random division. The Kennard-Stone, minimal test set dissimilarity, and sphere exclusion algorithms were used as the rational division methods. The hierarchical clustering, random forest, and k-nearest neighbor (kNN) methods were used to develop QSAR models based on the training sets. For kNN QSAR, multiple training and test sets were generated, and multiple QSAR models were built. The results of this study indicate that models based on rational division methods generate better statistical results for the test sets than models based on random division, but the predictive power of both types of models are comparable.

  12. Radio Occultation Investigation of the Rings of Saturn and Uranus

    NASA Technical Reports Server (NTRS)

    Marouf, Essam A.

    1997-01-01

    The proposed work addresses two main objectives: (1) to pursue the development of the random diffraction screen model for analytical/computational characterization of the extinction and near-forward scattering by ring models that include particle crowding, uniform clustering, and clustering along preferred orientations (anisotropy). The characterization is crucial for proper interpretation of past (Voyager) and future (Cassini) ring, occultation observations in terms of physical ring properties, and is needed to address outstanding puzzles in the interpretation of the Voyager radio occultation data sets; (2) to continue the development of spectral analysis techniques to identify and characterize the power scattered by all features of Saturn's rings that can be resolved in the Voyager radio occultation observations, and to use the results to constrain the maximum particle size and its abundance. Characterization of the variability of surface mass density among the main ring, features and within individual features is important for constraining the ring mass and is relevant to investigations of ring dynamics and origin. We completed the developed of the stochastic geometry (random screen) model for the interaction of electromagnetic waves with of planetary ring models; used the model to relate the oblique optical depth and the angular spectrum of the near forward scattered signal to statistical averages of the stochastic geometry of the randomly blocked area. WE developed analytical results based on the assumption of Poisson statistics for particle positions, and investigated the dependence of the oblique optical depth and angular spectrum on the fractional area blocked, vertical ring profile, and incidence angle when the volume fraction is small. Demonstrated agreement with the classical radiative transfer predictions for oblique incidence. Also developed simulation procedures to generate statistical realizations of random screens corresponding to uniformly packed ring models, and used the results to characterize dependence of the extinction and near-forward scattering on ring thickness, packing fraction, and the ring opening angle.

  13. MIXED MODEL AND ESTIMATING EQUATION APPROACHES FOR ZERO INFLATION IN CLUSTERED BINARY RESPONSE DATA WITH APPLICATION TO A DATING VIOLENCE STUDY1

    PubMed Central

    Fulton, Kara A.; Liu, Danping; Haynie, Denise L.; Albert, Paul S.

    2016-01-01

    The NEXT Generation Health study investigates the dating violence of adolescents using a survey questionnaire. Each student is asked to affirm or deny multiple instances of violence in his/her dating relationship. There is, however, evidence suggesting that students not in a relationship responded to the survey, resulting in excessive zeros in the responses. This paper proposes likelihood-based and estimating equation approaches to analyze the zero-inflated clustered binary response data. We adopt a mixed model method to account for the cluster effect, and the model parameters are estimated using a maximum-likelihood (ML) approach that requires a Gaussian–Hermite quadrature (GHQ) approximation for implementation. Since an incorrect assumption on the random effects distribution may bias the results, we construct generalized estimating equations (GEE) that do not require the correct specification of within-cluster correlation. In a series of simulation studies, we examine the performance of ML and GEE methods in terms of their bias, efficiency and robustness. We illustrate the importance of properly accounting for this zero inflation by reanalyzing the NEXT data where this issue has previously been ignored. PMID:26937263

  14. Triadic split-merge sampler

    NASA Astrophysics Data System (ADS)

    van Rossum, Anne C.; Lin, Hai Xiang; Dubbeldam, Johan; van der Herik, H. Jaap

    2018-04-01

    In machine vision typical heuristic methods to extract parameterized objects out of raw data points are the Hough transform and RANSAC. Bayesian models carry the promise to optimally extract such parameterized objects given a correct definition of the model and the type of noise at hand. A category of solvers for Bayesian models are Markov chain Monte Carlo methods. Naive implementations of MCMC methods suffer from slow convergence in machine vision due to the complexity of the parameter space. Towards this blocked Gibbs and split-merge samplers have been developed that assign multiple data points to clusters at once. In this paper we introduce a new split-merge sampler, the triadic split-merge sampler, that perform steps between two and three randomly chosen clusters. This has two advantages. First, it reduces the asymmetry between the split and merge steps. Second, it is able to propose a new cluster that is composed out of data points from two different clusters. Both advantages speed up convergence which we demonstrate on a line extraction problem. We show that the triadic split-merge sampler outperforms the conventional split-merge sampler. Although this new MCMC sampler is demonstrated in this machine vision context, its application extend to the very general domain of statistical inference.

  15. Randomly diluted eg orbital-ordered systems.

    PubMed

    Tanaka, T; Matsumoto, M; Ishihara, S

    2005-12-31

    Dilution effects on the long-range ordered state of the doubly degenerate e(g) orbital are investigated. Quenched impurities without the orbital degree of freedom are introduced in the orbital model where the long-range order is realized by the order-from-disorder mechanism. It is shown by Monte Carlo simulations and the cluster-expansion method that a decrease in the orbital-ordering temperature by dilution is substantially larger than that in the randomly diluted spin models. Tilting of orbital pseudospins around impurities is the essence of this dilution effect. The present theory provides a new viewpoint for the recent resonant x-ray scattering experiments in KCu(1-x)Zn(x)F(3).

  16. Impact of a social-emotional and character development program on school-level indicators of academic achievement, absenteeism, and disciplinary outcomes: A matched-pair, cluster randomized, controlled trial.

    PubMed

    Snyder, Frank; Flay, Brian; Vuchinich, Samuel; Acock, Alan; Washburn, Isaac; Beets, Michael; Li, Kin-Kit

    2010-01-01

    This paper reports the effects of a comprehensive elementary school-based social-emotional and character education program on school-level achievement, absenteeism, and disciplinary outcomes utilizing a matched-pair, cluster randomized, controlled design. The Positive Action Hawai'i trial included 20 racially/ethnically diverse schools (mean enrollment = 544) and was conducted from the 2002-03 through the 2005-06 academic years. Using school-level archival data, analyses comparing change from baseline (2002) to one-year post trial (2007) revealed that intervention schools scored 9.8% better on the TerraNova (2 nd ed.) test for reading and 8.8% on math; 20.7% better in Hawai'i Content and Performance Standards scores for reading and 51.4% better in math; and that intervention schools reported 15.2% lower absenteeism and fewer suspensions (72.6%) and retentions (72.7%). Overall, effect sizes were moderate to large (range 0.5-1.1) for all of the examined outcomes. Sensitivity analyses using permutation models and random-intercept growth curve models substantiated results. The results provide evidence that a comprehensive school-based program, specifically developed to target student behavior and character, can positively influence school-level achievement, attendance, and disciplinary outcomes concurrently.

  17. Microscopic Spin Model for the STOCK Market with Attractor Bubbling on Regular and Small-World Lattices

    NASA Astrophysics Data System (ADS)

    Krawiecki, A.

    A multi-agent spin model for changes of prices in the stock market based on the Ising-like cellular automaton with interactions between traders randomly varying in time is investigated by means of Monte Carlo simulations. The structure of interactions has topology of a small-world network obtained from regular two-dimensional square lattices with various coordination numbers by randomly cutting and rewiring edges. Simulations of the model on regular lattices do not yield time series of logarithmic price returns with statistical properties comparable with the empirical ones. In contrast, in the case of networks with a certain degree of randomness for a wide range of parameters the time series of the logarithmic price returns exhibit intermittent bursting typical of volatility clustering. Also the tails of distributions of returns obey a power scaling law with exponents comparable to those obtained from the empirical data.

  18. Headache cessation by an educational intervention in grammar schools: a cluster randomized trial.

    PubMed

    Albers, L; Heinen, F; Landgraf, M; Straube, A; Blum, B; Filippopulos, F; Lehmann, S; Mansmann, U; Berger, U; Akboga, Y; von Kries, R

    2015-02-01

    Headache is a common health problem in adolescents. There are a number of risk factors for headache in adolescents that are amenable to intervention. The aim of the study was to assess the effectiveness of a low-level headache prevention programme in the classroom setting to prevent these risk factors. In all, 1674 students in 8th-10th grade at 12 grammar schools in greater Munich, Germany, were cluster randomized into intervention and control groups. A standardized 60-min prevention lesson focusing on preventable risk factors for headache (physical inactivity, coffee consumption, alcohol consumption and smoking) and providing instructions on stress management and neck and shoulder muscle relaxation exercises was given in a classroom setting. Seven months later, students were reassessed. The main outcome parameter was headache cessation. Logistic regression models with random effects for cluster and adjustment for baseline risk factors were calculated. Nine hundred students (intervention group N = 450, control group N = 450) with headache at baseline and complete data for headache and confounders were included in the analysis. Headache cessation was observed in 9.78% of the control group compared with 16.22% in the intervention group (number needed to treat = 16). Accounting for cluster effects and confounders, the probability of headache cessation in the intervention group was 1.77 (95% confidence interval = [1.08; 2.90]) higher than in the control group. The effect was most pronounced in adolescents with tension-type headache: odds ratio = 2.11 (95% confidence interval = [1.15; 3.80]). Our study demonstrates the effectiveness of a one-time, classroom-based headache prevention programme. © 2014 EAN.

  19. Effects of physical activity on schoolchildren's academic performance: The Active Smarter Kids (ASK) cluster-randomized controlled trial.

    PubMed

    Resaland, Geir K; Aadland, Eivind; Moe, Vegard Fusche; Aadland, Katrine N; Skrede, Turid; Stavnsbo, Mette; Suominen, Laura; Steene-Johannessen, Jostein; Glosvik, Øyvind; Andersen, John R; Kvalheim, Olav M; Engelsrud, Gunn; Andersen, Lars B; Holme, Ingar M; Ommundsen, Yngvar; Kriemler, Susi; van Mechelen, Willem; McKay, Heather A; Ekelund, Ulf; Anderssen, Sigmund A

    2016-10-01

    To investigate the effect of a seven-month, school-based cluster-randomized controlled trial on academic performance in 10-year-old children. In total, 1129 fifth-grade children from 57 elementary schools in Sogn og Fjordane County, Norway, were cluster-randomized by school either to the intervention group or to the control group. The children in the 28 intervention schools participated in a physical activity intervention between November 2014 and June 2015 consisting of three components: 1) 90min/week of physically active educational lessons mainly carried out in the school playground; 2) 5min/day of physical activity breaks during classroom lessons; 3) 10min/day physical activity homework. Academic performance in numeracy, reading and English was measured using standardized Norwegian national tests. Physical activity was measured objectively by accelerometry. We found no effect of the intervention on academic performance in primary analyses (standardized difference 0.01-0.06, p>0.358). Subgroup analyses, however, revealed a favorable intervention effect for those who performed the poorest at baseline (lowest tertile) for numeracy (p=0.005 for the subgroup∗group interaction), compared to controls (standardized difference 0.62, 95% CI 0.19-1.07). This large, rigorously conducted cluster RCT in 10-year-old children supports the notion that there is still inadequate evidence to conclude that increased physical activity in school enhances academic achievement in all children. Still, combining physical activity and learning seems a viable model to stimulate learning in those academically weakest schoolchildren. Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.

  20. GPI-anchored proteins are confined in subdiffraction clusters at the apical surface of polarized epithelial cells.

    PubMed

    Paladino, Simona; Lebreton, Stéphanie; Lelek, Mickaël; Riccio, Patrizia; De Nicola, Sergio; Zimmer, Christophe; Zurzolo, Chiara

    2017-12-01

    Spatio-temporal compartmentalization of membrane proteins is critical for the regulation of diverse vital functions in eukaryotic cells. It was previously shown that, at the apical surface of polarized MDCK cells, glycosylphosphatidylinositol (GPI)-anchored proteins (GPI-APs) are organized in small cholesterol-independent clusters of single GPI-AP species (homoclusters), which are required for the formation of larger cholesterol-dependent clusters formed by multiple GPI-AP species (heteroclusters). This clustered organization is crucial for the biological activities of GPI-APs; hence, understanding the spatio-temporal properties of their membrane organization is of fundamental importance. Here, by using direct stochastic optical reconstruction microscopy coupled to pair correlation analysis (pc-STORM), we were able to visualize and measure the size of these clusters. Specifically, we show that they are non-randomly distributed and have an average size of 67 nm. We also demonstrated that polarized MDCK and non-polarized CHO cells have similar cluster distribution and size, but different sensitivity to cholesterol depletion. Finally, we derived a model that allowed a quantitative characterization of the cluster organization of GPI-APs at the apical surface of polarized MDCK cells for the first time. Experimental FRET (fluorescence resonance energy transfer)/FLIM (fluorescence-lifetime imaging microscopy) data were correlated to the theoretical predictions of the model. © 2017 The Author(s).

  1. Reference Values of Within-District Intraclass Correlations of Academic Achievement by District Characteristics: Results from a Meta-Analysis of District-Specific Values

    ERIC Educational Resources Information Center

    Hedberg, E. C.; Hedges, Larry V.

    2014-01-01

    Randomized experiments are often considered the strongest designs to study the impact of educational interventions. Perhaps the most prevalent class of designs used in large scale education experiments is the cluster randomized design in which entire schools are assigned to treatments. In cluster randomized trials (CRTs) that assign schools to…

  2. A cluster randomized theory-guided oral hygiene trial in adolescents-A latent growth model.

    PubMed

    Aleksejūnienė, J; Brukienė, V

    2018-05-01

    (i) To test whether theory-guided interventions are more effective than conventional dental instruction (CDI) for changing oral hygiene in adolescents and (ii) to examine whether such interventions equally benefit both genders and different socio-economic (SES) groups. A total of 244 adolescents were recruited from three schools, and cluster randomization allocated adolescents to one of the three types of interventions: two were theory-based interventions (Precaution Adoption Process Model or Authoritative Parenting Model) and CDI served as an active control. Oral hygiene levels % (OH) were assessed at baseline, after 3 months and after 12 months. A complete data set was available for 166 adolescents (the total follow-up rate: 69%). There were no significant differences in baseline OH between those who participated throughout the study and those who dropped out. Bivariate and multivariate analyses showed that theory-guided interventions produced significant improvements in oral hygiene and that there were no significant gender or socio-economic differences. Theory-guided interventions produced more positive changes in OH than CDI, and these changes did not differ between gender and SES groups. © 2017 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  3. What is the role and authority of gatekeepers in cluster randomized trials in health research?

    PubMed Central

    2012-01-01

    This article is part of a series of papers examining ethical issues in cluster randomized trials (CRTs) in health research. In the introductory paper in this series, we set out six areas of inquiry that must be addressed if the CRT is to be set on a firm ethical foundation. This paper addresses the sixth of the questions posed, namely, what is the role and authority of gatekeepers in CRTs in health research? ‘Gatekeepers’ are individuals or bodies that represent the interests of cluster members, clusters, or organizations. The need for gatekeepers arose in response to the difficulties in obtaining informed consent because of cluster randomization, cluster-level interventions, and cluster size. In this paper, we call for a more restrictive understanding of the role and authority of gatekeepers. Previous papers in this series have provided solutions to the challenges posed by informed consent in CRTs without the need to invoke gatekeepers. We considered that consent to randomization is not required when cluster members are approached for consent at the earliest opportunity and before any study interventions or data-collection procedures have started. Further, when cluster-level interventions or cluster size means that obtaining informed consent is not possible, a waiver of consent may be appropriate. In this paper, we suggest that the role of gatekeepers in protecting individual interests in CRTs should be limited. Generally, gatekeepers do not have the authority to provide proxy consent for cluster members. When a municipality or other community has a legitimate political authority that is empowered to make such decisions, cluster permission may be appropriate; however, gatekeepers may usefully protect cluster interests in other ways. Cluster consultation may ensure that the CRT addresses local health needs, and is conducted in accord with local values and customs. Gatekeepers may also play an important role in protecting the interests of organizations, such as hospitals, nursing homes, general practices, and schools. In these settings, permission to access the organization relies on resource implications and adherence to institutional policies. PMID:22834691

  4. The role of gender in a smoking cessation intervention: a cluster randomized clinical trial.

    PubMed

    Puente, Diana; Cabezas, Carmen; Rodriguez-Blanco, Teresa; Fernández-Alonso, Carmen; Cebrian, Tránsito; Torrecilla, Miguel; Clemente, Lourdes; Martín, Carlos

    2011-05-23

    The prevalence of smoking in Spain is high in both men and women. The aim of our study was to evaluate the role of gender in the effectiveness of a specific smoking cessation intervention conducted in Spain. This study was a secondary analysis of a cluster randomized clinical trial in which the randomization unit was the Basic Care Unit (family physician and nurse who care for the same group of patients). The intervention consisted of a six-month period of implementing the recommendations of a Clinical Practice Guideline. A total of 2,937 current smokers at 82 Primary Care Centers in 13 different regions of Spain were included (2003-2005). The success rate was measured by a six-month continued abstinence rate at the one-year follow-up. A logistic mixed-effects regression model, taking Basic Care Units as random-effect parameter, was performed in order to analyze gender as a predictor of smoking cessation. At the one-year follow-up, the six-month continuous abstinence quit rate was 9.4% in men and 8.5% in women (p = 0.400). The logistic mixed-effects regression model showed that women did not have a higher odds of being an ex-smoker than men after the analysis was adjusted for confounders (OR adjusted = 0.9, 95% CI = 0.7-1.2). Gender does not appear to be a predictor of smoking cessation at the one-year follow-up in individuals presenting at Primary Care Centers. CLINICALTRIALS.GOV IDENTIFIER: NCT00125905.

  5. The Wilcoxon signed rank test for paired comparisons of clustered data.

    PubMed

    Rosner, Bernard; Glynn, Robert J; Lee, Mei-Ling T

    2006-03-01

    The Wilcoxon signed rank test is a frequently used nonparametric test for paired data (e.g., consisting of pre- and posttreatment measurements) based on independent units of analysis. This test cannot be used for paired comparisons arising from clustered data (e.g., if paired comparisons are available for each of two eyes of an individual). To incorporate clustering, a generalization of the randomization test formulation for the signed rank test is proposed, where the unit of randomization is at the cluster level (e.g., person), while the individual paired units of analysis are at the subunit within cluster level (e.g., eye within person). An adjusted variance estimate of the signed rank test statistic is then derived, which can be used for either balanced (same number of subunits per cluster) or unbalanced (different number of subunits per cluster) data, with an exchangeable correlation structure, with or without tied values. The resulting test statistic is shown to be asymptotically normal as the number of clusters becomes large, if the cluster size is bounded. Simulation studies are performed based on simulating correlated ranked data from a signed log-normal distribution. These studies indicate appropriate type I error for data sets with > or =20 clusters and a superior power profile compared with either the ordinary signed rank test based on the average cluster difference score or the multivariate signed rank test of Puri and Sen. Finally, the methods are illustrated with two data sets, (i) an ophthalmologic data set involving a comparison of electroretinogram (ERG) data in retinitis pigmentosa (RP) patients before and after undergoing an experimental surgical procedure, and (ii) a nutritional data set based on a randomized prospective study of nutritional supplements in RP patients where vitamin E intake outside of study capsules is compared before and after randomization to monitor compliance with nutritional protocols.

  6. Average structure and local configuration of excess oxygen in UO(2+x).

    PubMed

    Wang, Jianwei; Ewing, Rodney C; Becker, Udo

    2014-03-19

    Determination of the local configuration of interacting defects in a crystalline, periodic solid is problematic because defects typically do not have a long-range periodicity. Uranium dioxide, the primary fuel for fission reactors, exists in hyperstoichiometric form, UO(2+x). Those excess oxygen atoms occur as interstitial defects, and these defects are not random but rather partially ordered. The widely-accepted model to date, the Willis cluster based on neutron diffraction, cannot be reconciled with the first-principles molecular dynamics simulations present here. We demonstrate that the Willis cluster is a fair representation of the numerical ratio of different interstitial O atoms; however, the model does not represent the actual local configuration. The simulations show that the average structure of UO(2+x) involves a combination of defect structures including split di-interstitial, di-interstitial, mono-interstitial, and the Willis cluster, and the latter is a transition state that provides for the fast diffusion of the defect cluster. The results provide new insights in differentiating the average structure from the local configuration of defects in a solid and the transport properties of UO(2+x).

  7. Sample size calculations for the design of cluster randomized trials: A summary of methodology.

    PubMed

    Gao, Fei; Earnest, Arul; Matchar, David B; Campbell, Michael J; Machin, David

    2015-05-01

    Cluster randomized trial designs are growing in popularity in, for example, cardiovascular medicine research and other clinical areas and parallel statistical developments concerned with the design and analysis of these trials have been stimulated. Nevertheless, reviews suggest that design issues associated with cluster randomized trials are often poorly appreciated and there remain inadequacies in, for example, describing how the trial size is determined and the associated results are presented. In this paper, our aim is to provide pragmatic guidance for researchers on the methods of calculating sample sizes. We focus attention on designs with the primary purpose of comparing two interventions with respect to continuous, binary, ordered categorical, incidence rate and time-to-event outcome variables. Issues of aggregate and non-aggregate cluster trials, adjustment for variation in cluster size and the effect size are detailed. The problem of establishing the anticipated magnitude of between- and within-cluster variation to enable planning values of the intra-cluster correlation coefficient and the coefficient of variation are also described. Illustrative examples of calculations of trial sizes for each endpoint type are included. Copyright © 2015 Elsevier Inc. All rights reserved.

  8. Hurdle models for multilevel zero-inflated data via h-likelihood.

    PubMed

    Molas, Marek; Lesaffre, Emmanuel

    2010-12-30

    Count data often exhibit overdispersion. One type of overdispersion arises when there is an excess of zeros in comparison with the standard Poisson distribution. Zero-inflated Poisson and hurdle models have been proposed to perform a valid likelihood-based analysis to account for the surplus of zeros. Further, data often arise in clustered, longitudinal or multiple-membership settings. The proper analysis needs to reflect the design of a study. Typically random effects are used to account for dependencies in the data. We examine the h-likelihood estimation and inference framework for hurdle models with random effects for complex designs. We extend the h-likelihood procedures to fit hurdle models, thereby extending h-likelihood to truncated distributions. Two applications of the methodology are presented. Copyright © 2010 John Wiley & Sons, Ltd.

  9. Kinetics of Aggregation with Choice

    DOE PAGES

    Ben-Naim, Eli; Krapivsky, Paul

    2016-12-01

    Here we generalize the ordinary aggregation process to allow for choice. In ordinary aggregation, two random clusters merge and form a larger aggregate. In our implementation of choice, a target cluster and two candidate clusters are randomly selected and the target cluster merges with the larger of the two candidate clusters.We study the long-time asymptotic behavior and find that as in ordinary aggregation, the size density adheres to the standard scaling form. However, aggregation with choice exhibits a number of different features. First, the density of the smallest clusters exhibits anomalous scaling. Second, both the small-size and the large-size tailsmore » of the density are overpopulated, at the expense of the density of moderate-size clusters. Finally, we also study the complementary case where the smaller candidate cluster participates in the aggregation process and find an abundance of moderate clusters at the expense of small and large clusters. Additionally, we investigate aggregation processes with choice among multiple candidate clusters and a symmetric implementation where the choice is between two pairs of clusters.« less

  10. Sample size calculations for stepped wedge and cluster randomised trials: a unified approach

    PubMed Central

    Hemming, Karla; Taljaard, Monica

    2016-01-01

    Objectives To clarify and illustrate sample size calculations for the cross-sectional stepped wedge cluster randomized trial (SW-CRT) and to present a simple approach for comparing the efficiencies of competing designs within a unified framework. Study Design and Setting We summarize design effects for the SW-CRT, the parallel cluster randomized trial (CRT), and the parallel cluster randomized trial with before and after observations (CRT-BA), assuming cross-sectional samples are selected over time. We present new formulas that enable trialists to determine the required cluster size for a given number of clusters. We illustrate by example how to implement the presented design effects and give practical guidance on the design of stepped wedge studies. Results For a fixed total cluster size, the choice of study design that provides the greatest power depends on the intracluster correlation coefficient (ICC) and the cluster size. When the ICC is small, the CRT tends to be more efficient; when the ICC is large, the SW-CRT tends to be more efficient and can serve as an alternative design when the CRT is an infeasible design. Conclusion Our unified approach allows trialists to easily compare the efficiencies of three competing designs to inform the decision about the most efficient design in a given scenario. PMID:26344808

  11. Edge union of networks on the same vertex set

    NASA Astrophysics Data System (ADS)

    Loe, Chuan Wen; Jeldtoft Jensen, Henrik

    2013-06-01

    Random network generators such as Erdős-Rényi, Watts-Strogatz and Barabási-Albert models are used as models to study real-world networks. Let G1(V, E1) and G2(V, E2) be two such networks on the same vertex set V. This paper studies the degree distribution and clustering coefficient of the resultant networks, G(V, E1∪E2).

  12. Prediction of the effect of formulation on the toxicity of chemicals.

    PubMed

    Mistry, Pritesh; Neagu, Daniel; Sanchez-Ruiz, Antonio; Trundle, Paul R; Vessey, Jonathan D; Gosling, John Paul

    2017-01-01

    Two approaches for the prediction of which of two vehicles will result in lower toxicity for anticancer agents are presented. Machine-learning models are developed using decision tree, random forest and partial least squares methodologies and statistical evidence is presented to demonstrate that they represent valid models. Separately, a clustering method is presented that allows the ordering of vehicles by the toxicity they show for chemically-related compounds.

  13. Sample size determination for GEE analyses of stepped wedge cluster randomized trials.

    PubMed

    Li, Fan; Turner, Elizabeth L; Preisser, John S

    2018-06-19

    In stepped wedge cluster randomized trials, intact clusters of individuals switch from control to intervention from a randomly assigned period onwards. Such trials are becoming increasingly popular in health services research. When a closed cohort is recruited from each cluster for longitudinal follow-up, proper sample size calculation should account for three distinct types of intraclass correlations: the within-period, the inter-period, and the within-individual correlations. Setting the latter two correlation parameters to be equal accommodates cross-sectional designs. We propose sample size procedures for continuous and binary responses within the framework of generalized estimating equations that employ a block exchangeable within-cluster correlation structure defined from the distinct correlation types. For continuous responses, we show that the intraclass correlations affect power only through two eigenvalues of the correlation matrix. We demonstrate that analytical power agrees well with simulated power for as few as eight clusters, when data are analyzed using bias-corrected estimating equations for the correlation parameters concurrently with a bias-corrected sandwich variance estimator. © 2018, The International Biometric Society.

  14. Do Interim Assessments Reduce the Race and SES Achievement Gaps?

    ERIC Educational Resources Information Center

    Konstantopoulos, Spyros; Li, Wei; Miller, Shazia R.; van der Ploeg, Arie

    2017-01-01

    The authors examined differential effects of interim assessments on minority and low socioeconomic status students' achievement in Grades K-6. They conducted a large-scale cluster randomized experiment in 2009-2010 to evaluate the impact of Indiana's policy initiative introducing interim assessments statewide. The authors used 2-level models to…

  15. Writing Week-Journals to Improve the Writing Quality of Fourth-Graders' Compositions

    ERIC Educational Resources Information Center

    Rosário, Pedro; Högemann, Julia; Núñez, José Carlos; Vallejo, Guillermo; Cunha, Jennifer; Oliveira, Vera; Fuentes, Sonia; Rodrigues, Celestino

    2017-01-01

    Students' writing problems are a global educational concern and is in need of particular attention. This study aims to examine the impact of providing extra writing opportunities (i.e., writing journals) on the quality of writing compositions. A longitudinal cluster-randomized controlled design using a multilevel modeling analysis with 182 fourth…

  16. Power Analysis for Models of Change in Cluster Randomized Designs

    ERIC Educational Resources Information Center

    Li, Wei; Konstantopoulos, Spyros

    2017-01-01

    Field experiments in education frequently assign entire groups such as schools to treatment or control conditions. These experiments incorporate sometimes a longitudinal component where for example students are followed over time to assess differences in the average rate of linear change, or rate of acceleration. In this study, we provide methods…

  17. Invasion Percolation and Global Optimization

    NASA Astrophysics Data System (ADS)

    Barabási, Albert-László

    1996-05-01

    Invasion bond percolation (IBP) is mapped exactly into Prim's algorithm for finding the shortest spanning tree of a weighted random graph. Exploring this mapping, which is valid for arbitrary dimensions and lattices, we introduce a new IBP model that belongs to the same universality class as IBP and generates the minimal energy tree spanning the IBP cluster.

  18. Assessing Impacts of "Math in Focus," a "Singapore Math" Program

    ERIC Educational Resources Information Center

    Jaciw, Andrew P.; Hegseth, Whitney Michelle; Lin, Li; Toby, Megan; Newman, Denis; Ma, Boya; Zacamy, Jenna

    2016-01-01

    This study investigates, through a cluster randomized trial, the impact of "Math in Focus," a core mathematics program modeled after instructional approaches used in Singapore, on third- through fifth-grade students' achievement in mathematics. The program is currently being used in more than 400 school districts in the United States.…

  19. A Statistical Model for Misreported Binary Outcomes in Clustered RCTs of Education Interventions

    ERIC Educational Resources Information Center

    Schochet, Peter Z.

    2013-01-01

    In randomized control trials (RCTs) of educational interventions, there is a growing literature on impact estimation methods to adjust for missing student outcome data using such methods as multiple imputation, the construction of nonresponse weights, casewise deletion, and maximum likelihood methods (see, for example, Allison, 2002; Graham, 2009;…

  20. Effectiveness of a self-management program for dual sensory impaired seniors in aged care settings: study protocol for a cluster randomized controlled trial.

    PubMed

    Roets-Merken, Lieve M; Graff, Maud J L; Zuidema, Sytse U; Hermsen, Pieter G J M; Teerenstra, Steven; Kempen, Gertrudis I J M; Vernooij-Dassen, Myrra J F J

    2013-10-07

    Five to 25 percent of residents in aged care settings have a combined hearing and visual sensory impairment. Usual care is generally restricted to single sensory impairment, neglecting the consequences of dual sensory impairment on social participation and autonomy. The aim of this study is to evaluate the effectiveness of a self-management program for seniors who acquired dual sensory impairment at old age. In a cluster randomized, single-blind controlled trial, with aged care settings as the unit of randomization, the effectiveness of a self-management program will be compared to usual care. A minimum of 14 and maximum of 20 settings will be randomized to either the intervention cluster or the control cluster, aiming to include a total of 132 seniors with dual sensory impairment. Each senior will be linked to a licensed practical nurse working at the setting. During a five to six month intervention period, nurses at the intervention clusters will be trained in a self-management program to support and empower seniors to use self-management strategies. In two separate diaries, nurses keep track of the interviews with the seniors and their reflections on their own learning process. Nurses of the control clusters offer care as usual. At senior level, the primary outcome is the social participation of the seniors measured using the Hearing Handicap Questionnaire and the Activity Card Sort, and secondary outcomes are mood, autonomy and quality of life. At nurse level, the outcome is job satisfaction. Effectiveness will be evaluated using linear mixed model analysis. The results of this study will provide evidence for the effectiveness of the Self-Management Program for seniors with dual sensory impairment living in aged care settings. The findings are expected to contribute to the knowledge on the program's potential to enhance social participation and autonomy of the seniors, as well as increasing the job satisfaction of the licensed practical nurses. Furthermore, an extensive process evaluation will take place which will offer insight in the quality and feasibility of the sampling and intervention process. If it is shown to be effective and feasible, this Self-Management Program could be widely disseminated. ClinicalTrials.gov, NCT01217502.

  1. Effectiveness of a self-management program for dual sensory impaired seniors in aged care settings: study protocol for a cluster randomized controlled trial

    PubMed Central

    2013-01-01

    Background Five to 25 percent of residents in aged care settings have a combined hearing and visual sensory impairment. Usual care is generally restricted to single sensory impairment, neglecting the consequences of dual sensory impairment on social participation and autonomy. The aim of this study is to evaluate the effectiveness of a self-management program for seniors who acquired dual sensory impairment at old age. Methods/Design In a cluster randomized, single-blind controlled trial, with aged care settings as the unit of randomization, the effectiveness of a self-management program will be compared to usual care. A minimum of 14 and maximum of 20 settings will be randomized to either the intervention cluster or the control cluster, aiming to include a total of 132 seniors with dual sensory impairment. Each senior will be linked to a licensed practical nurse working at the setting. During a five to six month intervention period, nurses at the intervention clusters will be trained in a self-management program to support and empower seniors to use self-management strategies. In two separate diaries, nurses keep track of the interviews with the seniors and their reflections on their own learning process. Nurses of the control clusters offer care as usual. At senior level, the primary outcome is the social participation of the seniors measured using the Hearing Handicap Questionnaire and the Activity Card Sort, and secondary outcomes are mood, autonomy and quality of life. At nurse level, the outcome is job satisfaction. Effectiveness will be evaluated using linear mixed model analysis. Discussion The results of this study will provide evidence for the effectiveness of the Self-Management Program for seniors with dual sensory impairment living in aged care settings. The findings are expected to contribute to the knowledge on the program’s potential to enhance social participation and autonomy of the seniors, as well as increasing the job satisfaction of the licensed practical nurses. Furthermore, an extensive process evaluation will take place which will offer insight in the quality and feasibility of the sampling and intervention process. If it is shown to be effective and feasible, this Self-Management Program could be widely disseminated. Clinical trials registration ClinicalTrials.gov, NCT01217502. PMID:24099315

  2. Reducing salt intake for prevention of cardiovascular diseases in high-risk patients by advanced health education intervention (RESIP-CVD study), Northern Thailand: study protocol for a cluster randomized trial

    PubMed Central

    2012-01-01

    Background Decreasing salt consumption can prevent cardiovascular diseases (CVD). Practically, it is difficult to promote people’s awareness of daily salt intake and to change their eating habits in terms of reducing salt intake for better cardiovascular health. Health education programs visualizing daily dietary salt content and intake may promote lifestyle changes in patients at high risk of cardiovascular diseases. Methods/Design This is a cluster randomized trial. A total of 800 high-CVD-risk patients attending diabetes and hypertension clinics at health centers in Muang District, Chiang Rai province, Thailand, will be studied with informed consent. A health center recruiting 100 participants is a cluster, the unit of randomization. Eight clusters will be randomized into intervention and control arms and followed up for 1 year. Within the intervention clusters the following will be undertaken: (1) salt content in the daily diet will be measured and shown to study participants; (2) 24-hour salt intake will be estimated in overnight-collected urine and the results shown to the participants; (3) a dietician will assist small group health education classes in cooking meals with less salt. The primary outcome is blood pressure change at the 1-year follow-up. Secondary outcomes at the 1-year follow-up are estimated 24-hoursalt intake, incidence of CVD events and CVD death. The intention-to-treat analysis will be followed. Blood pressure and estimated 24-hour salt intake will be compared between intervention and control groups at the cluster and individual level at the 1-year follow-up. Clinical CVD events and deaths will be analyzed by time-event analysis. Retinal blood vessel calibers of CVD-risk patients will be assessed cross-sectionally. Behavioral change to reduce salt intake and the influencing factors will be determined by structured equation model (SEM). Multilevel regression analyses will be applied. Finally, the cost effectiveness of the intervention will be analyzed. Discussion This study is unique as it will recruit the individuals most vulnerable to CVD morbidity and mortality by applying the general Framingham CVD risk scoring system. Dietary salt reduction will be applied as a prioritized, community level intervention for the prevention of CVD in a developing country. Trial registration ISRCTN39416277 PMID:22947342

  3. Reducing salt intake for prevention of cardiovascular diseases in high-risk patients by advanced health education intervention (RESIP-CVD study), Northern Thailand: study protocol for a cluster randomized trial.

    PubMed

    Aung, Myo Nyein; Yuasa, Motoyuki; Moolphate, Saiyud; Nedsuwan, Supalert; Yokokawa, Hidehiro; Kitajima, Tsutomu; Minematsu, Kazuo; Tanimura, Susumu; Fukuda, Hiroshi; Hiratsuka, Yoshimune; Ono, Koichi; Kawai, Sachio; Marui, Eiji

    2012-09-04

    Decreasing salt consumption can prevent cardiovascular diseases (CVD). Practically, it is difficult to promote people's awareness of daily salt intake and to change their eating habits in terms of reducing salt intake for better cardiovascular health. Health education programs visualizing daily dietary salt content and intake may promote lifestyle changes in patients at high risk of cardiovascular diseases. This is a cluster randomized trial. A total of 800 high-CVD-risk patients attending diabetes and hypertension clinics at health centers in Muang District, Chiang Rai province, Thailand, will be studied with informed consent. A health center recruiting 100 participants is a cluster, the unit of randomization. Eight clusters will be randomized into intervention and control arms and followed up for 1 year. Within the intervention clusters the following will be undertaken: (1) salt content in the daily diet will be measured and shown to study participants; (2) 24-hour salt intake will be estimated in overnight-collected urine and the results shown to the participants; (3) a dietician will assist small group health education classes in cooking meals with less salt. The primary outcome is blood pressure change at the 1-year follow-up. Secondary outcomes at the 1-year follow-up are estimated 24-hoursalt intake, incidence of CVD events and CVD death. The intention-to-treat analysis will be followed.Blood pressure and estimated 24-hour salt intake will be compared between intervention and control groups at the cluster and individual level at the 1-year follow-up. Clinical CVD events and deaths will be analyzed by time-event analysis. Retinal blood vessel calibers of CVD-risk patients will be assessed cross-sectionally. Behavioral change to reduce salt intake and the influencing factors will be determined by structured equation model (SEM). Multilevel regression analyses will be applied. Finally, the cost effectiveness of the intervention will be analyzed. This study is unique as it will recruit the individuals most vulnerable to CVD morbidity and mortality by applying the general Framingham CVD risk scoring system. Dietary salt reduction will be applied as a prioritized, community level intervention for the prevention of CVD in a developing country. ISRCTN39416277.

  4. Impact of Text Message Reminders on Caregivers’ Adherence to a Home Fortification Program Against Child Anemia in Rural Western China: A Cluster-Randomized Controlled Trial

    PubMed Central

    Zhou, Huan; Sun, Shuai; Sylvia, Sean; Yue, Ai; Shi, Yaojiang; Zhang, Linxiu; Medina, Alexis; Rozelle, Scott

    2016-01-01

    Objectives. To test whether text message reminders sent to caregivers improve the effectiveness of a home micronutrient fortification program in western China. Methods. We carried out a cluster-randomized controlled trial in 351 villages (clusters) in Shaanxi Province in 2013 and 2014, enrolling children aged 6 to 12 months. We randomly assigned each village to 1 of 3 groups: free delivery group, text messaging group, or control group. We collected information on compliance with treatments and hemoglobin concentrations from all children at baseline and 6-month follow-up. We estimated the intent-to-treat effects on compliance and child anemia using a logistic regression model. Results. There were 1393 eligible children. We found that assignment to the text messaging group led to an increase in full compliance (marginal effect = 0.10; 95% confidence interval [CI] = 0.03, 0.16) compared with the free delivery group and decrease in the rate of anemia at end line relative to the control group (marginal effect = −0.07; 95% CI = −0.12, −0.01), but not relative to the free delivery group (marginal effect = −0.03; 95% CI = −0.09, 0.03). Conclusions. Text messages improved compliance of caregivers to a home fortification program and children’s nutrition. PMID:27077354

  5. Ranking and clustering of nodes in networks with smart teleportation

    NASA Astrophysics Data System (ADS)

    Lambiotte, R.; Rosvall, M.

    2012-05-01

    Random teleportation is a necessary evil for ranking and clustering directed networks based on random walks. Teleportation enables ergodic solutions, but the solutions must necessarily depend on the exact implementation and parametrization of the teleportation. For example, in the commonly used PageRank algorithm, the teleportation rate must trade off a heavily biased solution with a uniform solution. Here we show that teleportation to links rather than nodes enables a much smoother trade-off and effectively more robust results. We also show that, by not recording the teleportation steps of the random walker, we can further reduce the effect of teleportation with dramatic effects on clustering.

  6. Implications of crater distributions on Venus

    NASA Technical Reports Server (NTRS)

    Kaula, W. M.

    1993-01-01

    The horizontal locations of craters on Venus are consistent with randomness. However, (1) randomness does not make crater counts useless for age indications; (2) consistency does not imply necessity or optimality; and (3) horizontal location is not the only reference frame against which to test models. Re (1), the apparent smallness of resurfacing areas means that a region on the order of one percent of the planet with a typical number of craters, 5-15, will have a range of feature ages of several 100 My. Re (2), models of resurfacing somewhat similar to Earth's can be found that are also consistent and more optimal than random: i.e., resurfacing occurring in clusters, that arise and die away in lime intervals on the order of 50 My. These agree with the observation that there are more areas of high crater density, and fewer of moderate density, than optimal for random. Re (3), 799 crater elevations were tested; there are more at low elevations and fewer at high elevations than optimal for random: i.e., 54.6 percent below the median. Only one of 40 random sets of 799 was as extreme.

  7. Scattering Properties of Heterogeneous Mineral Particles with Absorbing Inclusions

    NASA Technical Reports Server (NTRS)

    Dlugach, Janna M.; Mishchenko, Michael I.

    2015-01-01

    We analyze the results of numerically exact computer modeling of scattering and absorption properties of randomly oriented poly-disperse heterogeneous particles obtained by placing microscopic absorbing grains randomly on the surfaces of much larger spherical mineral hosts or by imbedding them randomly inside the hosts. These computations are paralleled by those for heterogeneous particles obtained by fully encapsulating fractal-like absorbing clusters in the mineral hosts. All computations are performed using the superposition T-matrix method. In the case of randomly distributed inclusions, the results are compared with the outcome of Lorenz-Mie computations for an external mixture of the mineral hosts and absorbing grains. We conclude that internal aggregation can affect strongly both the integral radiometric and differential scattering characteristics of the heterogeneous particle mixtures.

  8. Random variability explains apparent global clustering of large earthquakes

    USGS Publications Warehouse

    Michael, A.J.

    2011-01-01

    The occurrence of 5 Mw ≥ 8.5 earthquakes since 2004 has created a debate over whether or not we are in a global cluster of large earthquakes, temporarily raising risks above long-term levels. I use three classes of statistical tests to determine if the record of M ≥ 7 earthquakes since 1900 can reject a null hypothesis of independent random events with a constant rate plus localized aftershock sequences. The data cannot reject this null hypothesis. Thus, the temporal distribution of large global earthquakes is well-described by a random process, plus localized aftershocks, and apparent clustering is due to random variability. Therefore the risk of future events has not increased, except within ongoing aftershock sequences, and should be estimated from the longest possible record of events.

  9. Cluster Randomized Trial of a Church-Based Peer Counselor and Tailored Newsletter Intervention to Promote Colorectal Cancer Screening and Physical Activity among Older African Americans

    ERIC Educational Resources Information Center

    Leone, Lucia A.; Allicock, Marlyn; Pignone, Michael P.; Walsh, Joan F.; Johnson, La-Shell; Armstrong-Brown, Janelle; Carr, Carol C.; Langford, Aisha; Ni, Andy; Resnicow, Ken; Campbell, Marci K.

    2016-01-01

    Action Through Churches in Time to Save Lives (ACTS) of Wellness was a cluster randomized controlled trial developed to promote colorectal cancer screening and physical activity (PA) within urban African American churches. Churches were recruited from North Carolina (n = 12) and Michigan (n = 7) and were randomized to intervention (n = 10) or…

  10. Cost-Effectiveness of a Long-Term Internet-Delivered Worksite Health Promotion Programme on Physical Activity and Nutrition: A Cluster Randomized Controlled Trial

    ERIC Educational Resources Information Center

    Robroek, Suzan J. W.; Polinder, Suzanne; Bredt, Folef J.; Burdorf, Alex

    2012-01-01

    This study aims to evaluate the cost-effectiveness of a long-term workplace health promotion programme on physical activity (PA) and nutrition. In total, 924 participants enrolled in a 2-year cluster randomized controlled trial, with departments (n = 74) within companies (n = 6) as the unit of randomization. The intervention was compared with a…

  11. Lumbar Imaging with Reporting of Epidemiology (LIRE)- Protocol for a Pragmatic Cluster Randomized Trial

    PubMed Central

    Jarvik, Jeffrey G.; Comstock, Bryan A.; James, Kathryn T.; Avins, Andrew L.; Bresnahan, Brian W.; Deyo, Richard A.; Luetmer, Patrick H.; Friedly, Janna L.; Meier, Eric N.; Cherkin, Daniel C.; Gold, Laura S.; Rundell, Sean D.; Halabi, Safwan S.; Kallmes, David F.; Tan, Katherine W.; Turner, Judith A.; Kessler, Larry G.; Lavallee, Danielle C.; Stephens, Kari A.; Heagerty, Patrick J.

    2015-01-01

    Background Diagnostic imaging is often the first step in evaluating patients with back pain and likely functions as a “gateway” to a subsequent cascade of interventions. However, lumbar spine imaging frequently reveals incidental findings among normal, pain-free individuals suggesting that treatment of these “abnormalities” may not be warranted. Our prior work suggested that inserting the prevalence of imaging findings in patients without back pain into spine imaging reports may reduce subsequent interventions. We are now conducting a pragmatic cluster randomized clinical trial to test the hypothesis that inserting this prevalence data into lumbar spine imaging reports for studies ordered by primary care providers will reduce subsequent spine-related interventions. Methods/Design We are using a stepped wedge design that sequentially randomizes 100 primary care clinics at four health systems to receive either standard lumbar spine imaging reports, or reports containing prevalence data for common imaging findings in patients without back pain. We capture all outcomes passively through the electronic medical record. Our primary outcome is spine-related intervention intensity based on Relative Value Units (RVUs) during the following year. Secondary outcomes include subsequent prescriptions for opioid analgesics and cross-sectional lumbar spine re-imaging. Discussion If our study shows that adding prevalence data to spine imaging reports decreases subsequent back-related RVUs, this intervention could be easily generalized and applied to other kinds of testing, as well as other conditions where incidental findings may be common. Our study also serves as a model for cluster randomized trials that are minimal risk and highly pragmatic. PMID:26493088

  12. Research on Some Bus Transport Networks with Random Overlapping Clique Structure

    NASA Astrophysics Data System (ADS)

    Yang, Xu-Hua; Wang, Bo; Wang, Wan-Liang; Sun, You-Xian

    2008-11-01

    On the basis of investigating the statistical data of bus transport networks of three big cities in China, we propose that each bus route is a clique (maximal complete subgraph) and a bus transport network (BTN) consists of a lot of cliques, which intensively connect and overlap with each other. We study the network properties, which include the degree distribution, multiple edges' overlapping time distribution, distribution of the overlap size between any two overlapping cliques, distribution of the number of cliques that a node belongs to. Naturally, the cliques also constitute a network, with the overlapping nodes being their multiple links. We also research its network properties such as degree distribution, clustering, average path length, and so on. We propose that a BTN has the properties of random clique increment and random overlapping clique, at the same time, a BTN is a small-world network with highly clique-clustered and highly clique-overlapped. Finally, we introduce a BTN evolution model, whose simulation results agree well with the statistical laws that emerge in real BTNs.

  13. Best (but oft-forgotten) practices: designing, analyzing, and reporting cluster randomized controlled trials.

    PubMed

    Brown, Andrew W; Li, Peng; Bohan Brown, Michelle M; Kaiser, Kathryn A; Keith, Scott W; Oakes, J Michael; Allison, David B

    2015-08-01

    Cluster randomized controlled trials (cRCTs; also known as group randomized trials and community-randomized trials) are multilevel experiments in which units that are randomly assigned to experimental conditions are sets of grouped individuals, whereas outcomes are recorded at the individual level. In human cRCTs, clusters that are randomly assigned are typically families, classrooms, schools, worksites, or counties. With growing interest in community-based, public health, and policy interventions to reduce obesity or improve nutrition, the use of cRCTs has increased. Errors in the design, analysis, and interpretation of cRCTs are unfortunately all too common. This situation seems to stem in part from investigator confusion about how the unit of randomization affects causal inferences and the statistical procedures required for the valid estimation and testing of effects. In this article, we provide a brief introduction and overview of the importance of cRCTs and highlight and explain important considerations for the design, analysis, and reporting of cRCTs by using published examples. © 2015 American Society for Nutrition.

  14. Design of a Phase III cluster randomized trial to assess the efficacy and safety of a malaria transmission blocking vaccine.

    PubMed

    Delrieu, Isabelle; Leboulleux, Didier; Ivinson, Karen; Gessner, Bradford D

    2015-03-24

    Vaccines interrupting Plasmodium falciparum malaria transmission targeting sexual, sporogonic, or mosquito-stage antigens (SSM-VIMT) are currently under development to reduce malaria transmission. An international group of malaria experts was established to evaluate the feasibility and optimal design of a Phase III cluster randomized trial (CRT) that could support regulatory review and approval of an SSM-VIMT. The consensus design is a CRT with a sentinel population randomly selected from defined inner and buffer zones in each cluster, a cluster size sufficient to assess true vaccine efficacy in the inner zone, and inclusion of ongoing assessment of vaccine impact stratified by distance of residence from the cluster edge. Trials should be conducted first in areas of moderate transmission, where SSM-VIMT impact should be greatest. Sample size estimates suggest that such a trial is feasible, and within the range of previously supported trials of malaria interventions, although substantial issues to implementation exist. Copyright © 2015 Elsevier Ltd. All rights reserved.

  15. Nonconventional screening of the Coulomb interaction in FexOy clusters: An ab initio study

    NASA Astrophysics Data System (ADS)

    Peters, L.; Şaşıoǧlu, E.; Rossen, S.; Friedrich, C.; Blügel, S.; Katsnelson, M. I.

    2017-04-01

    From microscopic point-dipole model calculations of the screening of the Coulomb interaction in nonpolar systems by polarizable atoms, it is known that screening strongly depends on dimensionality. For example, in one-dimensional systems, the short-range interaction is screened, while the long-range interaction is antiscreened. This antiscreening is also observed in some zero-dimensional structures, i.e., molecular systems. By means of ab initio calculations in conjunction with the random-phase approximation (RPA) within the FLAPW method, we study screening of the Coulomb interaction in FexOy clusters. For completeness, these results are compared with their bulk counterpart magnetite. It appears that the on-site Coulomb interaction is very well screened both in the clusters and bulk. On the other hand, for the intersite Coulomb interaction, the important observation is made that it is almost constant throughout the clusters, while for the bulk it is almost completely screened. More precisely and interestingly, in the clusters antiscreening is observed by means of ab initio calculations.

  16. Edible oil structures at low and intermediate concentrations. I. Modeling, computer simulation, and predictions for X ray scattering

    NASA Astrophysics Data System (ADS)

    Pink, David A.; Quinn, Bonnie; Peyronel, Fernanda; Marangoni, Alejandro G.

    2013-12-01

    Triacylglycerols (TAGs) are biologically important molecules which form the recently discovered highly anisotropic crystalline nanoplatelets (CNPs) and, ultimately, the large-scale fat crystal networks in edible oils. Identifying the hierarchies of these networks and how they spontaneously self-assemble is important to understanding their functionality and oil binding capacity. We have modelled CNPs and studied how they aggregate under the assumption that all CNPs are present before aggregation begins and that their solubility in the liquid oil is very low. We represented CNPs as rigid planar arrays of spheres with diameter ≈50 nm and defined the interaction between spheres in terms of a Hamaker coefficient, A, and a binding energy, VB. We studied three cases: weak binding, |VB|/kBT ≪ 1, physically realistic binding, VB = Vd(R, Δ), so that |VB|/kBT ≈ 1, and Strong binding with |VB|/kBT ≫ 1. We divided the concentration of CNPs, ϕ, with 0≤ϕ= 10-2 (solid fat content) ≤1, into two regions: Low and intermediate concentrations with 0<ϕ<0.25 and high concentrations with 0.25 < ϕ and considered only the first case. We employed Monte Carlo computer simulation to model CNP aggregation and analyzed them using static structure functions, S(q). We found that strong binding cases formed aggregates with fractal dimension, D, 1.7≤D ≤1.8, in accord with diffusion limited cluster-cluster aggregation (DLCA) and weak binding formed aggregates with D =3, indicating a random distribution of CNPs. We found that models with physically realistic intermediate binding energies formed linear multilayer stacks of CNPs (TAGwoods) with fractal dimension D =1 for ϕ =0.06,0.13, and 0.22. TAGwood lengths were greater at lower ϕ than at higher ϕ, where some of the aggregates appeared as thick CNPs. We increased the spatial scale and modelled the TAGwoods as rigid linear arrays of spheres of diameter ≈500 nm, interacting via the attractive van der Waals interaction. We found that TAGwoods aggregated via DLCA into clusters with fractal dimension D =1.7-1.8. As the simulations were run further, TAGwoods relaxed their positions in order to maximize the attractive interaction making the process look like reaction limited cluster-cluster aggregation with the fractal dimension increasing to D =2.0-2.1. For higher concentrations of CNPs, many TAGwood clusters were formed and, because of their weak interactions, were distributed randomly with D =3.0. We summarize the hierarchy of structures and make predictions for X-ray scattering.

  17. Epidemic Threshold in Structured Scale-Free Networks

    NASA Astrophysics Data System (ADS)

    EguíLuz, VíCtor M.; Klemm, Konstantin

    2002-08-01

    We analyze the spreading of viruses in scale-free networks with high clustering and degree correlations, as found in the Internet graph. For the susceptible-infected-susceptible model of epidemics the prevalence undergoes a phase transition at a finite threshold of the transmission probability. Comparing with the absence of a finite threshold in networks with purely random wiring, our result suggests that high clustering (modularity) and degree correlations protect scale-free networks against the spreading of viruses. We introduce and verify a quantitative description of the epidemic threshold based on the connectivity of the neighborhoods of the hubs.

  18. The Implications of "Contamination" for Experimental Design in Education

    ERIC Educational Resources Information Center

    Rhoads, Christopher H.

    2011-01-01

    Experimental designs that randomly assign entire clusters of individuals (e.g., schools and classrooms) to treatments are frequently advocated as a way of guarding against contamination of the estimated average causal effect of treatment. However, in the absence of contamination, experimental designs that randomly assign intact clusters to…

  19. Propensity score matching with clustered data. An application to the estimation of the impact of caesarean section on the Apgar score.

    PubMed

    Arpino, Bruno; Cannas, Massimo

    2016-05-30

    This article focuses on the implementation of propensity score matching for clustered data. Different approaches to reduce bias due to cluster-level confounders are considered and compared using Monte Carlo simulations. We investigated methods that exploit the clustered structure of the data in two ways: in the estimation of the propensity score model (through the inclusion of fixed or random effects) or in the implementation of the matching algorithm. In addition to a pure within-cluster matching, we also assessed the performance of a new approach, 'preferential' within-cluster matching. This approach first searches for control units to be matched to treated units within the same cluster. If matching is not possible within-cluster, then the algorithm searches in other clusters. All considered approaches successfully reduced the bias due to the omission of a cluster-level confounder. The preferential within-cluster matching approach, combining the advantages of within-cluster and between-cluster matching, showed a relatively good performance both in the presence of big and small clusters, and it was often the best method. An important advantage of this approach is that it reduces the number of unmatched units as compared with a pure within-cluster matching. We applied these methods to the estimation of the effect of caesarean section on the Apgar score using birth register data. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  20. Resonance, criticality, and emergence in city traffic investigated in cellular automaton models.

    PubMed

    Varas, A; Cornejo, M D; Toledo, B A; Muñoz, V; Rogan, J; Zarama, R; Valdivia, J A

    2009-11-01

    The complex behavior that occurs when traffic lights are synchronized is studied for a row of interacting cars. The system is modeled through a cellular automaton. Two strategies are considered: all lights in phase and a "green wave" with a propagating green signal. It is found that the mean velocity near the resonant condition follows a critical scaling law. For the green wave, it is shown that the mean velocity scaling law holds even for random separation between traffic lights and is not dependent on the density. This independence on car density is broken when random perturbations are considered in the car velocity. Random velocity perturbations also have the effect of leading the system to an emergent state, where cars move in clusters, but with an average velocity which is independent of traffic light switching for large injection rates.

  1. Psychosocial education improves low back pain beliefs: results from a cluster randomized clinical trial (NCT00373009) in a primary prevention setting.

    PubMed

    George, Steven Z; Teyhen, Deydre S; Wu, Samuel S; Wright, Alison C; Dugan, Jessica L; Yang, Guijun; Robinson, Michael E; Childs, John D

    2009-07-01

    The general population has a pessimistic view of low back pain (LBP), and evidence-based information has been used to positively influence LBP beliefs in previously reported mass media studies. However, there is a lack of randomized trials investigating whether LBP beliefs can be modified in primary prevention settings. This cluster randomized clinical trial investigated the effect of an evidence-based psychosocial educational program (PSEP) on LBP beliefs for soldiers completing military training. A military setting was selected for this clinical trial, because LBP is a common cause of soldier disability. Companies of soldiers (n = 3,792) were recruited, and cluster randomized to receive a PSEP or no education (control group, CG). The PSEP consisted of an interactive seminar, and soldiers were issued the Back Book for reference material. The primary outcome measure was the back beliefs questionnaire (BBQ), which assesses inevitable consequences of and ability to cope with LBP. The BBQ was administered before randomization and 12 weeks later. A linear mixed model was fitted for the BBQ at the 12-week follow-up, and a generalized linear mixed model was fitted for the dichotomous outcomes on BBQ change of greater than two points. Sensitivity analyses were performed to account for drop out. BBQ scores (potential range: 9-45) improved significantly from baseline of 25.6 +/- 5.7 (mean +/- SD) to 26.9 +/- 6.2 for those receiving the PSEP, while there was a significant decline from 26.1 +/- 5.7 to 25.6 +/- 6.0 for those in the CG. The adjusted mean BBQ score at follow-up for those receiving the PSEP was 1.49 points higher than those in the CG (P < 0.0001). The adjusted odds ratio of BBQ improvement of greater than two points for those receiving the PSEP was 1.51 (95% CI = 1.22-1.86) times that of those in the CG. BBQ improvement was also mildly associated with race and college education. Sensitivity analyses suggested minimal influence of drop out. In conclusion, soldiers that received the PSEP had an improvement in their beliefs related to the inevitable consequences of and ability to cope with LBP. This is the first randomized trial to show positive influence on LBP beliefs in a primary prevention setting, and these findings have potentially important public health implications for prevention of LBP.

  2. Addressing the complexity of water chemistry in environmental fate modeling for engineered nanoparticles.

    PubMed

    Sani-Kast, Nicole; Scheringer, Martin; Slomberg, Danielle; Labille, Jérôme; Praetorius, Antonia; Ollivier, Patrick; Hungerbühler, Konrad

    2015-12-01

    Engineered nanoparticle (ENP) fate models developed to date - aimed at predicting ENP concentration in the aqueous environment - have limited applicability because they employ constant environmental conditions along the modeled system or a highly specific environmental representation; both approaches do not show the effects of spatial and/or temporal variability. To address this conceptual gap, we developed a novel modeling strategy that: 1) incorporates spatial variability in environmental conditions in an existing ENP fate model; and 2) analyzes the effect of a wide range of randomly sampled environmental conditions (representing variations in water chemistry). This approach was employed to investigate the transport of nano-TiO2 in the Lower Rhône River (France) under numerous sets of environmental conditions. The predicted spatial concentration profiles of nano-TiO2 were then grouped according to their similarity by using cluster analysis. The analysis resulted in a small number of clusters representing groups of spatial concentration profiles. All clusters show nano-TiO2 accumulation in the sediment layer, supporting results from previous studies. Analysis of the characteristic features of each cluster demonstrated a strong association between the water conditions in regions close to the ENP emission source and the cluster membership of the corresponding spatial concentration profiles. In particular, water compositions favoring heteroaggregation between the ENPs and suspended particulate matter resulted in clusters of low variability. These conditions are, therefore, reliable predictors of the eventual fate of the modeled ENPs. The conclusions from this study are also valid for ENP fate in other large river systems. Our results, therefore, shift the focus of future modeling and experimental research of ENP environmental fate to the water characteristic in regions near the expected ENP emission sources. Under conditions favoring heteroaggregation in these regions, the fate of the ENPs can be readily predicted. Copyright © 2014 Elsevier B.V. All rights reserved.

  3. DOE Office of Scientific and Technical Information (OSTI.GOV)

    ALAM,TODD M.

    Monte Carlo simulations of phosphate tetrahedron connectivity distributions in alkali and alkaline earth phosphate glasses are reported. By utilizing a discrete bond model, the distribution of next-nearest neighbor connectivities between phosphate polyhedron for random, alternating and clustering bonding scenarios was evaluated as a function of the relative bond energy difference. The simulated distributions are compared to experimentally observed connectivities reported for solid-state two-dimensional exchange and double-quantum NMR experiments of phosphate glasses. These Monte Carlo simulations demonstrate that the polyhedron connectivity is best described by a random distribution in lithium phosphate and calcium phosphate glasses.

  4. Electrostatic effects on clustering and ion dynamics in ionomer melts

    NASA Astrophysics Data System (ADS)

    Ma, Boran; Nguyen, Trung; Pryamitsyn, Victor; Olvera de La Cruz, Monica

    An understanding of the relationships between ionomer chain morphology, dynamics and counter-ion mobility is a key factor in the design of ion conducting membranes for battery applications. In this study, we investigate the influence of electrostatic coupling between randomly charged copolymers (ionomers) and counter ions on the structural and dynamic features of a model system of ionomer melts. Using coarse-grained molecular dynamics (CGMD) simulations, we found that variations in electrostatic coupling strength (Γ) remarkably affect the formation of ion-counter ion clusters, ion mobility, and polymer dynamics for a range of charged monomer fractions. Specifically, an increase in Γ leads to larger ionic cluster sizes and reduced polymer and ion mobility. Analysis of the distribution of the radius of gyration of the clusters further reveals that the fractal dimension of the ion clusters is nearly independent from Γ for all the cases studied. Finally, at sufficiently high values of Γ, we observed arrested heterogeneous ions mobility, which is correlated with an increase in ion cluster size. These findings provide insight into the role of electrostatics in governing the nanostructures formed by ionomers.

  5. Impact of Facility- and Community-Based Peer Support Models on Maternal Uptake and Retention in Malawi's Option B+ HIV Prevention of Mother-to-Child Transmission Program: A 3-Arm Cluster Randomized Controlled Trial (PURE Malawi).

    PubMed

    Phiri, Sam; Tweya, Hannock; van Lettow, Monique; Rosenberg, Nora E; Trapence, Clement; Kapito-Tembo, Atupele; Kaunda-Khangamwa, Blessings; Kasende, Florence; Kayoyo, Virginia; Cataldo, Fabian; Stanley, Christopher; Gugsa, Salem; Sampathkumar, Veena; Schouten, Erik; Chiwaula, Levison; Eliya, Michael; Chimbwandira, Frank; Hosseinipour, Mina C

    2017-06-01

    Many sub-Saharan African countries have adopted Option B+, a prevention of mother-to-child transmission approach providing HIV-infected pregnant and lactating women with immediate lifelong antiretroviral therapy. High maternal attrition has been observed in Option B+. Peer-based support may improve retention. A 3-arm stratified cluster randomized controlled trial was conducted in Malawi to assess whether facility- and community-based peer support would improve Option B+ uptake and retention compared with standard of care (SOC). In SOC, no enhancements were made (control). In facility-based and community-based models, peers provided patient education, support groups, and patient tracing. Uptake was defined as attending a second scheduled follow-up visit. Retention was defined as being alive and in-care at 2 years without defaulting. Attrition was defined as death, default, or stopping antiretroviral therapy. Generalized estimating equations were used to estimate risk differences (RDs) in uptake. Cox proportional hazards regression with shared frailties was used to estimate hazard of attrition. Twenty-one facilities were randomized and enrolled 1269 women: 447, 428, and 394 in facilities that implemented SOC, facility-based, and community-based peer support models, respectively. Mean age was 27 years. Uptake was higher in facility-based (86%; RD: 6%, confidence interval [CI]: -3% to 15%) and community-based (90%; RD: 9%, CI: 1% to 18%) models compared with SOC (81%). At 24 months, retention was higher in facility-based (80%; RD: 13%, CI: 1% to 26%) and community-based (83%; RD: 16%, CI: 3% to 30%) models compared with SOC (66%). Facility- and community-based peer support interventions can benefit maternal uptake and retention in Option B+.

  6. Mean-cluster approach indicates cell sorting time scales are determined by collective dynamics

    NASA Astrophysics Data System (ADS)

    Beatrici, Carine P.; de Almeida, Rita M. C.; Brunnet, Leonardo G.

    2017-03-01

    Cell migration is essential to cell segregation, playing a central role in tissue formation, wound healing, and tumor evolution. Considering random mixtures of two cell types, it is still not clear which cell characteristics define clustering time scales. The mass of diffusing clusters merging with one another is expected to grow as td /d +2 when the diffusion constant scales with the inverse of the cluster mass. Cell segregation experiments deviate from that behavior. Explanations for that could arise from specific microscopic mechanisms or from collective effects, typical of active matter. Here we consider a power law connecting diffusion constant and cluster mass to propose an analytic approach to model cell segregation where we explicitly take into account finite-size corrections. The results are compared with active matter model simulations and experiments available in the literature. To investigate the role played by different mechanisms we considered different hypotheses describing cell-cell interaction: differential adhesion hypothesis and different velocities hypothesis. We find that the simulations yield normal diffusion for long time intervals. Analytic and simulation results show that (i) cluster evolution clearly tends to a scaling regime, disrupted only at finite-size limits; (ii) cluster diffusion is greatly enhanced by cell collective behavior, such that for high enough tendency to follow the neighbors, cluster diffusion may become independent of cluster size; (iii) the scaling exponent for cluster growth depends only on the mass-diffusion relation, not on the detailed local segregation mechanism. These results apply for active matter systems in general and, in particular, the mechanisms found underlying the increase in cell sorting speed certainly have deep implications in biological evolution as a selection mechanism.

  7. Large scale structure in universes dominated by cold dark matter

    NASA Technical Reports Server (NTRS)

    Bond, J. Richard

    1986-01-01

    The theory of Gaussian random density field peaks is applied to a numerical study of the large-scale structure developing from adiabatic fluctuations in models of biased galaxy formation in universes with Omega = 1, h = 0.5 dominated by cold dark matter (CDM). The angular anisotropy of the cross-correlation function demonstrates that the far-field regions of cluster-scale peaks are asymmetric, as recent observations indicate. These regions will generate pancakes or filaments upon collapse. One-dimensional singularities in the large-scale bulk flow should arise in these CDM models, appearing as pancakes in position space. They are too rare to explain the CfA bubble walls, but pancakes that are just turning around now are sufficiently abundant and would appear to be thin walls normal to the line of sight in redshift space. Large scale streaming velocities are significantly smaller than recent observations indicate. To explain the reported 700 km/s coherent motions, mass must be significantly more clustered than galaxies with a biasing factor of less than 0.4 and a nonlinear redshift at cluster scales greater than one for both massive neutrino and cold models.

  8. Sensitivity Analysis of Multiple Informant Models When Data are Not Missing at Random

    PubMed Central

    Blozis, Shelley A.; Ge, Xiaojia; Xu, Shu; Natsuaki, Misaki N.; Shaw, Daniel S.; Neiderhiser, Jenae; Scaramella, Laura; Leve, Leslie; Reiss, David

    2014-01-01

    Missing data are common in studies that rely on multiple informant data to evaluate relationships among variables for distinguishable individuals clustered within groups. Estimation of structural equation models using raw data allows for incomplete data, and so all groups may be retained even if only one member of a group contributes data. Statistical inference is based on the assumption that data are missing completely at random or missing at random. Importantly, whether or not data are missing is assumed to be independent of the missing data. A saturated correlates model that incorporates correlates of the missingness or the missing data into an analysis and multiple imputation that may also use such correlates offer advantages over the standard implementation of SEM when data are not missing at random because these approaches may result in a data analysis problem for which the missingness is ignorable. This paper considers these approaches in an analysis of family data to assess the sensitivity of parameter estimates to assumptions about missing data, a strategy that may be easily implemented using SEM software. PMID:25221420

  9. Naming games in two-dimensional and small-world-connected random geometric networks.

    PubMed

    Lu, Qiming; Korniss, G; Szymanski, B K

    2008-01-01

    We investigate a prototypical agent-based model, the naming game, on two-dimensional random geometric networks. The naming game [Baronchelli, J. Stat. Mech.: Theory Exp. (2006) P06014] is a minimal model, employing local communications that captures the emergence of shared communication schemes (languages) in a population of autonomous semiotic agents. Implementing the naming games with local broadcasts on random geometric graphs, serves as a model for agreement dynamics in large-scale, autonomously operating wireless sensor networks. Further, it captures essential features of the scaling properties of the agreement process for spatially embedded autonomous agents. Among the relevant observables capturing the temporal properties of the agreement process, we investigate the cluster-size distribution and the distribution of the agreement times, both exhibiting dynamic scaling. We also present results for the case when a small density of long-range communication links are added on top of the random geometric graph, resulting in a "small-world"-like network and yielding a significantly reduced time to reach global agreement. We construct a finite-size scaling analysis for the agreement times in this case.

  10. Evaluation of Seeds of Science/Roots of Reading: Effective Tools for Developing Literacy through Science in the Early Grades-Light Energy Unit. CRESST Report 781

    ERIC Educational Resources Information Center

    Goldschmidt, Pete; Jung, Hyekyung

    2011-01-01

    This evaluation focuses on the Seeds of Science/Roots of Reading: Effective Tools for Developing Literacy through Science in the Early Grades ("Seeds/Roots") model of science-literacy integration. The evaluation is based on a cluster randomized design of 100 teachers, half of which were in the treatment group. Multi-level models are employed to…

  11. Decrease in musculoskeletal pain after 4 and 12 months of an aerobic exercise intervention: a worksite RCT among cleaners.

    PubMed

    Korshøj, Mette; Birk Jørgensen, Marie; Lidegaard, Mark; Mortensen, Ole Steen; Krustrup, Peter; Holtermann, Andreas; Søgaard, Karen

    2017-07-01

    Prevalence of musculoskeletal pain is high in jobs with high physical work demands. An aerobic exercise intervention targeting cardiovascular health was evaluated for its long term side effects on musculoskeletal pain. The objective was to investigate if aerobic exercise affects level of musculoskeletal pain from baseline to 4- and 12-months follow-up. One-hundred-and-sixteen cleaners aged 18-65 years were cluster-randomized. The aerobic exercise group ( n = 57) received worksite aerobic exercise (30 min twice a week) and the reference group ( n = 59) lectures in health promotion. Strata were formed according to closest manager (total 11 strata); clusters were set within strata (total 40 clusters, 20 in each group). Musculoskeletal pain data from eight body regions was collected at baseline and after 4- and 12-months follow-up. The participants stated highest pain in the last month on a scale from 0, stating no pain, up to 10, stating worst possible pain. A repeated-measure 2 × 2 multi-adjusted mixed-models design was applied to compare the between-groups differences in an intention to treat analysis. Participants were entered as a random effect nested in clusters to account for the cluster-based randomization. Clinically significant reductions (>30%, f  2 > 0.25) in the aerobic exercise group, compared to the reference group, in pain intensity in neck, shoulders, arms/wrists were found at 12-months follow-up, and a tendency ( p = 0.07, f  2 = 0.18) to an increase for the knees. At 4-months follow-up the only significant between-group change was an increase in hip pain. This study indicates that aerobic exercise reduces musculoskeletal pain in the upper extremities, but as an unintended side effect may increase pain in the lower extremities. Aerobic exercise interventions among workers standing or walking in the majority of the working hours should tailor exercise to only maintain the positive effect on musculoskeletal pain.

  12. Serological Markers of Sand Fly Exposure to Evaluate Insecticidal Nets against Visceral Leishmaniasis in India and Nepal: A Cluster-Randomized Trial

    PubMed Central

    Gidwani, Kamlesh; Picado, Albert; Rijal, Suman; Singh, Shri Prakash; Roy, Lalita; Volfova, Vera; Andersen, Elisabeth Wreford; Uranw, Surendra; Ostyn, Bart; Sudarshan, Medhavi; Chakravarty, Jaya; Volf, Petr; Sundar, Shyam; Boelaert, Marleen; Rogers, Matthew Edward

    2011-01-01

    Background Visceral leishmaniasis is the world' second largest vector-borne parasitic killer and a neglected tropical disease, prevalent in poor communities. Long-lasting insecticidal nets (LNs) are a low cost proven vector intervention method for malaria control; however, their effectiveness against visceral leishmaniasis (VL) is unknown. This study quantified the effect of LNs on exposure to the sand fly vector of VL in India and Nepal during a two year community intervention trial. Methods As part of a paired-cluster randomized controlled clinical trial in VL-endemic regions of India and Nepal we tested the effect of LNs on sand fly biting by measuring the antibody response of subjects to the saliva of Leishmania donovani vector Phlebotomus argentipes and the sympatric (non-vector) Phlebotomus papatasi. Fifteen to 20 individuals above 15 years of age from 26 VL endemic clusters were asked to provide a blood sample at baseline, 12 and 24 months post-intervention. Results A total of 305 individuals were included in the study, 68 participants provided two blood samples and 237 gave three samples. A random effect linear regression model showed that cluster-wide distribution of LNs reduced exposure to P. argentipes by 12% at 12 months (effect 0.88; 95% CI 0.83–0.94) and 9% at 24 months (effect 0.91; 95% CI 0.80–1.02) in the intervention group compared to control adjusting for baseline values and pair. Similar results were obtained for P. papatasi. Conclusions This trial provides evidence that LNs have a limited effect on sand fly exposure in VL endemic communities in India and Nepal and supports the use of sand fly saliva antibodies as a marker to evaluate vector control interventions. PMID:21931871

  13. Serological markers of sand fly exposure to evaluate insecticidal nets against visceral leishmaniasis in India and Nepal: a cluster-randomized trial.

    PubMed

    Gidwani, Kamlesh; Picado, Albert; Rijal, Suman; Singh, Shri Prakash; Roy, Lalita; Volfova, Vera; Andersen, Elisabeth Wreford; Uranw, Surendra; Ostyn, Bart; Sudarshan, Medhavi; Chakravarty, Jaya; Volf, Petr; Sundar, Shyam; Boelaert, Marleen; Rogers, Matthew Edward

    2011-09-01

    Visceral leishmaniasis is the world' second largest vector-borne parasitic killer and a neglected tropical disease, prevalent in poor communities. Long-lasting insecticidal nets (LNs) are a low cost proven vector intervention method for malaria control; however, their effectiveness against visceral leishmaniasis (VL) is unknown. This study quantified the effect of LNs on exposure to the sand fly vector of VL in India and Nepal during a two year community intervention trial. As part of a paired-cluster randomized controlled clinical trial in VL-endemic regions of India and Nepal we tested the effect of LNs on sand fly biting by measuring the antibody response of subjects to the saliva of Leishmania donovani vector Phlebotomus argentipes and the sympatric (non-vector) Phlebotomus papatasi. Fifteen to 20 individuals above 15 years of age from 26 VL endemic clusters were asked to provide a blood sample at baseline, 12 and 24 months post-intervention. A total of 305 individuals were included in the study, 68 participants provided two blood samples and 237 gave three samples. A random effect linear regression model showed that cluster-wide distribution of LNs reduced exposure to P. argentipes by 12% at 12 months (effect 0.88; 95% CI 0.83-0.94) and 9% at 24 months (effect 0.91; 95% CI 0.80-1.02) in the intervention group compared to control adjusting for baseline values and pair. Similar results were obtained for P. papatasi. This trial provides evidence that LNs have a limited effect on sand fly exposure in VL endemic communities in India and Nepal and supports the use of sand fly saliva antibodies as a marker to evaluate vector control interventions.

  14. Clustering and phase transitions on a neutral landscape

    NASA Astrophysics Data System (ADS)

    Scott, Adam D.; King, Dawn M.; Marić, Nevena; Bahar, Sonya

    2013-06-01

    Recent computational studies have shown that speciation can occur under neutral conditions, i.e., when the simulated organisms all have identical fitness. These works bear comparison with mathematical studies of clustering on neutral landscapes in the context of branching and coalescing random walks. Here, we show that sympatric clustering/speciation can occur on a neutral landscape whose dimensions specify only the simulated organisms’ phenotypes. We demonstrate that clustering occurs not only in the case of assortative mating, but also in the case of asexual fission; it is not observed in the control case of random mating. We find that the population size and the number of clusters undergo a second-order non-equilibrium phase transition as the maximum mutation size is varied.

  15. Reporting and methodological quality of sample size calculations in cluster randomized trials could be improved: a review.

    PubMed

    Rutterford, Clare; Taljaard, Monica; Dixon, Stephanie; Copas, Andrew; Eldridge, Sandra

    2015-06-01

    To assess the quality of reporting and accuracy of a priori estimates used in sample size calculations for cluster randomized trials (CRTs). We reviewed 300 CRTs published between 2000 and 2008. The prevalence of reporting sample size elements from the 2004 CONSORT recommendations was evaluated and a priori estimates compared with those observed in the trial. Of the 300 trials, 166 (55%) reported a sample size calculation. Only 36 of 166 (22%) reported all recommended descriptive elements. Elements specific to CRTs were the worst reported: a measure of within-cluster correlation was specified in only 58 of 166 (35%). Only 18 of 166 articles (11%) reported both a priori and observed within-cluster correlation values. Except in two cases, observed within-cluster correlation values were either close to or less than a priori values. Even with the CONSORT extension for cluster randomization, the reporting of sample size elements specific to these trials remains below that necessary for transparent reporting. Journal editors and peer reviewers should implement stricter requirements for authors to follow CONSORT recommendations. Authors should report observed and a priori within-cluster correlation values to enable comparisons between these over a wider range of trials. Copyright © 2015 The Authors. Published by Elsevier Inc. All rights reserved.

  16. [Analysis of Time-to-onset of Interstitial Lung Disease after the Administration of Small Molecule Molecularly-targeted Drugs].

    PubMed

    Komada, Fusao

    2018-01-01

     The aim of this study was to investigate the time-to-onset of drug-induced interstitial lung disease (DILD) following the administration of small molecule molecularly-targeted drugs via the use of the spontaneous adverse reaction reporting system of the Japanese Adverse Drug Event Report database. DILD datasets for afatinib, alectinib, bortezomib, crizotinib, dasatinib, erlotinib, everolimus, gefitinib, imatinib, lapatinib, nilotinib, osimertinib, sorafenib, sunitinib, temsirolimus, and tofacitinib were used to calculate the median onset times of DILD and the Weibull distribution parameters, and to perform the hierarchical cluster analysis. The median onset times of DILD for afatinib, bortezomib, crizotinib, erlotinib, gefitinib, and nilotinib were within one month. The median onset times of DILD for dasatinib, everolimus, lapatinib, osimertinib, and temsirolimus ranged from 1 to 2 months. The median onset times of the DILD for alectinib, imatinib, and tofacitinib ranged from 2 to 3 months. The median onset times of the DILD for sunitinib and sorafenib ranged from 8 to 9 months. Weibull distributions for these drugs when using the cluster analysis showed that there were 4 clusters. Cluster 1 described a subgroup with early to later onset DILD and early failure type profiles or a random failure type profile. Cluster 2 exhibited early failure type profiles or a random failure type profile with early onset DILD. Cluster 3 exhibited a random failure type profile or wear out failure type profiles with later onset DILD. Cluster 4 exhibited an early failure type profile or a random failure type profile with the latest onset DILD.

  17. Multilevel Multidimensional Item Response Model with a Multilevel Latent Covariate

    ERIC Educational Resources Information Center

    Cho, Sun-Joo; Bottge, Brian A.

    2015-01-01

    In a pretest-posttest cluster-randomized trial, one of the methods commonly used to detect an intervention effect involves controlling pre-test scores and other related covariates while estimating an intervention effect at post-test. In many applications in education, the total post-test and pre-test scores that ignores measurement error in the…

  18. Online versus Face-to-Face Training of Critical Time Intervention: A Matching Cluster Randomized Trial

    ERIC Educational Resources Information Center

    Olivet, Jeffrey; Zerger, Suzanne; Greene, R. Neil; Kenney, Rachael R.; Herman, Daniel B.

    2016-01-01

    This study examined the effectiveness of online education to providers who serve people experiencing homelessness, comparing online and face-to-face training of Critical Time Intervention (CTI), an evidence-based case management model. The authors recruited 184 staff from nineteen homeless service agencies to participate in one of two training…

  19. Predicting the random drift of MEMS gyroscope based on K-means clustering and OLS RBF Neural Network

    NASA Astrophysics Data System (ADS)

    Wang, Zhen-yu; Zhang, Li-jie

    2017-10-01

    Measure error of the sensor can be effectively compensated with prediction. Aiming at large random drift error of MEMS(Micro Electro Mechanical System))gyroscope, an improved learning algorithm of Radial Basis Function(RBF) Neural Network(NN) based on K-means clustering and Orthogonal Least-Squares (OLS) is proposed in this paper. The algorithm selects the typical samples as the initial cluster centers of RBF NN firstly, candidates centers with K-means algorithm secondly, and optimizes the candidate centers with OLS algorithm thirdly, which makes the network structure simpler and makes the prediction performance better. Experimental results show that the proposed K-means clustering OLS learning algorithm can predict the random drift of MEMS gyroscope effectively, the prediction error of which is 9.8019e-007°/s and the prediction time of which is 2.4169e-006s

  20. Reliability Evaluation for Clustered WSNs under Malware Propagation

    PubMed Central

    Shen, Shigen; Huang, Longjun; Liu, Jianhua; Champion, Adam C.; Yu, Shui; Cao, Qiying

    2016-01-01

    We consider a clustered wireless sensor network (WSN) under epidemic-malware propagation conditions and solve the problem of how to evaluate its reliability so as to ensure efficient, continuous, and dependable transmission of sensed data from sensor nodes to the sink. Facing the contradiction between malware intention and continuous-time Markov chain (CTMC) randomness, we introduce a strategic game that can predict malware infection in order to model a successful infection as a CTMC state transition. Next, we devise a novel measure to compute the Mean Time to Failure (MTTF) of a sensor node, which represents the reliability of a sensor node continuously performing tasks such as sensing, transmitting, and fusing data. Since clustered WSNs can be regarded as parallel-serial-parallel systems, the reliability of a clustered WSN can be evaluated via classical reliability theory. Numerical results show the influence of parameters such as the true positive rate and the false positive rate on a sensor node’s MTTF. Furthermore, we validate the method of reliability evaluation for a clustered WSN according to the number of sensor nodes in a cluster, the number of clusters in a route, and the number of routes in the WSN. PMID:27294934

  1. Reliability Evaluation for Clustered WSNs under Malware Propagation.

    PubMed

    Shen, Shigen; Huang, Longjun; Liu, Jianhua; Champion, Adam C; Yu, Shui; Cao, Qiying

    2016-06-10

    We consider a clustered wireless sensor network (WSN) under epidemic-malware propagation conditions and solve the problem of how to evaluate its reliability so as to ensure efficient, continuous, and dependable transmission of sensed data from sensor nodes to the sink. Facing the contradiction between malware intention and continuous-time Markov chain (CTMC) randomness, we introduce a strategic game that can predict malware infection in order to model a successful infection as a CTMC state transition. Next, we devise a novel measure to compute the Mean Time to Failure (MTTF) of a sensor node, which represents the reliability of a sensor node continuously performing tasks such as sensing, transmitting, and fusing data. Since clustered WSNs can be regarded as parallel-serial-parallel systems, the reliability of a clustered WSN can be evaluated via classical reliability theory. Numerical results show the influence of parameters such as the true positive rate and the false positive rate on a sensor node's MTTF. Furthermore, we validate the method of reliability evaluation for a clustered WSN according to the number of sensor nodes in a cluster, the number of clusters in a route, and the number of routes in the WSN.

  2. The effects of bilingual language proficiency on recall accuracy and semantic clustering in free recall output: evidence for shared semantic associations across languages.

    PubMed

    Francis, Wendy S; Taylor, Randolph S; Gutiérrez, Marisela; Liaño, Mary K; Manzanera, Diana G; Penalver, Renee M

    2018-05-19

    Two experiments investigated how well bilinguals utilise long-standing semantic associations to encode and retrieve semantic clusters in verbal episodic memory. In Experiment 1, Spanish-English bilinguals (N = 128) studied and recalled word and picture sets. Word recall was equivalent in L1 and L2, picture recall was better in L1 than in L2, and the picture superiority effect was stronger in L1 than in L2. Semantic clustering in word and picture recall was equivalent in L1 and L2. In Experiment 2, Spanish-English bilinguals (N = 128) and English-speaking monolinguals (N = 128) studied and recalled word sequences that contained semantically related pairs. Data were analyzed using a multinomial processing tree approach, the pair-clustering model. Cluster formation was more likely for semantically organised than for randomly ordered word sequences. Probabilities of cluster formation, cluster retrieval, and retrieval of unclustered items did not differ across languages or language groups. Language proficiency has little if any impact on the utilisation of long-standing semantic associations, which are language-general.

  3. Sample Size Estimation in Cluster Randomized Educational Trials: An Empirical Bayes Approach

    ERIC Educational Resources Information Center

    Rotondi, Michael A.; Donner, Allan

    2009-01-01

    The educational field has now accumulated an extensive literature reporting on values of the intraclass correlation coefficient, a parameter essential to determining the required size of a planned cluster randomized trial. We propose here a simple simulation-based approach including all relevant information that can facilitate this task. An…

  4. The Effectiveness of Healthy Start Home Visit Program: Cluster Randomized Controlled Trial

    ERIC Educational Resources Information Center

    Leung, Cynthia; Tsang, Sandra; Heung, Kitty

    2015-01-01

    Purpose: The study reported the effectiveness of a home visit program for disadvantaged Chinese parents with preschool children, using cluster randomized controlled trial design. Method: Participants included 191 parents and their children from 24 preschools, with 84 dyads (12 preschools) in the intervention group and 107 dyads (12 preschools) in…

  5. Standardized Effect Size Measures for Mediation Analysis in Cluster-Randomized Trials

    ERIC Educational Resources Information Center

    Stapleton, Laura M.; Pituch, Keenan A.; Dion, Eric

    2015-01-01

    This article presents 3 standardized effect size measures to use when sharing results of an analysis of mediation of treatment effects for cluster-randomized trials. The authors discuss 3 examples of mediation analysis (upper-level mediation, cross-level mediation, and cross-level mediation with a contextual effect) with demonstration of the…

  6. Intraclass Correlations and Covariate Outcome Correlations for Planning Two-and Three-Level Cluster-Randomized Experiments in Education

    ERIC Educational Resources Information Center

    Hedges, Larry V.; Hedberg, E. C.

    2013-01-01

    Background: Cluster-randomized experiments that assign intact groups such as schools or school districts to treatment conditions are increasingly common in educational research. Such experiments are inherently multilevel designs whose sensitivity (statistical power and precision of estimates) depends on the variance decomposition across levels.…

  7. Fit 5 Kids TV reduction program for Latino preschoolers: A cluster randomized controlled trial

    USDA-ARS?s Scientific Manuscript database

    Reducing Latino preschoolers' TV viewing is needed to reduce their risk of obesity and other chronic diseases. This study's objective was to evaluate the Fit 5 Kids (F5K) TV reduction program's impact on Latino preschooler's TV viewing. The study design was a cluster randomized controlled trial (RCT...

  8. Intraclass Correlations and Covariate Outcome Correlations for Planning 2 and 3 Level Cluster Randomized Experiments in Education

    ERIC Educational Resources Information Center

    Hedges, Larry V.; Hedberg, Eric C.

    2013-01-01

    Background: Cluster randomized experiments that assign intact groups such as schools or school districts to treatment conditions are increasingly common in educational research. Such experiments are inherently multilevel designs whose sensitivity (statistical power and precision of estimates) depends on the variance decomposition across levels.…

  9. Infrared Extinction Performance of Randomly Oriented Microbial-Clustered Agglomerate Materials.

    PubMed

    Li, Le; Hu, Yihua; Gu, Youlin; Zhao, Xinying; Xu, Shilong; Yu, Lei; Zheng, Zhi Ming; Wang, Peng

    2017-11-01

    In this study, the spatial structure of randomly distributed clusters of fungi An0429 spores was simulated using a cluster aggregation (CCA) model, and the single scattering parameters of fungi An0429 spores were calculated using the discrete dipole approximation (DDA) method. The transmittance of 10.6 µm infrared (IR) light in the aggregated fungi An0429 spores swarm is simulated by using the Monte Carlo method. Several parameters that affect the transmittance of 10.6 µm IR light, such as the number and radius of original fungi An0429 spores, porosity of aggregated fungi An0429 spores, and density of aggregated fungi An0429 spores of the formation aerosol area were discussed. Finally, the transmittances of microbial materials with different qualities were measured in the dynamic test platform. The simulation results showed that the parameters analyzed were closely connected with the extinction performance of fungi An0429 spores. By controlling the value of the influencing factors, the transmittance could be lower than a certain threshold to meet the requirement of attenuation in application. In addition, the experimental results showed that the Monte Carlo method could well reflect the attenuation law of IR light in fungi An0429 spore agglomerates swarms.

  10. Effective spreading from multiple leaders identified by percolation in the susceptible-infected-recovered (SIR) model

    NASA Astrophysics Data System (ADS)

    Ji, Shenggong; Lü, Linyuan; Yeung, Chi Ho; Hu, Yanqing

    2017-07-01

    Social networks constitute a new platform for information propagation, but its success is crucially dependent on the choice of spreaders who initiate the spreading of information. In this paper, we remove edges in a network at random and the network segments into isolated clusters. The most important nodes in each cluster then form a set of influential spreaders, such that news propagating from them would lead to extensive coverage and minimal redundancy. The method utilizes the similarities between the segmented networks before percolation and the coverage of information propagation in each social cluster to obtain a set of distributed and coordinated spreaders. Our tests of implementing the susceptible-infected-recovered model on Facebook and Enron email networks show that this method outperforms conventional centrality-based methods in terms of spreadability and coverage redundancy. The suggested way of identifying influential spreaders thus sheds light on a new paradigm of information propagation in social networks.

  11. Anomalously slow relaxation of the system of liquid clusters in a disordered nanoporous medium according to the self-organized criticality scenario

    NASA Astrophysics Data System (ADS)

    Borman, V. D.; Tronin, V. N.; Byrkin, V. A.

    2016-04-01

    We propose a physical model of a relaxation of states of clusters of nonwetting liquid confined in a random nanoporous medium. The relaxation is occurred by the self-organized criticality (SOC) scenario. Process is characterized by waiting for fluctuation necessary for overcoming of a local energy barrier with the subsequent avalanche hydrodynamic extrusion of the liquid by surface forces of the nonwetting frame. The dependence of the interaction between local configurations on the number of filled pores belonging to the infinite percolation cluster of filled pores serves as an internal feedback initiating the SOC process. The calculations give a power-law time dependence of the relative volume θ of the confined liquid θ ∼t-ν (ν ∼ 0.2) as in the picture of relaxation in the mean field approximation. The model of the relaxation of the porous medium with the nonwetting liquid demonstrates possible mechanisms and scenarios of SOC for relaxation of other disordered systems.

  12. New method for estimating clustering of DNA lesions induced by physical/chemical mutagens using fluorescence anisotropy.

    PubMed

    Akamatsu, Ken; Shikazono, Naoya; Saito, Takeshi

    2017-11-01

    We have developed a new method for estimating the localization of DNA damage such as apurinic/apyrimidinic sites (APs) on DNA using fluorescence anisotropy. This method is aimed at characterizing clustered DNA damage produced by DNA-damaging agents such as ionizing radiation and genotoxic chemicals. A fluorescent probe with an aminooxy group (AlexaFluor488) was used to label APs. We prepared a pUC19 plasmid with APs by heating under acidic conditions as a model for damaged DNA, and subsequently labeled the APs. We found that the observed fluorescence anisotropy (r obs ) decreases as averaged AP density (λ AP : number of APs per base pair) increases due to homo-FRET, and that the APs were randomly distributed. We applied this method to three DNA-damaging agents, 60 Co γ-rays, methyl methanesulfonate (MMS), and neocarzinostatin (NCS). We found that r obs -λ AP relationships differed significantly between MMS and NCS. At low AP density (λ AP  < 0.001), the APs induced by MMS seemed to not be closely distributed, whereas those induced by NCS were remarkably clustered. In contrast, the AP clustering induced by 60 Co γ-rays was similar to, but potentially more likely to occur than, random distribution. This simple method can be used to estimate mutagenicity of ionizing radiation and genotoxic chemicals. Copyright © 2017 Elsevier Inc. All rights reserved.

  13. Bayesian Nonparametric Inference – Why and How

    PubMed Central

    Müller, Peter; Mitra, Riten

    2013-01-01

    We review inference under models with nonparametric Bayesian (BNP) priors. The discussion follows a set of examples for some common inference problems. The examples are chosen to highlight problems that are challenging for standard parametric inference. We discuss inference for density estimation, clustering, regression and for mixed effects models with random effects distributions. While we focus on arguing for the need for the flexibility of BNP models, we also review some of the more commonly used BNP models, thus hopefully answering a bit of both questions, why and how to use BNP. PMID:24368932

  14. A Cluster Randomized Evaluation of a Health Department Data to Care Intervention Designed to Increase Engagement in HIV Care and Antiretroviral Use.

    PubMed

    Dombrowski, Julia C; Hughes, James P; Buskin, Susan E; Bennett, Amy; Katz, David; Fleming, Mark; Nunez, Angela; Golden, Matthew R

    2018-06-01

    Many US health departments have implemented Data to Care interventions, which use HIV surveillance data to identify persons who are inadequately engaged in HIV medical care and assist them with care reengagement, but the effectiveness of this strategy is uncertain. We conducted a stepped-wedge, cluster-randomized evaluation of a Data to Care intervention in King County, Washington, 2011 to 2014. Persons diagnosed as having HIV for at least 6 months were eligible based on 1 of 2 criteria: (1) viral load (VL) greater than 500 copies/mL and CD4 less than 350 cells/μL at the last report in the past 12 months or (2) no CD4 or VL reported to the health department for at least 12 months. The intervention included medical provider contact, patient contact, and a structured individual interview. Health department staff assisted patients with reengagement using health systems navigation, brief counseling, and referral to support services. We clustered all eligible cases in the county by the last known medical provider and randomized the order of clusters for intervention, creating contemporaneous intervention and control periods (cases in later clusters contributed person-time to the control period at the same time that cases in earlier clusters contributed person-time to the intervention period). We compared the time to viral suppression (VL <200 copies/mL) for individuals during intervention and control periods using a Cox proportional hazards model. We identified 997 persons (intention to treat [ITT]), 18% of whom had moved or died. Of the remaining 822 (modified ITT), 161 (20%) had an undetectable VL reported before contact and 164 (20%) completed the individual interview. The hazard ratio (HR) for time to viral suppression did not differ between the intervention and control periods in ITT (HR, 1.21 [95% confidence interval, 0.85-1.71]) or modified ITT (HR, 1.18 [95% confidence interval, 0.83-1.68]) analysis. The Data to Care intervention did not impact time to viral suppression.

  15. Assessing the feasibility of interrupting the transmission of soil-transmitted helminths through mass drug administration: The DeWorm3 cluster randomized trial protocol

    PubMed Central

    Ajjampur, Sitara S. Rao; Anderson, Roy M.; Bailey, Robin; Gardiner, Iain; Halliday, Katherine E.; Ibikounle, Moudachirou; Kalua, Khumbo; Kang, Gagandeep; Littlewood, D. Timothy J.; Luty, Adrian J. F.; Means, Arianna Rubin; Oswald, William; Pullan, Rachel L.; Sarkar, Rajiv; Schär, Fabian; Szpiro, Adam; Truscott, James E.; Werkman, Marleen; Yard, Elodie; Walson, Judd L.

    2018-01-01

    Current control strategies for soil-transmitted helminths (STH) emphasize morbidity control through mass drug administration (MDA) targeting preschool- and school-age children, women of childbearing age and adults in certain high-risk occupations such as agricultural laborers or miners. This strategy is effective at reducing morbidity in those treated but, without massive economic development, it is unlikely it will interrupt transmission. MDA will therefore need to continue indefinitely to maintain benefit. Mathematical models suggest that transmission interruption may be achievable through MDA alone, provided that all age groups are targeted with high coverage. The DeWorm3 Project will test the feasibility of interrupting STH transmission using biannual MDA targeting all age groups. Study sites (population ≥80,000) have been identified in Benin, Malawi and India. Each site will be divided into 40 clusters, to be randomized 1:1 to three years of twice-annual community-wide MDA or standard-of-care MDA, typically annual school-based deworming. Community-wide MDA will be delivered door-to-door, while standard-of-care MDA will be delivered according to national guidelines. The primary outcome is transmission interruption of the STH species present at each site, defined as weighted cluster-level prevalence ≤2% by quantitative polymerase chain reaction (qPCR), 24 months after the final round of MDA. Secondary outcomes include the endline prevalence of STH, overall and by species, and the endline prevalence of STH among children under five as an indicator of incident infections. Secondary analyses will identify cluster-level factors associated with transmission interruption. Prevalence will be assessed using qPCR of stool samples collected from a random sample of cluster residents at baseline, six months after the final round of MDA and 24 months post-MDA. A smaller number of individuals in each cluster will be followed with annual sampling to monitor trends in prevalence and reinfection throughout the trial. Trial registration ClinicalTrials.gov NCT03014167 PMID:29346377

  16. Testing feedback message framing and comparators to address prescribing of high-risk medications in nursing homes: protocol for a pragmatic, factorial, cluster-randomized trial.

    PubMed

    Ivers, Noah M; Desveaux, Laura; Presseau, Justin; Reis, Catherine; Witteman, Holly O; Taljaard, Monica K; McCleary, Nicola; Thavorn, Kednapa; Grimshaw, Jeremy M

    2017-07-14

    Audit and feedback (AF) interventions that leverage routine administrative data offer a scalable and relatively low-cost method to improve processes of care. AF interventions are usually designed to highlight discrepancies between desired and actual performance and to encourage recipients to act to address such discrepancies. Comparing to a regional average is a common approach, but more recipients would have a discrepancy if compared to a higher-than-average level of performance. In addition, how recipients perceive and respond to discrepancies may depend on how the feedback itself is framed. We aim to evaluate the effectiveness of different comparators and framing in feedback on high-risk prescribing in nursing homes. This is a pragmatic, 2 × 2 factorial, cluster-randomized controlled trial testing variations in the comparator and framing on the effectiveness of quarterly AF in changing high-risk prescribing in nursing homes in Ontario, Canada. We grouped homes that share physicians into clusters and randomized these clusters into the four experimental conditions. Outcomes will be assessed after 6 months; all primary analyses will be by intention-to-treat. The primary outcome (monthly number of high-risk medications received by each patient) will be analysed using a general linear mixed effects regression model. We will present both four-arm and factorial analyses. With 160 clusters and an average of 350 beds per cluster, assuming no interaction and similar effects for each intervention, we anticipate 90% power to detect an absolute mean difference of 0.3 high-risk medications prescribed. A mixed-methods process evaluation will explore potential mechanisms underlying the observed effects, exploring targeted constructs including intention, self-efficacy, outcome expectations, descriptive norms, and goal prioritization. An economic analysis will examine cost-effectiveness analysis from the perspective of the publicly funded health care system. This protocol describes the rationale and methodology of a trial testing manipulations of theory-informed components of an audit and feedback intervention to determine how to improve an existing intervention and provide generalizable insights for implementation science. NCT02979964.

  17. A semiparametric Bayesian proportional hazards model for interval censored data with frailty effects.

    PubMed

    Henschel, Volkmar; Engel, Jutta; Hölzel, Dieter; Mansmann, Ulrich

    2009-02-10

    Multivariate analysis of interval censored event data based on classical likelihood methods is notoriously cumbersome. Likelihood inference for models which additionally include random effects are not available at all. Developed algorithms bear problems for practical users like: matrix inversion, slow convergence, no assessment of statistical uncertainty. MCMC procedures combined with imputation are used to implement hierarchical models for interval censored data within a Bayesian framework. Two examples from clinical practice demonstrate the handling of clustered interval censored event times as well as multilayer random effects for inter-institutional quality assessment. The software developed is called survBayes and is freely available at CRAN. The proposed software supports the solution of complex analyses in many fields of clinical epidemiology as well as health services research.

  18. Study protocol of Prednisone in episodic Cluster Headache (PredCH): a randomized, double-blind, placebo-controlled parallel group trial to evaluate the efficacy and safety of oral prednisone as an add-on therapy in the prophylactic treatment of episodic cluster headache with verapamil

    PubMed Central

    2013-01-01

    Background Episodic cluster headache (ECH) is a primary headache disorder that severely impairs patient’s quality of life. First-line therapy in the initiation of a prophylactic treatment is verapamil. Due to its delayed onset of efficacy and the necessary slow titration of dosage for tolerability reasons prednisone is frequently added by clinicians to the initial prophylactic treatment of a cluster episode. This treatment strategy is thought to effectively reduce the number and intensity of cluster attacks in the beginning of a cluster episode (before verapamil is effective). This study will assess the efficacy and safety of oral prednisone as an add-on therapy to verapamil and compare it to a monotherapy with verapamil in the initial prophylactic treatment of a cluster episode. Methods and design PredCH is a prospective, randomized, double-blind, placebo-controlled trial with parallel study arms. Eligible patients with episodic cluster headache will be randomized to a treatment intervention with prednisone or a placebo arm. The multi-center trial will be conducted in eight German headache clinics that specialize in the treatment of ECH. Discussion PredCH is designed to assess whether oral prednisone added to first-line agent verapamil helps reduce the number and intensity of cluster attacks in the beginning of a cluster episode as compared to monotherapy with verapamil. Trial registration German Clinical Trials Register DRKS00004716 PMID:23889923

  19. Accelerating Information Retrieval from Profile Hidden Markov Model Databases.

    PubMed

    Tamimi, Ahmad; Ashhab, Yaqoub; Tamimi, Hashem

    2016-01-01

    Profile Hidden Markov Model (Profile-HMM) is an efficient statistical approach to represent protein families. Currently, several databases maintain valuable protein sequence information as profile-HMMs. There is an increasing interest to improve the efficiency of searching Profile-HMM databases to detect sequence-profile or profile-profile homology. However, most efforts to enhance searching efficiency have been focusing on improving the alignment algorithms. Although the performance of these algorithms is fairly acceptable, the growing size of these databases, as well as the increasing demand for using batch query searching approach, are strong motivations that call for further enhancement of information retrieval from profile-HMM databases. This work presents a heuristic method to accelerate the current profile-HMM homology searching approaches. The method works by cluster-based remodeling of the database to reduce the search space, rather than focusing on the alignment algorithms. Using different clustering techniques, 4284 TIGRFAMs profiles were clustered based on their similarities. A representative for each cluster was assigned. To enhance sensitivity, we proposed an extended step that allows overlapping among clusters. A validation benchmark of 6000 randomly selected protein sequences was used to query the clustered profiles. To evaluate the efficiency of our approach, speed and recall values were measured and compared with the sequential search approach. Using hierarchical, k-means, and connected component clustering techniques followed by the extended overlapping step, we obtained an average reduction in time of 41%, and an average recall of 96%. Our results demonstrate that representation of profile-HMMs using a clustering-based approach can significantly accelerate data retrieval from profile-HMM databases.

  20. Random isotropic one-dimensional XY-model

    NASA Astrophysics Data System (ADS)

    Gonçalves, L. L.; Vieira, A. P.

    1998-01-01

    The 1D isotropic s = ½XY-model ( N sites), with random exchange interaction in a transverse random field is considered. The random variables satisfy bimodal quenched distributions. The solution is obtained by using the Jordan-Wigner fermionization and a canonical transformation, reducing the problem to diagonalizing an N × N matrix, corresponding to a system of N noninteracting fermions. The calculations are performed numerically for N = 1000, and the field-induced magnetization at T = 0 is obtained by averaging the results for the different samples. For the dilute case, in the uniform field limit, the magnetization exhibits various discontinuities, which are the consequence of the existence of disconnected finite clusters distributed along the chain. Also in this limit, for finite exchange constants J A and J B, as the probability of J A varies from one to zero, the saturation field is seen to vary from Γ A to Γ B, where Γ A(Γ B) is the value of the saturation field for the pure case with exchange constant equal to J A(J B) .

  1. Robust Bayesian clustering.

    PubMed

    Archambeau, Cédric; Verleysen, Michel

    2007-01-01

    A new variational Bayesian learning algorithm for Student-t mixture models is introduced. This algorithm leads to (i) robust density estimation, (ii) robust clustering and (iii) robust automatic model selection. Gaussian mixture models are learning machines which are based on a divide-and-conquer approach. They are commonly used for density estimation and clustering tasks, but are sensitive to outliers. The Student-t distribution has heavier tails than the Gaussian distribution and is therefore less sensitive to any departure of the empirical distribution from Gaussianity. As a consequence, the Student-t distribution is suitable for constructing robust mixture models. In this work, we formalize the Bayesian Student-t mixture model as a latent variable model in a different way from Svensén and Bishop [Svensén, M., & Bishop, C. M. (2005). Robust Bayesian mixture modelling. Neurocomputing, 64, 235-252]. The main difference resides in the fact that it is not necessary to assume a factorized approximation of the posterior distribution on the latent indicator variables and the latent scale variables in order to obtain a tractable solution. Not neglecting the correlations between these unobserved random variables leads to a Bayesian model having an increased robustness. Furthermore, it is expected that the lower bound on the log-evidence is tighter. Based on this bound, the model complexity, i.e. the number of components in the mixture, can be inferred with a higher confidence.

  2. The role of gender in a smoking cessation intervention: a cluster randomized clinical trial

    PubMed Central

    2011-01-01

    Background The prevalence of smoking in Spain is high in both men and women. The aim of our study was to evaluate the role of gender in the effectiveness of a specific smoking cessation intervention conducted in Spain. Methods This study was a secondary analysis of a cluster randomized clinical trial in which the randomization unit was the Basic Care Unit (family physician and nurse who care for the same group of patients). The intervention consisted of a six-month period of implementing the recommendations of a Clinical Practice Guideline. A total of 2,937 current smokers at 82 Primary Care Centers in 13 different regions of Spain were included (2003-2005). The success rate was measured by a six-month continued abstinence rate at the one-year follow-up. A logistic mixed-effects regression model, taking Basic Care Units as random-effect parameter, was performed in order to analyze gender as a predictor of smoking cessation. Results At the one-year follow-up, the six-month continuous abstinence quit rate was 9.4% in men and 8.5% in women (p = 0.400). The logistic mixed-effects regression model showed that women did not have a higher odds of being an ex-smoker than men after the analysis was adjusted for confounders (OR adjusted = 0.9, 95% CI = 0.7-1.2). Conclusions Gender does not appear to be a predictor of smoking cessation at the one-year follow-up in individuals presenting at Primary Care Centers. ClinicalTrials.gov Identifier NCT00125905. PMID:21605389

  3. Transportability of an Evidence-Based Early Childhood Intervention in a Low-Income African Country: Results of a Cluster Randomized Controlled Study.

    PubMed

    Huang, Keng-Yen; Nakigudde, Janet; Rhule, Dana; Gumikiriza-Onoria, Joy Louise; Abura, Gloria; Kolawole, Bukky; Ndyanabangi, Sheila; Kim, Sharon; Seidman, Edward; Ogedegbe, Gbenga; Brotman, Laurie Miller

    2017-11-01

    Children in Sub-Saharan Africa (SSA) are burdened by significant unmet mental health needs. Despite the successes of numerous school-based interventions for promoting child mental health, most evidence-based interventions (EBIs) are not available in SSA. This study investigated the implementation quality and effectiveness of one component of an EBI from a developed country (USA) in a SSA country (Uganda). The EBI component, Professional Development, was provided by trained Ugandan mental health professionals to Ugandan primary school teachers. It included large-group experiential training and small-group coaching to introduce and support a range of evidence-based practices (EBPs) to create nurturing and predictable classroom experiences. The study was guided by the Consolidated Framework for Implementation Research, the Teacher Training Implementation Model, and the RE-AIM evaluation framework. Effectiveness outcomes were studied using a cluster randomized design, in which 10 schools were randomized to intervention and wait-list control conditions. A total of 79 early childhood teachers participated. Teacher knowledge and the use of EBPs were assessed at baseline and immediately post-intervention (4-5 months later). A sample of 154 parents was randomly selected to report on child behavior at baseline and post-intervention. Linear mixed effect modeling was applied to examine effectiveness outcomes. Findings support the feasibility of training Ugandan mental health professionals to provide Professional Development for Ugandan teachers. Professional Development was delivered with high levels of fidelity and resulted in improved teacher EBP knowledge and the use of EBPs in the classroom, and child social competence.

  4. Assessing the interruption of the transmission of human helminths with mass drug administration alone: optimizing the design of cluster randomized trials.

    PubMed

    Anderson, Roy; Farrell, Sam; Turner, Hugo; Walson, Judd; Donnelly, Christl A; Truscott, James

    2017-02-17

    A method is outlined for the use of an individual-based stochastic model of parasite transmission dynamics to assess different designs for a cluster randomized trial in which mass drug administration (MDA) is employed in attempts to eliminate the transmission of soil-transmitted helminths (STH) in defined geographic locations. The hypothesis to be tested is: Can MDA alone interrupt the transmission of STH species in defined settings? Clustering is at a village level and the choice of clusters of villages is stratified by transmission intensity (low, medium and high) and parasite species mix (either Ascaris, Trichuris or hookworm dominant). The methodological approach first uses an age-structured deterministic model to predict the MDA coverage required for treating pre-school aged children (Pre-SAC), school aged children (SAC) and adults (Adults) to eliminate transmission (crossing the breakpoint in transmission created by sexual mating in dioecious helminths) with 3 rounds of annual MDA. Stochastic individual-based models are then used to calculate the positive and negative predictive values (PPV and NPV, respectively, for observing elimination or the bounce back of infection) for a defined prevalence of infection 2 years post the cessation of MDA. For the arm only involving the treatment of Pre-SAC and SAC, the failure rate is predicted to be very high (particularly for hookworm-infected villages) unless transmission intensity is very low (R 0 , or the effective reproductive number R, just above unity in value). The calculations are designed to consider various trial arms and stratifications; namely, community-based treatment and Pre-SAC and SAC only treatment (the two arms of the trial), different STH transmission settings of low, medium and high, and different STH species mixes. Results are considered in the light of the complications introduced by the choice of statistic to define success or failure, varying adherence to treatment, migration and parameter uncertainty.

  5. Precision of systematic and random sampling in clustered populations: habitat patches and aggregating organisms.

    PubMed

    McGarvey, Richard; Burch, Paul; Matthews, Janet M

    2016-01-01

    Natural populations of plants and animals spatially cluster because (1) suitable habitat is patchy, and (2) within suitable habitat, individuals aggregate further into clusters of higher density. We compare the precision of random and systematic field sampling survey designs under these two processes of species clustering. Second, we evaluate the performance of 13 estimators for the variance of the sample mean from a systematic survey. Replicated simulated surveys, as counts from 100 transects, allocated either randomly or systematically within the study region, were used to estimate population density in six spatial point populations including habitat patches and Matérn circular clustered aggregations of organisms, together and in combination. The standard one-start aligned systematic survey design, a uniform 10 x 10 grid of transects, was much more precise. Variances of the 10 000 replicated systematic survey mean densities were one-third to one-fifth of those from randomly allocated transects, implying transect sample sizes giving equivalent precision by random survey would need to be three to five times larger. Organisms being restricted to patches of habitat was alone sufficient to yield this precision advantage for the systematic design. But this improved precision for systematic sampling in clustered populations is underestimated by standard variance estimators used to compute confidence intervals. True variance for the survey sample mean was computed from the variance of 10 000 simulated survey mean estimates. Testing 10 published and three newly proposed variance estimators, the two variance estimators (v) that corrected for inter-transect correlation (ν₈ and ν(W)) were the most accurate and also the most precise in clustered populations. These greatly outperformed the two "post-stratification" variance estimators (ν₂ and ν₃) that are now more commonly applied in systematic surveys. Similar variance estimator performance rankings were found with a second differently generated set of spatial point populations, ν₈ and ν(W) again being the best performers in the longer-range autocorrelated populations. However, no systematic variance estimators tested were free from bias. On balance, systematic designs bring more narrow confidence intervals in clustered populations, while random designs permit unbiased estimates of (often wider) confidence interval. The search continues for better estimators of sampling variance for the systematic survey mean.

  6. A clinical carepath for obese pregnant women: A pragmatic pilot cluster randomized controlled trial.

    PubMed

    McDonald, Sarah D; Viaje, Kristen A; Rooney, Rebecca A; Jarde, Alexander; Giglia, Lucia; Maxwell, Cynthia V; Small, David; Kelly, Tracy Pearce; Midwifery, B H Sc; Sabatino, Lisa; Thabane, Lehana

    2018-05-17

    Obese women are at increased risks for complications during pregnancy, birth and in their infants. Although guidelines have been established for the clinical care of obese pregnant women, management is sometimes suboptimal. Our goal was to determine the feasibility of implementing and testing a clinical carepath for obese pregnant women compared to standard care, in a pilot cluster randomized controlled trial (RCT). A pragmatic pilot cluster RCT was conducted, randomly allocating eight clinics to the carepath or standard care for obese pregnant women. Women were eligible if they had a prepregnancy body mass index of ≥ 30 kg/m 2 and a viable singleton < 21 weeks. The primary outcomes were the feasibility of conducting a full-scale cluster RCT (defined as > 80%: randomization of clinics, use in eligible women, and completeness of follow-up) and of the intervention (defined as > 80%: compliance with each step in the carepath, and recommendation of the carepath by clinicians to a colleague). All eight approached clinics agreed to participate and were randomized. Half of the intervention clinics used the carepath, resulting in < 80% uptake of eligible women. High follow-up (99.5%) was achieved, in 188 of 189 women. The carepath was feasible for numerous guideline-directed recommendations for screening, but less so for counselling topics. When the carepath was used in the majority of women, all clinicians, most of whom were midwives, reported they would recommend it to a colleague. The intervention group had significantly higher overall adherence to the guideline recommendations compared to control (relative risk 1.71, 95% confidence interval 1.57-1.87). In this pragmatic pilot cluster RCT, a guideline-directed clinical carepath improved some aspects of care of obese pregnant women and was recommended by clinicians, particularly midwives. A cluster RCT may not be feasible in a mix of obstetric and midwifery clinics, but may be feasible in midwifery clinics. This pragmatic pilot cluster RCT was registered on clinicaltrials.gov (identifier: NCT02534051 ).

  7. Descriptive epidemiology of typhoid fever during an epidemic in Harare, Zimbabwe, 2012.

    PubMed

    Polonsky, Jonathan A; Martínez-Pino, Isabel; Nackers, Fabienne; Chonzi, Prosper; Manangazira, Portia; Van Herp, Michel; Maes, Peter; Porten, Klaudia; Luquero, Francisco J

    2014-01-01

    Typhoid fever remains a significant public health problem in developing countries. In October 2011, a typhoid fever epidemic was declared in Harare, Zimbabwe - the fourth enteric infection epidemic since 2008. To orient control activities, we described the epidemiology and spatiotemporal clustering of the epidemic in Dzivaresekwa and Kuwadzana, the two most affected suburbs of Harare. A typhoid fever case-patient register was analysed to describe the epidemic. To explore clustering, we constructed a dataset comprising GPS coordinates of case-patient residences and randomly sampled residential locations (spatial controls). The scale and significance of clustering was explored with Ripley K functions. Cluster locations were determined by a random labelling technique and confirmed using Kulldorff's spatial scan statistic. We analysed data from 2570 confirmed and suspected case-patients, and found significant spatiotemporal clustering of typhoid fever in two non-overlapping areas, which appeared to be linked to environmental sources. Peak relative risk was more than six times greater than in areas lying outside the cluster ranges. Clusters were identified in similar geographical ranges by both random labelling and Kulldorff's spatial scan statistic. The spatial scale at which typhoid fever clustered was highly localised, with significant clustering at distances up to 4.5 km and peak levels at approximately 3.5 km. The epicentre of infection transmission shifted from one cluster to the other during the course of the epidemic. This study demonstrated highly localised clustering of typhoid fever during an epidemic in an urban African setting, and highlights the importance of spatiotemporal analysis for making timely decisions about targetting prevention and control activities and reinforcing treatment during epidemics. This approach should be integrated into existing surveillance systems to facilitate early detection of epidemics and identify their spatial range.

  8. Descriptive Epidemiology of Typhoid Fever during an Epidemic in Harare, Zimbabwe, 2012

    PubMed Central

    Polonsky, Jonathan A.; Martínez-Pino, Isabel; Nackers, Fabienne; Chonzi, Prosper; Manangazira, Portia; Van Herp, Michel; Maes, Peter; Porten, Klaudia; Luquero, Francisco J.

    2014-01-01

    Background Typhoid fever remains a significant public health problem in developing countries. In October 2011, a typhoid fever epidemic was declared in Harare, Zimbabwe - the fourth enteric infection epidemic since 2008. To orient control activities, we described the epidemiology and spatiotemporal clustering of the epidemic in Dzivaresekwa and Kuwadzana, the two most affected suburbs of Harare. Methods A typhoid fever case-patient register was analysed to describe the epidemic. To explore clustering, we constructed a dataset comprising GPS coordinates of case-patient residences and randomly sampled residential locations (spatial controls). The scale and significance of clustering was explored with Ripley K functions. Cluster locations were determined by a random labelling technique and confirmed using Kulldorff's spatial scan statistic. Principal Findings We analysed data from 2570 confirmed and suspected case-patients, and found significant spatiotemporal clustering of typhoid fever in two non-overlapping areas, which appeared to be linked to environmental sources. Peak relative risk was more than six times greater than in areas lying outside the cluster ranges. Clusters were identified in similar geographical ranges by both random labelling and Kulldorff's spatial scan statistic. The spatial scale at which typhoid fever clustered was highly localised, with significant clustering at distances up to 4.5 km and peak levels at approximately 3.5 km. The epicentre of infection transmission shifted from one cluster to the other during the course of the epidemic. Conclusions This study demonstrated highly localised clustering of typhoid fever during an epidemic in an urban African setting, and highlights the importance of spatiotemporal analysis for making timely decisions about targetting prevention and control activities and reinforcing treatment during epidemics. This approach should be integrated into existing surveillance systems to facilitate early detection of epidemics and identify their spatial range. PMID:25486292

  9. Review of Recent Methodological Developments in Group-Randomized Trials: Part 1—Design

    PubMed Central

    Li, Fan; Gallis, John A.; Prague, Melanie; Murray, David M.

    2017-01-01

    In 2004, Murray et al. reviewed methodological developments in the design and analysis of group-randomized trials (GRTs). We have highlighted the developments of the past 13 years in design with a companion article to focus on developments in analysis. As a pair, these articles update the 2004 review. We have discussed developments in the topics of the earlier review (e.g., clustering, matching, and individually randomized group-treatment trials) and in new topics, including constrained randomization and a range of randomized designs that are alternatives to the standard parallel-arm GRT. These include the stepped-wedge GRT, the pseudocluster randomized trial, and the network-randomized GRT, which, like the parallel-arm GRT, require clustering to be accounted for in both their design and analysis. PMID:28426295

  10. Review of Recent Methodological Developments in Group-Randomized Trials: Part 1-Design.

    PubMed

    Turner, Elizabeth L; Li, Fan; Gallis, John A; Prague, Melanie; Murray, David M

    2017-06-01

    In 2004, Murray et al. reviewed methodological developments in the design and analysis of group-randomized trials (GRTs). We have highlighted the developments of the past 13 years in design with a companion article to focus on developments in analysis. As a pair, these articles update the 2004 review. We have discussed developments in the topics of the earlier review (e.g., clustering, matching, and individually randomized group-treatment trials) and in new topics, including constrained randomization and a range of randomized designs that are alternatives to the standard parallel-arm GRT. These include the stepped-wedge GRT, the pseudocluster randomized trial, and the network-randomized GRT, which, like the parallel-arm GRT, require clustering to be accounted for in both their design and analysis.

  11. Implementing international osteoarthritis treatment guidelines in primary health care: study protocol for the SAMBA stepped wedge cluster randomized controlled trial.

    PubMed

    Østerås, Nina; van Bodegom-Vos, Leti; Dziedzic, Krysia; Moseng, Tuva; Aas, Eline; Andreassen, Øyvor; Mdala, Ibrahim; Natvig, Bård; Røtterud, Jan Harald; Schjervheim, Unni-Berit; Vlieland, Thea Vliet; Hagen, Kåre Birger

    2015-12-02

    Previous research indicates that people with osteoarthritis (OA) are not receiving the recommended and optimal treatment. Based on international treatment recommendations for hip and knee OA and previous research, the SAMBA model for integrated OA care in Norwegian primary health care has been developed. The model includes physiotherapist (PT) led patient OA education sessions and an exercise programme lasting 8-12 weeks. This study aims to assess the effectiveness, feasibility, and costs of a tailored strategy to implement the SAMBA model. A cluster randomized controlled trial with stepped wedge design including an effect, process, and cost evaluation will be conducted in six municipalities (clusters) in Norway. The municipalities will be randomized for time of crossover from current usual care to the implementation of the SAMBA model by a tailored strategy. The tailored strategy includes interactive workshops for general practitioners (GPs) and PTs in primary care covering the SAMBA model for integrated OA care, educational material, educational outreach visits, feedback, and reminder material. Outcomes will be measured at the patient, GP, and PT levels using self-report, semi-structured interviews, and register based data. The primary outcome measure is patient-reported quality of care (OsteoArthritis Quality Indicator questionnaire) at 6-month follow-up. Secondary outcomes include referrals to PT, imaging, and referrals to the orthopaedic surgeon as well as participants' treatment satisfaction, symptoms, physical activity level, body weight, and self-reported and measured lower limb function. The actual exposure to the tailor made implementation strategy and user experiences will be measured in a process evaluation. In the economic evaluation, the difference in costs of usual OA care and the SAMBA model for integrated OA care will be compared with the difference in health outcomes and reported by the incremental cost-effectiveness ratio (ICER). The results from the present study will add to the current knowledge on tailored strategies, which aims to improve the uptake of evidence-based OA care recommendations and improve the quality of OA care in primary health care. The new knowledge can be used in national and international initiatives designed to improve the quality of OA care. ClinicalTrials.gov NCT02333656.

  12. Radial alignment of elliptical galaxies by the tidal force of a cluster of galaxies

    NASA Astrophysics Data System (ADS)

    Rong, Yu; Yi, Shu-Xu; Zhang, Shuang-Nan; Tu, Hong

    2015-08-01

    Unlike the random radial orientation distribution of field elliptical galaxies, galaxies in a cluster are expected to point preferentially towards the centre of the cluster, as a result of the cluster's tidal force on its member galaxies. In this work, an analytic model is formulated to simulate this effect. The deformation time-scale of a galaxy in a cluster is usually much shorter than the time-scale of change of the tidal force; the dynamical process of tidal interaction within the galaxy can thus be ignored. The equilibrium shape of a galaxy is then assumed to be the surface of equipotential that is the sum of the self-gravitational potential of the galaxy and the tidal potential of the cluster at this location. We use a Monte Carlo method to calculate the radial orientation distribution of cluster galaxies, by assuming a Navarro-Frenk-White mass profile for the cluster and the initial ellipticity of field galaxies. The radial angles show a single-peak distribution centred at zero. The Monte Carlo simulations also show that a shift of the reference centre from the real cluster centre weakens the anisotropy of the radial angle distribution. Therefore, the expected radial alignment cannot be revealed if the distribution of spatial position angle is used instead of that of radial angle. The observed radial orientations of elliptical galaxies in cluster Abell 2744 are consistent with the simulated distribution.

  13. Applying the Anderson-Darling test to suicide clusters: evidence of contagion at U. S. universities?

    PubMed

    MacKenzie, Donald W

    2013-01-01

    Suicide clusters at Cornell University and the Massachusetts Institute of Technology (MIT) prompted popular and expert speculation of suicide contagion. However, some clustering is to be expected in any random process. This work tested whether suicide clusters at these two universities differed significantly from those expected under a homogeneous Poisson process, in which suicides occur randomly and independently of one another. Suicide dates were collected for MIT and Cornell for 1990-2012. The Anderson-Darling statistic was used to test the goodness-of-fit of the intervals between suicides to distribution expected under the Poisson process. Suicides at MIT were consistent with the homogeneous Poisson process, while those at Cornell showed clustering inconsistent with such a process (p = .05). The Anderson-Darling test provides a statistically powerful means to identify suicide clustering in small samples. Practitioners can use this method to test for clustering in relevant communities. The difference in clustering behavior between the two institutions suggests that more institutions should be studied to determine the prevalence of suicide clustering in universities and its causes.

  14. Cluster mass inference via random field theory.

    PubMed

    Zhang, Hui; Nichols, Thomas E; Johnson, Timothy D

    2009-01-01

    Cluster extent and voxel intensity are two widely used statistics in neuroimaging inference. Cluster extent is sensitive to spatially extended signals while voxel intensity is better for intense but focal signals. In order to leverage strength from both statistics, several nonparametric permutation methods have been proposed to combine the two methods. Simulation studies have shown that of the different cluster permutation methods, the cluster mass statistic is generally the best. However, to date, there is no parametric cluster mass inference available. In this paper, we propose a cluster mass inference method based on random field theory (RFT). We develop this method for Gaussian images, evaluate it on Gaussian and Gaussianized t-statistic images and investigate its statistical properties via simulation studies and real data. Simulation results show that the method is valid under the null hypothesis and demonstrate that it can be more powerful than the cluster extent inference method. Further, analyses with a single subject and a group fMRI dataset demonstrate better power than traditional cluster size inference, and good accuracy relative to a gold-standard permutation test.

  15. Sun protection at elementary schools: a cluster randomized trial.

    PubMed

    Hunter, Seft; Love-Jackson, Kymia; Abdulla, Rania; Zhu, Weiwei; Lee, Ji-Hyun; Wells, Kristen J; Roetzheim, Richard

    2010-04-07

    Elementary schools represent both a source of childhood sun exposure and a setting for educational interventions. Sun Protection of Florida's Children was a cluster randomized trial promoting hat use at (primary outcome) and outside of schools among fourth-grade students during August 8, 2006, through May 22, 2007. Twenty-two schools were randomly assigned to the intervention (1115 students) or control group (1376 students). Intervention schools received classroom sessions targeting sun protection attitudes and social norms. Each student attending an intervention school received two free wide-brimmed hats. Hat use at school was measured by direct observation and hat use outside of school was measured by self-report. A subgroup of 378 students (178 in the intervention group and 200 in the control group) underwent serial measurements of skin pigmentation to explore potential physiological effects of the intervention. Generalized linear mixed models were used to evaluate the intervention effect by accounting for the cluster randomized trial design. All P values were two-sided and were claimed as statistically significant at a level of .05. The percentage of students observed wearing hats at control schools remained essentially unchanged during the school year (baseline = 2%, fall = 0%, and spring = 1%) but increased statistically significantly at intervention schools (baseline = 2%, fall = 30%, and spring = 41%) (P < .001 for intervention effect comparing the change in rate of hat use over time at intervention vs control schools). Self-reported use of hats outside of school did not change statistically significantly during the study (control: baseline = 14%, fall = 14%, and spring = 11%; intervention: baseline = 24%, fall = 24%, and spring = 23%) nor did measures of skin pigmentation. The intervention increased use of hats among fourth-grade students at school but had no effect on self-reported wide-brimmed hat use outside of school or on measures of skin pigmentation.

  16. Using Geographic Information Systems and Spatial Analysis Methods to Assess Household Water Access and Sanitation Coverage in the SHINE Trial.

    PubMed

    Ntozini, Robert; Marks, Sara J; Mangwadu, Goldberg; Mbuya, Mduduzi N N; Gerema, Grace; Mutasa, Batsirai; Julian, Timothy R; Schwab, Kellogg J; Humphrey, Jean H; Zungu, Lindiwe I

    2015-12-15

    Access to water and sanitation are important determinants of behavioral responses to hygiene and sanitation interventions. We estimated cluster-specific water access and sanitation coverage to inform a constrained randomization technique in the SHINE trial. Technicians and engineers inspected all public access water sources to ascertain seasonality, function, and geospatial coordinates. Households and water sources were mapped using open-source geospatial software. The distance from each household to the nearest perennial, functional, protected water source was calculated, and for each cluster, the median distance and the proportion of households within <500 m and >1500 m of such a water source. Cluster-specific sanitation coverage was ascertained using a random sample of 13 households per cluster. These parameters were included as covariates in randomization to optimize balance in water and sanitation access across treatment arms at the start of the trial. The observed high variability between clusters in both parameters suggests that constraining on these factors was needed to reduce risk of bias. © The Author 2015. Published by Oxford University Press for the Infectious Diseases Society of America.

  17. Applying the Transtheoretical Model to evaluate the effect of a call-recall program in enhancing Pap smear practice: a cluster randomized trial.

    PubMed

    Abdullah, Fauziah; Su, Tin Tin

    2013-01-01

    The objective of this study was to evaluate the effect of a call-recall approach in enhancing Pap smear practice by changes of motivation stage among non-compliant women. A cluster randomized controlled trial with parallel and un-blinded design was conducted between January and November 2010 in 40 public secondary schools in Malaysia among 403 female teachers who never or infrequently attended for a Pap test. A cluster randomization was applied in assigning schools to both groups. An intervention group received an invitation and reminder (call-recall program) for a Pap test (20 schools with 201 participants), while the control group received usual care from the existing cervical screening program (20 schools with 202 participants). Multivariate logistic regression was performed to determine the effect of the intervention program on the action stage (Pap smear uptake) at 24 weeks. In both groups, pre-contemplation stage was found as the highest proportion of changes in stages. At 24 weeks, an intervention group showed two times more in the action stage than control group (adjusted odds ratio 2.44, 95% CI 1.29-4.62). The positive effect of a call-recall approach in motivating women to change the behavior of screening practice should be appreciated by policy makers and health care providers in developing countries as an intervention to enhance Pap smear uptake. Copyright © 2013 Elsevier Inc. All rights reserved.

  18. Use of LANDSAT imagery for wildlife habitat mapping in northeast and eastcentral Alaska

    NASA Technical Reports Server (NTRS)

    Lent, P. C. (Principal Investigator)

    1976-01-01

    The author has identified the following significant results. There is strong indication that spatially rare feature classes may be missed in clustering classifications based on 2% random sampling. Therefore, it seems advisable to augment random sampling for cluster analysis with directed sampling of any spatially rare features which are relevant to the analysis.

  19. Efficacy of a Universal Parent Training Program (HOPE-20): Cluster Randomized Controlled Trial

    ERIC Educational Resources Information Center

    Leung, Cynthia; Tsang, Sandra; Kwan, H. W.

    2017-01-01

    Objective: This study examined the efficacy of Hands-On Parent Empowerment-20 (HOPE-20) program. Methods: Eligible participants were parents residing in Hong Kong with target children aged 2 years attending nursery schools. Cluster randomized control trial was adopted, with 10 schools (110 participants) assigned to intervention group and 8 schools…

  20. A Multisite Cluster Randomized Field Trial of Open Court Reading

    ERIC Educational Resources Information Center

    Borman, Geoffrey D.; Dowling, N. Maritza; Schneck, Carrie

    2008-01-01

    In this article, the authors report achievement outcomes of a multisite cluster randomized field trial of Open Court Reading 2005 (OCR), a K-6 literacy curriculum published by SRA/McGraw-Hill. The participants are 49 first-grade through fifth-grade classrooms from predominantly minority and poor contexts across the nation. Blocking by grade level…

  1. Nanosecond laser-cluster interactions at 109-1012 W/cm 2

    NASA Astrophysics Data System (ADS)

    Singh, Rohtash; Tripathi, V. K.; Vatsa, R. K.; Das, D.

    2017-08-01

    An analytical model and a numerical code are developed to study the evolution of multiple charge states of ions by irradiating clusters of atoms of a high atomic number (e.g., Xe) by 1.06 μm and 0.53 μm nanosecond laser pulses of an intensity in the range of 109-1012 W/cm 2 . The laser turns clusters into plasma nanoballs. Initially, the momentum randomizing collisions of electrons are with neutrals, but soon these are taken over by collisions with ions. The ionization of an ion to the next higher state of ionization is taken to be caused by an energetic free electron impact, and the rates of impact ionization are suitably modelled by having an inverse exponential dependence of ionizing collision frequency on the ratio of ionization potential to electron temperature. Cluster expansion led adiabatic cooling is a major limiting mechanism on electron temperature. In the intensity range considered, ionization states up to 7 are expected with nanosecond pulses. Another possible mechanism, filamentation of the laser, has also been considered to account for the observation of higher charged states. However, filamentation is seen to be insufficient to cause substantial local enhancement in the intensity to affect electron heating rates.

  2. Sampling designs for HIV molecular epidemiology with application to Honduras.

    PubMed

    Shepherd, Bryan E; Rossini, Anthony J; Soto, Ramon Jeremias; De Rivera, Ivette Lorenzana; Mullins, James I

    2005-11-01

    Proper sampling is essential to characterize the molecular epidemiology of human immunodeficiency virus (HIV). HIV sampling frames are difficult to identify, so most studies use convenience samples. We discuss statistically valid and feasible sampling techniques that overcome some of the potential for bias due to convenience sampling and ensure better representation of the study population. We employ a sampling design called stratified cluster sampling. This first divides the population into geographical and/or social strata. Within each stratum, a population of clusters is chosen from groups, locations, or facilities where HIV-positive individuals might be found. Some clusters are randomly selected within strata and individuals are randomly selected within clusters. Variation and cost help determine the number of clusters and the number of individuals within clusters that are to be sampled. We illustrate the approach through a study designed to survey the heterogeneity of subtype B strains in Honduras.

  3. Avulsion Clusters in Alluvial Systems: An Example of Large-Scale Self-Organization in Ancient and Experimental Basins

    NASA Astrophysics Data System (ADS)

    Hajek, E.; Heller, P.; Huzurbazar, S.; Sheets, B.; Paola, C.

    2006-12-01

    The stratigraphic record of at least some alluvial basins exhibits a spatial structure that may reflect long time- scale (103-105 yr in natural basins) autogenic organization of river avulsions. Current models of avulsion-dominated alluvial sequences emphasize the spatial and temporal distribution of coarse-grained channel-belt deposits amid fine-grained floodplain materials. These models typically assume that individual avulsions move, either randomly or deterministically, to low spots distributed throughout the model space. However, our observations of ancient deposits and experimental stratigraphy indicate a previously unrecognized pattern of channel-belt organization, where clusters of closely-spaced channel-belt deposits are separated from each other by extensive intervals of overbank deposits. We explore potential causes of and controls on avulsion clustering with outcrop and subsurface data from Late Cretaceous/Early Paleogene fluvial deposits in the Rocky Mountains (including the Ferris, Lance, and Fort Union formations of Wyoming) and results of physical stratigraphy experiments from the St. Anthony Falls Lab, University of Minnesota. We use Ripley's K-function to determine the degree and scales of clustering in these basins with results that show moderate statistical clustering in experimental deposits and strong clustering in the Ferris Formation (Hanna Basin, Wyoming). External controls (base level, subsidence rate, and sediment/water supplies) were not varied during the experiment, and therefore not factors in cluster formation. Likewise, the stratigraphic context of the ancient system (including the absence of incised valleys and lack of faulting) suggests that obvious extrinsic controls, such as base level change and local tectonics, were not major influences on the development of clusters. We propose that avulsion clusters, as seen in this study, reflect a scale of self-organization in alluvial basins that is not usually recognized in stratigraphy. However cursory examination of other ancient systems suggests that such structure may be common in the rock record. Understanding mechanisms driving avulsion clustering will shed light on the dominant processes in alluvial basins over long time scales. Furthermore, characterizing autogenic avulsion clusters will be an important factor to consider when interpreting allogenic signals in ancient basin fills.

  4. Quantifying randomness in real networks

    NASA Astrophysics Data System (ADS)

    Orsini, Chiara; Dankulov, Marija M.; Colomer-de-Simón, Pol; Jamakovic, Almerima; Mahadevan, Priya; Vahdat, Amin; Bassler, Kevin E.; Toroczkai, Zoltán; Boguñá, Marián; Caldarelli, Guido; Fortunato, Santo; Krioukov, Dmitri

    2015-10-01

    Represented as graphs, real networks are intricate combinations of order and disorder. Fixing some of the structural properties of network models to their values observed in real networks, many other properties appear as statistical consequences of these fixed observables, plus randomness in other respects. Here we employ the dk-series, a complete set of basic characteristics of the network structure, to study the statistical dependencies between different network properties. We consider six real networks--the Internet, US airport network, human protein interactions, technosocial web of trust, English word network, and an fMRI map of the human brain--and find that many important local and global structural properties of these networks are closely reproduced by dk-random graphs whose degree distributions, degree correlations and clustering are as in the corresponding real network. We discuss important conceptual, methodological, and practical implications of this evaluation of network randomness, and release software to generate dk-random graphs.

  5. Efficient design of cluster randomized trials with treatment-dependent costs and treatment-dependent unknown variances.

    PubMed

    van Breukelen, Gerard J P; Candel, Math J J M

    2018-06-10

    Cluster randomized trials evaluate the effect of a treatment on persons nested within clusters, where treatment is randomly assigned to clusters. Current equations for the optimal sample size at the cluster and person level assume that the outcome variances and/or the study costs are known and homogeneous between treatment arms. This paper presents efficient yet robust designs for cluster randomized trials with treatment-dependent costs and treatment-dependent unknown variances, and compares these with 2 practical designs. First, the maximin design (MMD) is derived, which maximizes the minimum efficiency (minimizes the maximum sampling variance) of the treatment effect estimator over a range of treatment-to-control variance ratios. The MMD is then compared with the optimal design for homogeneous variances and costs (balanced design), and with that for homogeneous variances and treatment-dependent costs (cost-considered design). The results show that the balanced design is the MMD if the treatment-to control cost ratio is the same at both design levels (cluster, person) and within the range for the treatment-to-control variance ratio. It still is highly efficient and better than the cost-considered design if the cost ratio is within the range for the squared variance ratio. Outside that range, the cost-considered design is better and highly efficient, but it is not the MMD. An example shows sample size calculation for the MMD, and the computer code (SPSS and R) is provided as supplementary material. The MMD is recommended for trial planning if the study costs are treatment-dependent and homogeneity of variances cannot be assumed. © 2018 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

  6. Disorder-induced stiffness degradation of highly disordered porous materials

    NASA Astrophysics Data System (ADS)

    Laubie, Hadrien; Monfared, Siavash; Radjaï, Farhang; Pellenq, Roland; Ulm, Franz-Josef

    2017-09-01

    The effective mechanical behavior of multiphase solid materials is generally modeled by means of homogenization techniques that account for phase volume fractions and elastic moduli without considering the spatial distribution of the different phases. By means of extensive numerical simulations of randomly generated porous materials using the lattice element method, the role of local textural properties on the effective elastic properties of disordered porous materials is investigated and compared with different continuum micromechanics-based models. It is found that the pronounced disorder-induced stiffness degradation originates from stress concentrations around pore clusters in highly disordered porous materials. We identify a single disorder parameter, φsa, which combines a measure of the spatial disorder of pores (the clustering index, sa) with the pore volume fraction (the porosity, φ) to scale the disorder-induced stiffness degradation. Thus, we conclude that the classical continuum micromechanics models with one spherical pore phase, due to their underlying homogeneity assumption fall short of addressing the clustering effect, unless additional texture information is introduced, e.g. in form of the shift of the percolation threshold with disorder, or other functional relations between volume fractions and spatial disorder; as illustrated herein for a differential scheme model representative of a two-phase (solid-pore) composite model material.

  7. Novel approaches to pin cluster synchronization on complex dynamical networks in Lur'e forms

    NASA Astrophysics Data System (ADS)

    Tang, Ze; Park, Ju H.; Feng, Jianwen

    2018-04-01

    This paper investigates the cluster synchronization of complex dynamical networks consisted of identical or nonidentical Lur'e systems. Due to the special topology structure of the complex networks and the existence of stochastic perturbations, a kind of randomly occurring pinning controller is designed which not only synchronizes all Lur'e systems in the same cluster but also decreases the negative influence among different clusters. Firstly, based on an extended integral inequality, the convex combination theorem and S-procedure, the conditions for cluster synchronization of identical Lur'e networks are derived in a convex domain. Secondly, randomly occurring adaptive pinning controllers with two independent Bernoulli stochastic variables are designed and then sufficient conditions are obtained for the cluster synchronization on complex networks consisted of nonidentical Lur'e systems. In addition, suitable control gains for successful cluster synchronization of nonidentical Lur'e networks are acquired by designing some adaptive updating laws. Finally, we present two numerical examples to demonstrate the validity of the control scheme and the theoretical analysis.

  8. Predicting lower mantle heterogeneity from 4-D Earth models

    NASA Astrophysics Data System (ADS)

    Flament, Nicolas; Williams, Simon; Müller, Dietmar; Gurnis, Michael; Bower, Dan J.

    2016-04-01

    The Earth's lower mantle is characterized by two large-low-shear velocity provinces (LLSVPs), approximately ˜15000 km in diameter and 500-1000 km high, located under Africa and the Pacific Ocean. The spatial stability and chemical nature of these LLSVPs are debated. Here, we compare the lower mantle structure predicted by forward global mantle flow models constrained by tectonic reconstructions (Bower et al., 2015) to an analysis of five global tomography models. In the dynamic models, spanning 230 million years, slabs subducting deep into the mantle deform an initially uniform basal layer containing 2% of the volume of the mantle. Basal density, convective vigour (Rayleigh number Ra), mantle viscosity, absolute plate motions, and relative plate motions are varied in a series of model cases. We use cluster analysis to classify a set of equally-spaced points (average separation ˜0.45°) on the Earth's surface into two groups of points with similar variations in present-day temperature between 1000-2800 km depth, for each model case. Below ˜2400 km depth, this procedure reveals a high-temperature cluster in which mantle temperature is significantly larger than ambient and a low-temperature cluster in which mantle temperature is lower than ambient. The spatial extent of the high-temperature cluster is in first-order agreement with the outlines of the African and Pacific LLSVPs revealed by a similar cluster analysis of five tomography models (Lekic et al., 2012). Model success is quantified by computing the accuracy and sensitivity of the predicted temperature clusters in predicting the low-velocity cluster obtained from tomography (Lekic et al., 2012). In these cases, the accuracy varies between 0.61-0.80, where a value of 0.5 represents the random case, and the sensitivity ranges between 0.18-0.83. The largest accuracies and sensitivities are obtained for models with Ra ≈ 5 x 107, no asthenosphere (or an asthenosphere restricted to the oceanic domain), and a basal layer ˜ 4% denser than ambient mantle. Increasing convective vigour (Ra ≈ 5 x 108) or decreasing the density of the basal layer decreases both the accuracy and sensitivity of the predicted lower mantle structure. References: D. J. Bower, M. Gurnis, N. Flament, Assimilating lithosphere and slab history in 4-D Earth models. Phys. Earth Planet. Inter. 238, 8-22 (2015). V. Lekic, S. Cottaar, A. Dziewonski, B. Romanowicz, Cluster analysis of global lower mantle tomography: A new class of structure and implications for chemical heterogeneity. Earth Planet. Sci. Lett. 357, 68-77 (2012).

  9. Evaluation of the procedure 1A component of the 1980 US/Canada wheat and barley exploratory experiment

    NASA Technical Reports Server (NTRS)

    Chapman, G. M. (Principal Investigator); Carnes, J. G.

    1981-01-01

    Several techniques which use clusters generated by a new clustering algorithm, CLASSY, are proposed as alternatives to random sampling to obtain greater precision in crop proportion estimation: (1) Proportional Allocation/relative count estimator (PA/RCE) uses proportional allocation of dots to clusters on the basis of cluster size and a relative count cluster level estimate; (2) Proportional Allocation/Bayes Estimator (PA/BE) uses proportional allocation of dots to clusters and a Bayesian cluster-level estimate; and (3) Bayes Sequential Allocation/Bayesian Estimator (BSA/BE) uses sequential allocation of dots to clusters and a Bayesian cluster level estimate. Clustering in an effective method in making proportion estimates. It is estimated that, to obtain the same precision with random sampling as obtained by the proportional sampling of 50 dots with an unbiased estimator, samples of 85 or 166 would need to be taken if dot sets with AI labels (integrated procedure) or ground truth labels, respectively were input. Dot reallocation provides dot sets that are unbiased. It is recommended that these proportion estimation techniques are maintained, particularly the PA/BE because it provides the greatest precision.

  10. Analysis of a Spatial Point Pattern: Examining the Damage to Pavement and Pipes in Santa Clara Valley Resulting from the Loma Prieta Earthquake

    USGS Publications Warehouse

    Phelps, G.A.

    2008-01-01

    This report describes some simple spatial statistical methods to explore the relationships of scattered points to geologic or other features, represented by points, lines, or areas. It also describes statistical methods to search for linear trends and clustered patterns within the scattered point data. Scattered points are often contained within irregularly shaped study areas, necessitating the use of methods largely unexplored in the point pattern literature. The methods take advantage of the power of modern GIS toolkits to numerically approximate the null hypothesis of randomly located data within an irregular study area. Observed distributions can then be compared with the null distribution of a set of randomly located points. The methods are non-parametric and are applicable to irregularly shaped study areas. Patterns within the point data are examined by comparing the distribution of the orientation of the set of vectors defined by each pair of points within the data with the equivalent distribution for a random set of points within the study area. A simple model is proposed to describe linear or clustered structure within scattered data. A scattered data set of damage to pavement and pipes, recorded after the 1989 Loma Prieta earthquake, is used as an example to demonstrate the analytical techniques. The damage is found to be preferentially located nearer a set of mapped lineaments than randomly scattered damage, suggesting range-front faulting along the base of the Santa Cruz Mountains is related to both the earthquake damage and the mapped lineaments. The damage also exhibit two non-random patterns: a single cluster of damage centered in the town of Los Gatos, California, and a linear alignment of damage along the range front of the Santa Cruz Mountains, California. The linear alignment of damage is strongest between 45? and 50? northwest. This agrees well with the mean trend of the mapped lineaments, measured as 49? northwest.

  11. A cluster-randomized effectiveness trial of a physician-pharmacist collaborative model to improve blood pressure control.

    PubMed

    Carter, Barry L; Clarke, William; Ardery, Gail; Weber, Cynthia A; James, Paul A; Vander Weg, Mark; Chrischilles, Elizabeth A; Vaughn, Thomas; Egan, Brent M

    2010-07-01

    Numerous studies have demonstrated the value of team-based care to improve blood pressure (BP) control, but there is limited information on whether these models would be adopted in diverse populations. The purpose of this study was to evaluate whether a collaborative model between physicians and pharmacists can improve BP control in multiple primary care medical offices with diverse geographic and patient characteristics and whether long-term BP control can be sustained. This study is a randomized prospective trial in 27 primary care offices first stratified by the percentage of underrepresented minorities and the level of clinical pharmacy services within the office. Each office is then randomized to either a 9- or 24-month intervention or a control group. Patients will be enrolled in this study until 2012. The results of this study should provide information on whether this model can be implemented in large numbers of diverse offices, if it is effective in diverse populations, and whether BP control can be sustained long term. URL: http://www.clinicaltrials.gov. Unique identifier: NCT00935077.

  12. Spatial distribution and cluster analysis of retail drug shop characteristics and antimalarial behaviors as reported by private medicine retailers in western Kenya: informing future interventions.

    PubMed

    Rusk, Andria; Highfield, Linda; Wilkerson, J Michael; Harrell, Melissa; Obala, Andrew; Amick, Benjamin

    2016-02-19

    Efforts to improve malaria case management in sub-Saharan Africa have shifted focus to private antimalarial retailers to increase access to appropriate treatment. Demands to decrease intervention cost while increasing efficacy requires interventions tailored to geographic regions with demonstrated need. Cluster analysis presents an opportunity to meet this demand, but has not been applied to the retail sector or antimalarial retailer behaviors. This research conducted cluster analysis on medicine retailer behaviors in Kenya, to improve malaria case management and inform future interventions. Ninety-seven surveys were collected from medicine retailers working in the Webuye Health and Demographic Surveillance Site. Survey items included retailer training, education, antimalarial drug knowledge, recommending behavior, sales, and shop characteristics, and were analyzed using Kulldorff's spatial scan statistic. The Bernoulli purely spatial model for binomial data was used, comparing cases to controls. Statistical significance of found clusters was tested with a likelihood ratio test, using the null hypothesis of no clustering, and a p value based on 999 Monte Carlo simulations. The null hypothesis was rejected with p values of 0.05 or less. A statistically significant cluster of fewer than expected pharmacy-trained retailers was found (RR = .09, p = .001) when compared to the expected random distribution. Drug recommending behavior also yielded a statistically significant cluster, with fewer than expected retailers recommending the correct antimalarial medication to adults (RR = .018, p = .01), and fewer than expected shops selling that medication more often than outdated antimalarials when compared to random distribution (RR = 0.23, p = .007). All three of these clusters were co-located, overlapping in the northwest of the study area. Spatial clustering was found in the data. A concerning amount of correlation was found in one specific region in the study area where multiple behaviors converged in space, highlighting a prime target for interventions. These results also demonstrate the utility of applying geospatial methods in the study of medicine retailer behaviors, making the case for expanding this approach to other regions.

  13. Observed intra-cluster correlation coefficients in a cluster survey sample of patient encounters in general practice in Australia

    PubMed Central

    Knox, Stephanie A; Chondros, Patty

    2004-01-01

    Background Cluster sample study designs are cost effective, however cluster samples violate the simple random sample assumption of independence of observations. Failure to account for the intra-cluster correlation of observations when sampling through clusters may lead to an under-powered study. Researchers therefore need estimates of intra-cluster correlation for a range of outcomes to calculate sample size. We report intra-cluster correlation coefficients observed within a large-scale cross-sectional study of general practice in Australia, where the general practitioner (GP) was the primary sampling unit and the patient encounter was the unit of inference. Methods Each year the Bettering the Evaluation and Care of Health (BEACH) study recruits a random sample of approximately 1,000 GPs across Australia. Each GP completes details of 100 consecutive patient encounters. Intra-cluster correlation coefficients were estimated for patient demographics, morbidity managed and treatments received. Intra-cluster correlation coefficients were estimated for descriptive outcomes and for associations between outcomes and predictors and were compared across two independent samples of GPs drawn three years apart. Results Between April 1999 and March 2000, a random sample of 1,047 Australian general practitioners recorded details of 104,700 patient encounters. Intra-cluster correlation coefficients for patient demographics ranged from 0.055 for patient sex to 0.451 for language spoken at home. Intra-cluster correlations for morbidity variables ranged from 0.005 for the management of eye problems to 0.059 for management of psychological problems. Intra-cluster correlation for the association between two variables was smaller than the descriptive intra-cluster correlation of each variable. When compared with the April 2002 to March 2003 sample (1,008 GPs) the estimated intra-cluster correlation coefficients were found to be consistent across samples. Conclusions The demonstrated precision and reliability of the estimated intra-cluster correlations indicate that these coefficients will be useful for calculating sample sizes in future general practice surveys that use the GP as the primary sampling unit. PMID:15613248

  14. Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data

    PubMed Central

    Hallac, David; Vare, Sagar; Boyd, Stephen; Leskovec, Jure

    2018-01-01

    Subsequence clustering of multivariate time series is a useful tool for discovering repeated patterns in temporal data. Once these patterns have been discovered, seemingly complicated datasets can be interpreted as a temporal sequence of only a small number of states, or clusters. For example, raw sensor data from a fitness-tracking application can be expressed as a timeline of a select few actions (i.e., walking, sitting, running). However, discovering these patterns is challenging because it requires simultaneous segmentation and clustering of the time series. Furthermore, interpreting the resulting clusters is difficult, especially when the data is high-dimensional. Here we propose a new method of model-based clustering, which we call Toeplitz Inverse Covariance-based Clustering (TICC). Each cluster in the TICC method is defined by a correlation network, or Markov random field (MRF), characterizing the interdependencies between different observations in a typical subsequence of that cluster. Based on this graphical representation, TICC simultaneously segments and clusters the time series data. We solve the TICC problem through alternating minimization, using a variation of the expectation maximization (EM) algorithm. We derive closed-form solutions to efficiently solve the two resulting subproblems in a scalable way, through dynamic programming and the alternating direction method of multipliers (ADMM), respectively. We validate our approach by comparing TICC to several state-of-the-art baselines in a series of synthetic experiments, and we then demonstrate on an automobile sensor dataset how TICC can be used to learn interpretable clusters in real-world scenarios. PMID:29770257

  15. Biased phylodynamic inferences from analysing clusters of viral sequences

    PubMed Central

    Xiang, Fei; Frost, Simon D. W.

    2017-01-01

    Abstract Phylogenetic methods are being increasingly used to help understand the transmission dynamics of measurably evolving viruses, including HIV. Clusters of highly similar sequences are often observed, which appear to follow a ‘power law’ behaviour, with a small number of very large clusters. These clusters may help to identify subpopulations in an epidemic, and inform where intervention strategies should be implemented. However, clustering of samples does not necessarily imply the presence of a subpopulation with high transmission rates, as groups of closely related viruses can also occur due to non-epidemiological effects such as over-sampling. It is important to ensure that observed phylogenetic clustering reflects true heterogeneity in the transmitting population, and is not being driven by non-epidemiological effects. We qualify the effect of using a falsely identified ‘transmission cluster’ of sequences to estimate phylodynamic parameters including the effective population size and exponential growth rate under several demographic scenarios. Our simulation studies show that taking the maximum size cluster to re-estimate parameters from trees simulated under a randomly mixing, constant population size coalescent process systematically underestimates the overall effective population size. In addition, the transmission cluster wrongly resembles an exponential or logistic growth model 99% of the time. We also illustrate the consequences of false clusters in exponentially growing coalescent and birth-death trees, where again, the growth rate is skewed upwards. This has clear implications for identifying clusters in large viral databases, where a false cluster could result in wasted intervention resources. PMID:28852573

  16. Efficient gradient-based Monte Carlo simulation of materials: Applications to amorphous Si and Fe and Ni clusters

    NASA Astrophysics Data System (ADS)

    Limbu, Dil; Biswas, Parthapratim

    We present a simple and efficient Monte-Carlo (MC) simulation of Iron (Fe) and Nickel (Ni) clusters with N =5-100 and amorphous Silicon (a-Si) starting from a random configuration. Using Sutton-Chen and Finnis-Sinclair potentials for Ni (in fcc lattice) and Fe (in bcc lattice), and Stillinger-Weber potential for a-Si, respectively, the total energy of the system is optimized by employing MC moves that include both the stochastic nature of MC simulations and the gradient of the potential function. For both iron and nickel clusters, the energy of the configurations is found to be very close to the values listed in the Cambridge Cluster Database, whereas the maximum force on each cluster is found to be much lower than the corresponding value obtained from the optimized structural configurations reported in the database. An extension of the method to model the amorphous state of Si is presented and the results are compared with experimental data and those obtained from other simulation methods. The work is partially supported by the NSF under Grant Number DMR 1507166.

  17. Traveling-cluster approximation for uncorrelated amorphous systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sen, A.K.; Mills, R.; Kaplan, T.

    1984-11-15

    We have developed a formalism for including cluster effects in the one-electron Green's function for a positionally disordered (liquid or amorphous) system without any correlation among the scattering sites. This method is an extension of the technique known as the traveling-cluster approximation (TCA) originally obtained and applied to a substitutional alloy by Mills and Ratanavararaksa. We have also proved the appropriate fixed-point theorem, which guarantees, for a bounded local potential, that the self-consistent equations always converge upon iteration to a unique, Herglotz solution. To our knowledge, this is the only analytic theory for considering cluster effects. Furthermore, we have performedmore » some computer calculations in the pair TCA, for the model case of delta-function potentials on a one-dimensional random chain. These results have been compared with ''exact calculations'' (which, in principle, take into account all cluster effects) and with the coherent-potential approximation (CPA), which is the single-site TCA. The density of states for the pair TCA clearly shows some improvement over the CPA and yet, apparently, the pair approximation distorts some of the features of the exact results.« less

  18. Genomic prediction using different estimation methodology, blending and cross-validation techniques for growth traits and visual scores in Hereford and Braford cattle.

    PubMed

    Campos, G S; Reimann, F A; Cardoso, L L; Ferreira, C E R; Junqueira, V S; Schmidt, P I; Braccini Neto, J; Yokoo, M J I; Sollero, B P; Boligon, A A; Cardoso, F F

    2018-05-07

    The objective of the present study was to evaluate the accuracy and bias of direct and blended genomic predictions using different methods and cross-validation techniques for growth traits (weight and weight gains) and visual scores (conformation, precocity, muscling and size) obtained at weaning and at yearling in Hereford and Braford breeds. Phenotypic data contained 126,290 animals belonging to the Delta G Connection genetic improvement program, and a set of 3,545 animals genotyped with the 50K chip and 131 sires with the 777K. After quality control, 41,045 markers remained for all animals. An animal model was used to estimate (co)variances components and to predict breeding values, which were later used to calculate the deregressed estimated breeding values (DEBV). Animals with genotype and phenotype for the traits studied were divided into four or five groups by random and k-means clustering cross-validation strategies. The values of accuracy of the direct genomic values (DGV) were moderate to high magnitude for at weaning and at yearling traits, ranging from 0.19 to 0.45 for the k-means and 0.23 to 0.78 for random clustering among all traits. The greatest gain in relation to the pedigree BLUP (PBLUP) was 9.5% with the BayesB method with both the k-means and the random clustering. Blended genomic value accuracies ranged from 0.19 to 0.56 for k-means and from 0.21 to 0.82 for random clustering. The analyzes using the historical pedigree and phenotypes contributed additional information to calculate the GEBV and in general, the largest gains were for the single-step (ssGBLUP) method in bivariate analyses with a mean increase of 43.00% among all traits measured at weaning and of 46.27% for those evaluated at yearling. The accuracy values for the marker effects estimation methods were lower for k-means clustering, indicating that the training set relationship to the selection candidates is a major factor affecting accuracy of genomic predictions. The gains in accuracy obtained with genomic blending methods, mainly ssGBLUP in bivariate analyses, indicate that genomic predictions should be used as a tool to improve genetic gains in relation to the traditional PBLUP selection.

  19. Spatial modelling and mapping of female genital mutilation in Kenya.

    PubMed

    Achia, Thomas N O

    2014-03-25

    Female genital mutilation/cutting (FGM/C) is still prevalent in several communities in Kenya and other areas in Africa, as well as being practiced by some migrants from African countries living in other parts of the world. This study aimed at detecting clustering of FGM/C in Kenya, and identifying those areas within the country where women still intend to continue the practice. A broader goal of the study was to identify geographical areas where the practice continues unabated and where broad intervention strategies need to be introduced. The prevalence of FGM/C was investigated using the 2008 Kenya Demographic and Health Survey (KDHS) data. The 2008 KDHS used a multistage stratified random sampling plan to select women of reproductive age (15-49 years) and asked questions concerning their FGM/C status and their support for the continuation of FGM/C. A spatial scan statistical analysis was carried out using SaTScan™ to test for statistically significant clustering of the practice of FGM/C in the country. The risk of FGM/C was also modelled and mapped using a hierarchical spatial model under the Integrated Nested Laplace approximation approach using the INLA library in R. The prevalence of FGM/C stood at 28.2% and an estimated 10.3% of the women interviewed indicated that they supported the continuation of FGM. On the basis of the Deviance Information Criterion (DIC), hierarchical spatial models with spatially structured random effects were found to best fit the data for both response variables considered. Age, region, rural-urban classification, education, marital status, religion, socioeconomic status and media exposure were found to be significantly associated with FGM/C. The current FGM/C status of a woman was also a significant predictor of support for the continuation of FGM/C. Spatial scan statistics confirm FGM clusters in the North-Eastern and South-Western regions of Kenya (p<0.001). This suggests that the fight against FGM/C in Kenya is not yet over. There are still deep cultural and religious beliefs to be addressed in a bid to eradicate the practice. Interventions by government and other stakeholders must address these challenges and target the identified clusters.

  20. Randomization and resilience of brain functional networks as systems-level endophenotypes of schizophrenia.

    PubMed

    Lo, Chun-Yi Zac; Su, Tsung-Wei; Huang, Chu-Chung; Hung, Chia-Chun; Chen, Wei-Ling; Lan, Tsuo-Hung; Lin, Ching-Po; Bullmore, Edward T

    2015-07-21

    Schizophrenia is increasingly conceived as a disorder of brain network organization or dysconnectivity syndrome. Functional MRI (fMRI) networks in schizophrenia have been characterized by abnormally random topology. We tested the hypothesis that network randomization is an endophenotype of schizophrenia and therefore evident also in nonpsychotic relatives of patients. Head movement-corrected, resting-state fMRI data were acquired from 25 patients with schizophrenia, 25 first-degree relatives of patients, and 29 healthy volunteers. Graphs were used to model functional connectivity as a set of edges between regional nodes. We estimated the topological efficiency, clustering, degree distribution, resilience, and connection distance (in millimeters) of each functional network. The schizophrenic group demonstrated significant randomization of global network metrics (reduced clustering, greater efficiency), a shift in the degree distribution to a more homogeneous form (fewer hubs), a shift in the distance distribution (proportionally more long-distance edges), and greater resilience to targeted attack on network hubs. The networks of the relatives also demonstrated abnormal randomization and resilience compared with healthy volunteers, but they were typically less topologically abnormal than the patients' networks and did not have abnormal connection distances. We conclude that schizophrenia is associated with replicable and convergent evidence for functional network randomization, and a similar topological profile was evident also in nonpsychotic relatives, suggesting that this is a systems-level endophenotype or marker of familial risk. We speculate that the greater resilience of brain networks may confer some fitness advantages on nonpsychotic relatives that could explain persistence of this endophenotype in the population.

  1. Functional response and capture timing in an individual-based model: predation by northern squawfish (Ptychocheilus oregonensis) on juvenile salmonids in the Columbia River

    USGS Publications Warehouse

    Petersen, James H.; DeAngelis, Donald L.

    1992-01-01

    The behavior of individual northern squawfish (Ptychocheilus oregonensis) preying on juvenile salmonids was modeled to address questions about capture rate and the timing of prey captures (random versus contagious). Prey density, predator weight, prey weight, temperature, and diel feeding pattern were first incorporated into predation equations analogous to Holling Type 2 and Type 3 functional response models. Type 2 and Type 3 equations fit field data from the Columbia River equally well, and both models predicted predation rates on five of seven independent dates. Selecting a functional response type may be complicated by variable predation rates, analytical methods, and assumptions of the model equations. Using the Type 2 functional response, random versus contagious timing of prey capture was tested using two related models. ln the simpler model, salmon captures were assumed to be controlled by a Poisson renewal process; in the second model, several salmon captures were assumed to occur during brief "feeding bouts", modeled with a compound Poisson process. Salmon captures by individual northern squawfish were clustered through time, rather than random, based on comparison of model simulations and field data. The contagious-feeding result suggests that salmonids may be encountered as patches or schools in the river.

  2. A Motor-Gradient and Clustering Model of the Centripetal Motility of MTOCs in Meiosis I of Mouse Oocytes

    PubMed Central

    2016-01-01

    Asters nucleated by Microtubule (MT) organizing centers (MTOCs) converge on chromosomes during spindle assembly in mouse oocytes undergoing meiosis I. Time-lapse imaging suggests that this centripetal motion is driven by a biased ‘search-and-capture’ mechanism. Here, we develop a model of a random walk in a drift field to test the nature of the bias and the spatio-temporal dynamics of the search process. The model is used to optimize the spatial field of drift in simulations, by comparison to experimental motility statistics. In a second step, this optimized gradient is used to determine the location of immobilized dynein motors and MT polymerization parameters, since these are hypothesized to generate the gradient of forces needed to move MTOCs. We compare these scenarios to self-organized mechanisms by which asters have been hypothesized to find the cell-center- MT pushing at the cell-boundary and clustering motor complexes. By minimizing the error between simulation outputs and experiments, we find a model of “pulling” by a gradient of dynein motors alone can drive the centripetal motility. Interestingly, models of passive MT based “pushing” at the cortex, clustering by cross-linking motors and MT-dynamic instability gradients alone, by themselves do not result in the observed motility. The model predicts the sensitivity of the results to motor density and stall force, but not MTs per aster. A hybrid model combining a chromatin-centered immobilized dynein gradient, diffusible minus-end directed clustering motors and pushing at the cell cortex, is required to comprehensively explain the available data. The model makes experimentally testable predictions of a spatial bias and self-organized mechanisms by which MT asters can find the center of a large cell. PMID:27706163

  3. A Motor-Gradient and Clustering Model of the Centripetal Motility of MTOCs in Meiosis I of Mouse Oocytes.

    PubMed

    Khetan, Neha; Athale, Chaitanya A

    2016-10-01

    Asters nucleated by Microtubule (MT) organizing centers (MTOCs) converge on chromosomes during spindle assembly in mouse oocytes undergoing meiosis I. Time-lapse imaging suggests that this centripetal motion is driven by a biased 'search-and-capture' mechanism. Here, we develop a model of a random walk in a drift field to test the nature of the bias and the spatio-temporal dynamics of the search process. The model is used to optimize the spatial field of drift in simulations, by comparison to experimental motility statistics. In a second step, this optimized gradient is used to determine the location of immobilized dynein motors and MT polymerization parameters, since these are hypothesized to generate the gradient of forces needed to move MTOCs. We compare these scenarios to self-organized mechanisms by which asters have been hypothesized to find the cell-center- MT pushing at the cell-boundary and clustering motor complexes. By minimizing the error between simulation outputs and experiments, we find a model of "pulling" by a gradient of dynein motors alone can drive the centripetal motility. Interestingly, models of passive MT based "pushing" at the cortex, clustering by cross-linking motors and MT-dynamic instability gradients alone, by themselves do not result in the observed motility. The model predicts the sensitivity of the results to motor density and stall force, but not MTs per aster. A hybrid model combining a chromatin-centered immobilized dynein gradient, diffusible minus-end directed clustering motors and pushing at the cell cortex, is required to comprehensively explain the available data. The model makes experimentally testable predictions of a spatial bias and self-organized mechanisms by which MT asters can find the center of a large cell.

  4. Alcohol-Specific Parenting within a Cluster-Randomized Effectiveness Trial of a Swedish Primary Prevention Program

    ERIC Educational Resources Information Center

    Strandberg, Anna K.; Bodin, Maria C.

    2011-01-01

    Purpose: Within the framework of an ongoing cluster-randomized effectiveness trial of a parental prevention program, the aim of the present study is to investigate attitudes towards under-age drinking and use of program components, i.e. alcohol-specific parenting behaviors, in parents who did and did not take part in the programme.…

  5. The YouthMood Project: A Cluster Randomized Controlled Trial of an Online Cognitive Behavioral Program with Adolescents

    ERIC Educational Resources Information Center

    Calear, Alison L.; Christensen, Helen; Mackinnon, Andrew; Griffiths, Kathleen M.; O'Kearney, Richard

    2009-01-01

    The aim in the current study was to investigate the effectiveness of an online, self-directed cognitive-behavioral therapy program (MoodGYM) in preventing and reducing the symptoms of anxiety and depression in an adolescent school-based population. A cluster randomized controlled trial was conducted with 30 schools (N = 1,477) from across…

  6. Assessment Data-Informed Guidance to Individualize Kindergarten Reading Instruction: Findings from a Cluster-Randomized Control Field Trial

    ERIC Educational Resources Information Center

    Al Otaiba, Stephanie; Connor, Carol M.; Folsom, Jessica S.; Greulich, Luana; Meadows, Jane; Li, Zhi

    2011-01-01

    The purpose of this cluster-randomized control field trial was to examine whether kindergarten teachers could learn to differentiate classroom reading instruction using Individualized Student Instruction for Kindergarten (ISI-K) and to test the efficacy of differentiation on reading outcomes. The study involved 14 schools, 23 ISI-K (n = 305…

  7. The Effects of Therapist Competence in Assigning Homework in Cognitive Therapy with Cluster C Personality Disorders: Results from a Randomized Controlled Trial

    ERIC Educational Resources Information Center

    Ryum, Truls; Stiles, Tore C.; Svartberg, Martin; McCullough, Leigh

    2010-01-01

    Therapist competence in assigning homework was used to predict mid- and posttreatment outcome for patients with Cluster C personality disorders in cognitive therapy (CT). Twenty-five patients that underwent 40 sessions of CT were taken from a randomized controlled trial (Svartberg, Stiles, & Seltzer, 2004). Therapist competence in assigning…

  8. A General Framework for Power Analysis to Detect the Moderator Effects in Two- and Three-Level Cluster Randomized Trials

    ERIC Educational Resources Information Center

    Dong, Nianbo; Spybrook, Jessaca; Kelcey, Ben

    2016-01-01

    The purpose of this study is to propose a general framework for power analyses to detect the moderator effects in two- and three-level cluster randomized trials (CRTs). The study specifically aims to: (1) develop the statistical formulations for calculating statistical power, minimum detectable effect size (MDES) and its confidence interval to…

  9. The Impact of Curriculum-Based Professional Development on Science Instruction: Results from a Cluster-Randomized Trial

    ERIC Educational Resources Information Center

    Taylor, Joseph; Kowalski, Susan; Getty, Stephen; Wilson, Christopher; Carlson, Janet

    2011-01-01

    This research is part of a larger, IES-funded study titled: "Measuring the Efficacy and Student Achievement of Research-based Instructional Materials in High School Multidisciplinary Science" (Award # R305K060142). The larger study seeks to use a cluster-randomized trial design, with schools as the unit of assignment, to make causal…

  10. A Clustered Randomized Controlled Trial to Determine Impacts of the Harvest of the Month Program

    ERIC Educational Resources Information Center

    LaChausse, Robert G.

    2017-01-01

    The study purpose was to examine the impact of the Harvest of the Month (HOTM) program on fruit and vegetable (FV) consumption, FV preferences, other eating behaviors, physical activity and other variables related to healthy eating. A clustered randomized controlled trial was employed in 28 elementary schools. After parental consent was obtained,…

  11. Impact of a Social-Emotional and Character Development Program on School-Level Indicators of Academic Achievement, Absenteeism, and Disciplinary Outcomes: A Matched-Pair, Cluster-Randomized, Controlled Trial

    ERIC Educational Resources Information Center

    Snyder, Frank; Flay, Brian; Vuchinich, Samuel; Acock, Alan; Washburn, Isaac; Beets, Michael; Li, Kin-Kit

    2010-01-01

    This article reports the effects of a comprehensive elementary school-based social-emotional and character education program on school-level achievement, absenteeism, and disciplinary outcomes utilizing a matched-pair, cluster-randomized, controlled design. The "Positive Action" Hawai'i trial included 20 racially/ethnically diverse…

  12. Cluster Randomized-Controlled Trial of Interventions to Improve Health for Adults with Intellectual Disability Who Live in Private Dwellings

    ERIC Educational Resources Information Center

    Lennox, Nicholas; Bain, Chris; Rey-Conde, Therese; Taylor, Miriam; Boyle, Frances M.; Purdie, David M.; Ware, Robert S.

    2010-01-01

    Background: People with intellectual disability who live in the community often have poor health and healthcare, partly as a consequence of poor communication, recall difficulties and incomplete patient health information. Materials and Methods: A cluster randomized-controlled trial with 2 x 2 factorial design was conducted with adults with…

  13. Pathways to Health: A Cluster Randomized Trial of Nicotine Gum and Motivational Interviewing for Smoking Cessation in Low-Income Housing

    ERIC Educational Resources Information Center

    Okuyemi, Kolawole S.; James, Aimee S.; Mayo, Matthew S.; Nollen, Nicole; Catley, Delwyn; Choi, Won S.; Ahluwalia, Jasjit S.

    2007-01-01

    Despite high smoking rates among those living in poverty, few cessation studies are conducted in these populations. This cluster-randomized trial tested nicotine gum plus motivational interviewing (MI) for smoking cessation in 20 low-income housing developments (HDs). Intervention participants (10 HDs, n = 66) received educational materials, 8…

  14. Statistical mechanics of high-density bond percolation

    NASA Astrophysics Data System (ADS)

    Timonin, P. N.

    2018-05-01

    High-density (HD) percolation describes the percolation of specific κ -clusters, which are the compact sets of sites each connected to κ nearest filled sites at least. It takes place in the classical patterns of independently distributed sites or bonds in which the ordinary percolation transition also exists. Hence, the study of series of κ -type HD percolations amounts to the description of classical clusters' structure for which κ -clusters constitute κ -cores nested one into another. Such data are needed for description of a number of physical, biological, and information properties of complex systems on random lattices, graphs, and networks. They range from magnetic properties of semiconductor alloys to anomalies in supercooled water and clustering in biological and social networks. Here we present the statistical mechanics approach to study HD bond percolation on an arbitrary graph. It is shown that the generating function for κ -clusters' size distribution can be obtained from the partition function of the specific q -state Potts-Ising model in the q →1 limit. Using this approach we find exact κ -clusters' size distributions for the Bethe lattice and Erdos-Renyi graph. The application of the method to Euclidean lattices is also discussed.

  15. Disease Mapping of Zero-excessive Mesothelioma Data in Flanders

    PubMed Central

    Neyens, Thomas; Lawson, Andrew B.; Kirby, Russell S.; Nuyts, Valerie; Watjou, Kevin; Aregay, Mehreteab; Carroll, Rachel; Nawrot, Tim S.; Faes, Christel

    2016-01-01

    Purpose To investigate the distribution of mesothelioma in Flanders using Bayesian disease mapping models that account for both an excess of zeros and overdispersion. Methods The numbers of newly diagnosed mesothelioma cases within all Flemish municipalities between 1999 and 2008 were obtained from the Belgian Cancer Registry. To deal with overdispersion, zero-inflation and geographical association, the hurdle combined model was proposed, which has three components: a Bernoulli zero-inflation mixture component to account for excess zeros, a gamma random effect to adjust for overdispersion and a normal conditional autoregressive random effect to attribute spatial association. This model was compared with other existing methods in literature. Results The results indicate that hurdle models with a random effects term accounting for extra-variance in the Bernoulli zero-inflation component fit the data better than hurdle models that do not take overdispersion in the occurrence of zeros into account. Furthermore, traditional models that do not take into account excessive zeros but contain at least one random effects term that models extra-variance in the counts have better fits compared to their hurdle counterparts. In other words, the extra-variability, due to an excess of zeros, can be accommodated by spatially structured and/or unstructured random effects in a Poisson model such that the hurdle mixture model is not necessary. Conclusions Models taking into account zero-inflation do not always provide better fits to data with excessive zeros than less complex models. In this study, a simple conditional autoregressive model identified a cluster in mesothelioma cases near a former asbestos processing plant (Kapelle-op-den-Bos). This observation is likely linked with historical local asbestos exposures. Future research will clarify this. PMID:27908590

  16. Disease mapping of zero-excessive mesothelioma data in Flanders.

    PubMed

    Neyens, Thomas; Lawson, Andrew B; Kirby, Russell S; Nuyts, Valerie; Watjou, Kevin; Aregay, Mehreteab; Carroll, Rachel; Nawrot, Tim S; Faes, Christel

    2017-01-01

    To investigate the distribution of mesothelioma in Flanders using Bayesian disease mapping models that account for both an excess of zeros and overdispersion. The numbers of newly diagnosed mesothelioma cases within all Flemish municipalities between 1999 and 2008 were obtained from the Belgian Cancer Registry. To deal with overdispersion, zero inflation, and geographical association, the hurdle combined model was proposed, which has three components: a Bernoulli zero-inflation mixture component to account for excess zeros, a gamma random effect to adjust for overdispersion, and a normal conditional autoregressive random effect to attribute spatial association. This model was compared with other existing methods in literature. The results indicate that hurdle models with a random effects term accounting for extra variance in the Bernoulli zero-inflation component fit the data better than hurdle models that do not take overdispersion in the occurrence of zeros into account. Furthermore, traditional models that do not take into account excessive zeros but contain at least one random effects term that models extra variance in the counts have better fits compared to their hurdle counterparts. In other words, the extra variability, due to an excess of zeros, can be accommodated by spatially structured and/or unstructured random effects in a Poisson model such that the hurdle mixture model is not necessary. Models taking into account zero inflation do not always provide better fits to data with excessive zeros than less complex models. In this study, a simple conditional autoregressive model identified a cluster in mesothelioma cases near a former asbestos processing plant (Kapelle-op-den-Bos). This observation is likely linked with historical local asbestos exposures. Future research will clarify this. Copyright © 2016 Elsevier Inc. All rights reserved.

  17. Analysis of partially observed clustered data using generalized estimating equations and multiple imputation

    PubMed Central

    Aloisio, Kathryn M.; Swanson, Sonja A.; Micali, Nadia; Field, Alison; Horton, Nicholas J.

    2015-01-01

    Clustered data arise in many settings, particularly within the social and biomedical sciences. As an example, multiple–source reports are commonly collected in child and adolescent psychiatric epidemiologic studies where researchers use various informants (e.g. parent and adolescent) to provide a holistic view of a subject’s symptomatology. Fitzmaurice et al. (1995) have described estimation of multiple source models using a standard generalized estimating equation (GEE) framework. However, these studies often have missing data due to additional stages of consent and assent required. The usual GEE is unbiased when missingness is Missing Completely at Random (MCAR) in the sense of Little and Rubin (2002). This is a strong assumption that may not be tenable. Other options such as weighted generalized estimating equations (WEEs) are computationally challenging when missingness is non–monotone. Multiple imputation is an attractive method to fit incomplete data models while only requiring the less restrictive Missing at Random (MAR) assumption. Previously estimation of partially observed clustered data was computationally challenging however recent developments in Stata have facilitated their use in practice. We demonstrate how to utilize multiple imputation in conjunction with a GEE to investigate the prevalence of disordered eating symptoms in adolescents reported by parents and adolescents as well as factors associated with concordance and prevalence. The methods are motivated by the Avon Longitudinal Study of Parents and their Children (ALSPAC), a cohort study that enrolled more than 14,000 pregnant mothers in 1991–92 and has followed the health and development of their children at regular intervals. While point estimates were fairly similar to the GEE under MCAR, the MAR model had smaller standard errors, while requiring less stringent assumptions regarding missingness. PMID:25642154

  18. A cluster randomized control field trial of the ABRACADABRA web-based reading technology: replication and extension of basic findings

    PubMed Central

    Piquette, Noella A.; Savage, Robert S.; Abrami, Philip C.

    2014-01-01

    The present paper reports a cluster randomized control trial evaluation of teaching using ABRACADABRA (ABRA), an evidence-based and web-based literacy intervention (http://abralite.concordia.ca) with 107 kindergarten and 96 grade 1 children in 24 classes (12 intervention 12 control classes) from all 12 elementary schools in one school district in Canada. Children in the intervention condition received 10–12 h of whole class instruction using ABRA between pre- and post-test. Hierarchical linear modeling of post-test results showed significant gains in letter-sound knowledge for intervention classrooms over control classrooms. In addition, medium effect sizes were evident for three of five outcome measures favoring the intervention: letter-sound knowledge (d= +0.66), phonological blending (d = +0.52), and word reading (d = +0.52), over effect sizes for regular teaching. It is concluded that regular teaching with ABRA technology adds significantly to literacy in the early elementary years. PMID:25538663

  19. Bagging Voronoi classifiers for clustering spatial functional data

    NASA Astrophysics Data System (ADS)

    Secchi, Piercesare; Vantini, Simone; Vitelli, Valeria

    2013-06-01

    We propose a bagging strategy based on random Voronoi tessellations for the exploration of geo-referenced functional data, suitable for different purposes (e.g., classification, regression, dimensional reduction, …). Urged by an application to environmental data contained in the Surface Solar Energy database, we focus in particular on the problem of clustering functional data indexed by the sites of a spatial finite lattice. We thus illustrate our strategy by implementing a specific algorithm whose rationale is to (i) replace the original data set with a reduced one, composed by local representatives of neighborhoods covering the entire investigated area; (ii) analyze the local representatives; (iii) repeat the previous analysis many times for different reduced data sets associated to randomly generated different sets of neighborhoods, thus obtaining many different weak formulations of the analysis; (iv) finally, bag together the weak analyses to obtain a conclusive strong analysis. Through an extensive simulation study, we show that this new procedure - which does not require an explicit model for spatial dependence - is statistically and computationally efficient.

  20. Zipf's law from scale-free geometry.

    PubMed

    Lin, Henry W; Loeb, Abraham

    2016-03-01

    The spatial distribution of people exhibits clustering across a wide range of scales, from household (∼10(-2) km) to continental (∼10(4) km) scales. Empirical data indicate simple power-law scalings for the size distribution of cities (known as Zipf's law) and the population density fluctuations as a function of scale. Using techniques from random field theory and statistical physics, we show that these power laws are fundamentally a consequence of the scale-free spatial clustering of human populations and the fact that humans inhabit a two-dimensional surface. In this sense, the symmetries of scale invariance in two spatial dimensions are intimately connected to urban sociology. We test our theory by empirically measuring the power spectrum of population density fluctuations and show that the logarithmic slope α=2.04 ± 0.09, in excellent agreement with our theoretical prediction α=2. The model enables the analytic computation of many new predictions by importing the mathematical formalism of random fields.

  1. A tripartite clustering analysis on microRNA, gene and disease model.

    PubMed

    Shen, Chengcheng; Liu, Ying

    2012-02-01

    Alteration of gene expression in response to regulatory molecules or mutations could lead to different diseases. MicroRNAs (miRNAs) have been discovered to be involved in regulation of gene expression and a wide variety of diseases. In a tripartite biological network of human miRNAs, their predicted target genes and the diseases caused by altered expressions of these genes, valuable knowledge about the pathogenicity of miRNAs, involved genes and related disease classes can be revealed by co-clustering miRNAs, target genes and diseases simultaneously. Tripartite co-clustering can lead to more informative results than traditional co-clustering with only two kinds of members and pass the hidden relational information along the relation chain by considering multi-type members. Here we report a spectral co-clustering algorithm for k-partite graph to find clusters with heterogeneous members. We use the method to explore the potential relationships among miRNAs, genes and diseases. The clusters obtained from the algorithm have significantly higher density than randomly selected clusters, which means members in the same cluster are more likely to have common connections. Results also show that miRNAs in the same family based on the hairpin sequences tend to belong to the same cluster. We also validate the clustering results by checking the correlation of enriched gene functions and disease classes in the same cluster. Finally, widely studied miR-17-92 and its paralogs are analyzed as a case study to reveal that genes and diseases co-clustered with the miRNAs are in accordance with current research findings.

  2. Sampling in health geography: reconciling geographical objectives and probabilistic methods. An example of a health survey in Vientiane (Lao PDR)

    PubMed Central

    Vallée, Julie; Souris, Marc; Fournet, Florence; Bochaton, Audrey; Mobillion, Virginie; Peyronnie, Karine; Salem, Gérard

    2007-01-01

    Background Geographical objectives and probabilistic methods are difficult to reconcile in a unique health survey. Probabilistic methods focus on individuals to provide estimates of a variable's prevalence with a certain precision, while geographical approaches emphasise the selection of specific areas to study interactions between spatial characteristics and health outcomes. A sample selected from a small number of specific areas creates statistical challenges: the observations are not independent at the local level, and this results in poor statistical validity at the global level. Therefore, it is difficult to construct a sample that is appropriate for both geographical and probability methods. Methods We used a two-stage selection procedure with a first non-random stage of selection of clusters. Instead of randomly selecting clusters, we deliberately chose a group of clusters, which as a whole would contain all the variation in health measures in the population. As there was no health information available before the survey, we selected a priori determinants that can influence the spatial homogeneity of the health characteristics. This method yields a distribution of variables in the sample that closely resembles that in the overall population, something that cannot be guaranteed with randomly-selected clusters, especially if the number of selected clusters is small. In this way, we were able to survey specific areas while minimising design effects and maximising statistical precision. Application We applied this strategy in a health survey carried out in Vientiane, Lao People's Democratic Republic. We selected well-known health determinants with unequal spatial distribution within the city: nationality and literacy. We deliberately selected a combination of clusters whose distribution of nationality and literacy is similar to the distribution in the general population. Conclusion This paper describes the conceptual reasoning behind the construction of the survey sample and shows that it can be advantageous to choose clusters using reasoned hypotheses, based on both probability and geographical approaches, in contrast to a conventional, random cluster selection strategy. PMID:17543100

  3. Sampling in health geography: reconciling geographical objectives and probabilistic methods. An example of a health survey in Vientiane (Lao PDR).

    PubMed

    Vallée, Julie; Souris, Marc; Fournet, Florence; Bochaton, Audrey; Mobillion, Virginie; Peyronnie, Karine; Salem, Gérard

    2007-06-01

    Geographical objectives and probabilistic methods are difficult to reconcile in a unique health survey. Probabilistic methods focus on individuals to provide estimates of a variable's prevalence with a certain precision, while geographical approaches emphasise the selection of specific areas to study interactions between spatial characteristics and health outcomes. A sample selected from a small number of specific areas creates statistical challenges: the observations are not independent at the local level, and this results in poor statistical validity at the global level. Therefore, it is difficult to construct a sample that is appropriate for both geographical and probability methods. We used a two-stage selection procedure with a first non-random stage of selection of clusters. Instead of randomly selecting clusters, we deliberately chose a group of clusters, which as a whole would contain all the variation in health measures in the population. As there was no health information available before the survey, we selected a priori determinants that can influence the spatial homogeneity of the health characteristics. This method yields a distribution of variables in the sample that closely resembles that in the overall population, something that cannot be guaranteed with randomly-selected clusters, especially if the number of selected clusters is small. In this way, we were able to survey specific areas while minimising design effects and maximising statistical precision. We applied this strategy in a health survey carried out in Vientiane, Lao People's Democratic Republic. We selected well-known health determinants with unequal spatial distribution within the city: nationality and literacy. We deliberately selected a combination of clusters whose distribution of nationality and literacy is similar to the distribution in the general population. This paper describes the conceptual reasoning behind the construction of the survey sample and shows that it can be advantageous to choose clusters using reasoned hypotheses, based on both probability and geographical approaches, in contrast to a conventional, random cluster selection strategy.

  4. Effect of Reassuring Information About Musculoskeletal and Mental Health Complaints at the Workplace: A Cluster Randomized Trial of the atWork Intervention.

    PubMed

    Johnsen, Tone Langjordet; Eriksen, Hege Randi; Baste, Valborg; Indahl, Aage; Odeen, Magnus; Tveito, Torill Helene

    2018-05-21

    Purpose The purpose of this study was to investigate the possible difference between the Modified atWork intervention (MAW) and the Original atWork intervention (OAW) on sick leave and other health related outcomes. atWork is a group intervention using the workplace as an arena for distribution of evidence-based knowledge about musculoskeletal and mental health complaints. Methods A cluster randomized controlled trial with 93 kindergartens, comprising a total of 1011 employees, was conducted. Kindergartens were stratified by county and size and randomly allocated to MAW (45 clusters, 324 respondents) or OAW (48 clusters, 313 respondents). The randomization and intervention allocation processes were concealed. There was no blinding to group allocation. Primary outcome was register data on sick leave at cluster level. Secondary outcomes were health complaints, job satisfaction, social support, coping, and beliefs about musculoskeletal and mental health complaints, measured at the individual level. Results The MAW group reduced sick leave by 5.7% during the intervention year, while the OAW group had a 7.5% increase. Overall, the changes were not statistically significant, and no difference was detected between groups, based on 45 and 47 kindergartens. Compared to the OAW group, the MAW group had a smaller reduction for two of the statements concerning faulty beliefs about back pain, but believed less in the hereditary nature of depression. Conclusions The MAW did not have a different effect on sick leave at cluster level compared to the OAW. Trial registration https://Clinicaltrials.gov/ : NCT02396797. Registered March 23th, 2015.

  5. Viscoelasticity promotes collective swimming of sperm

    NASA Astrophysics Data System (ADS)

    Tung, Chih-Kuan; Harvey, Benedict B.; Fiore, Alyssa G.; Ardon, Florencia; Suarez, Susan S.; Wu, Mingming

    From flocking birds to swarming insects, interactions of organisms large and small lead to the emergence of collective dynamics. Here, we report striking collective swimming of bovine sperm, with sperm orienting in the same direction within each cluster, enabled by the viscoelasticity of the fluid. A long-chain polyacrylamide solution was used as a model viscoelastic fluid such that its rheology can be fine-tuned to mimic that of bovine cervical mucus. In viscoelastic fluid, sperm formed dynamic clusters, and the cluster size increased with elasticity of the polyacrylamide solution. In contrast, sperm swam randomly and individually in Newtonian fluids of similar viscosity. Analysis of the fluid motion surrounding individual swimming sperm indicated that sperm-fluid interaction is facilitated by the elastic component of the fluid. We note that almost all biological fluids (e.g. mucus and blood) are viscoelastic in nature, this finding highlights the importance of fluid elasticity in biological function. We will discuss what the orientation fluctuation within a cluster reveals about the interaction strength. Supported by NIH Grant 1R01HD070038.

  6. Atom Probe Tomography Analysis of the Distribution of Rhenium in Nickel Alloys

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mottura, A.; Warnken, N; Miller, Michael K

    2010-01-01

    Atom probe tomography (APT) is used to characterise the distributions of rhenium in a binary Ni-Re alloy and the nickel-based single-crystal CMSX-4 superalloy. A purpose-built algorithm is developed to quantify the size distribution of solute clusters, and applied to the APT datasets to critique the hypothesis that rhenium is prone to the formation of clusters in these systems. No evidence is found to indicate that rhenium forms solute clusters above the level expected from random fluctuations. In CMSX-4, enrichment of Re is detected in the matrix phase close to the matrix/precipitate ({gamma}/{gamma}{prime}) phase boundaries. Phase field modelling indicates that thismore » is due to the migration of the {gamma}/{gamma}{prime} interface during cooling from the temperature of operation. Thus, neither clustering of rhenium nor interface enrichments can be the cause of the enhancement in high temperature mechanical properties conferred by rhenium alloying.« less

  7. Cluster ensemble based on Random Forests for genetic data.

    PubMed

    Alhusain, Luluah; Hafez, Alaaeldin M

    2017-01-01

    Clustering plays a crucial role in several application domains, such as bioinformatics. In bioinformatics, clustering has been extensively used as an approach for detecting interesting patterns in genetic data. One application is population structure analysis, which aims to group individuals into subpopulations based on shared genetic variations, such as single nucleotide polymorphisms. Advances in DNA sequencing technology have facilitated the obtainment of genetic datasets with exceptional sizes. Genetic data usually contain hundreds of thousands of genetic markers genotyped for thousands of individuals, making an efficient means for handling such data desirable. Random Forests (RFs) has emerged as an efficient algorithm capable of handling high-dimensional data. RFs provides a proximity measure that can capture different levels of co-occurring relationships between variables. RFs has been widely considered a supervised learning method, although it can be converted into an unsupervised learning method. Therefore, RF-derived proximity measure combined with a clustering technique may be well suited for determining the underlying structure of unlabeled data. This paper proposes, RFcluE, a cluster ensemble approach for determining the underlying structure of genetic data based on RFs. The approach comprises a cluster ensemble framework to combine multiple runs of RF clustering. Experiments were conducted on high-dimensional, real genetic dataset to evaluate the proposed approach. The experiments included an examination of the impact of parameter changes, comparing RFcluE performance against other clustering methods, and an assessment of the relationship between the diversity and quality of the ensemble and its effect on RFcluE performance. This paper proposes, RFcluE, a cluster ensemble approach based on RF clustering to address the problem of population structure analysis and demonstrate the effectiveness of the approach. The paper also illustrates that applying a cluster ensemble approach, combining multiple RF clusterings, produces more robust and higher-quality results as a consequence of feeding the ensemble with diverse views of high-dimensional genetic data obtained through bagging and random subspace, the two key features of the RF algorithm.

  8. Significant locations in auxiliary data as seeds for typical use cases of point clustering

    NASA Astrophysics Data System (ADS)

    Kröger, Johannes

    2018-05-01

    Random greedy clustering and grid-based clustering are highly susceptible by their initial parameters. When used for point data clustering in maps they often change the apparent distribution of the underlying data. We propose a process that uses precomputed weighted seed points for the initialization of clusters, for example from local maxima in population density data. Exemplary results from the clustering of a dataset of petrol stations are presented.

  9. Composition, morphology, and growth of clusters in a gas of particles with random interactions

    NASA Astrophysics Data System (ADS)

    Azizi, Itay; Rabin, Yitzhak

    2018-03-01

    We use Langevin dynamics simulations to study the growth kinetics and the steady-state properties of condensed clusters in a dilute two-dimensional system of particles that are all different (APD) in the sense that each particle is characterized by a randomly chosen interaction parameter. The growth exponents, the transition temperatures, and the steady-state properties of the clusters and of the surrounding gas phase are obtained and compared with those of one-component systems. We investigate the fractionation phenomenon, i.e., how particles of different identities are distributed between the coexisting mother (gas) and daughter (clusters) phases. We study the local organization of particles inside clusters, according to their identity—neighbourhood identity ordering (NIO)—and compare the results with those of previous studies of NIO in dense APD systems.

  10. Subjective randomness as statistical inference.

    PubMed

    Griffiths, Thomas L; Daniels, Dylan; Austerweil, Joseph L; Tenenbaum, Joshua B

    2018-06-01

    Some events seem more random than others. For example, when tossing a coin, a sequence of eight heads in a row does not seem very random. Where do these intuitions about randomness come from? We argue that subjective randomness can be understood as the result of a statistical inference assessing the evidence that an event provides for having been produced by a random generating process. We show how this account provides a link to previous work relating randomness to algorithmic complexity, in which random events are those that cannot be described by short computer programs. Algorithmic complexity is both incomputable and too general to capture the regularities that people can recognize, but viewing randomness as statistical inference provides two paths to addressing these problems: considering regularities generated by simpler computing machines, and restricting the set of probability distributions that characterize regularity. Building on previous work exploring these different routes to a more restricted notion of randomness, we define strong quantitative models of human randomness judgments that apply not just to binary sequences - which have been the focus of much of the previous work on subjective randomness - but also to binary matrices and spatial clustering. Copyright © 2018 Elsevier Inc. All rights reserved.

  11. RRW: repeated random walks on genome-scale protein networks for local cluster discovery

    PubMed Central

    Macropol, Kathy; Can, Tolga; Singh, Ambuj K

    2009-01-01

    Background We propose an efficient and biologically sensitive algorithm based on repeated random walks (RRW) for discovering functional modules, e.g., complexes and pathways, within large-scale protein networks. Compared to existing cluster identification techniques, RRW implicitly makes use of network topology, edge weights, and long range interactions between proteins. Results We apply the proposed technique on a functional network of yeast genes and accurately identify statistically significant clusters of proteins. We validate the biological significance of the results using known complexes in the MIPS complex catalogue database and well-characterized biological processes. We find that 90% of the created clusters have the majority of their catalogued proteins belonging to the same MIPS complex, and about 80% have the majority of their proteins involved in the same biological process. We compare our method to various other clustering techniques, such as the Markov Clustering Algorithm (MCL), and find a significant improvement in the RRW clusters' precision and accuracy values. Conclusion RRW, which is a technique that exploits the topology of the network, is more precise and robust in finding local clusters. In addition, it has the added flexibility of being able to find multi-functional proteins by allowing overlapping clusters. PMID:19740439

  12. Multirate parallel distributed compensation of a cluster in wireless sensor and actor networks

    NASA Astrophysics Data System (ADS)

    Yang, Chun-xi; Huang, Ling-yun; Zhang, Hao; Hua, Wang

    2016-01-01

    The stabilisation problem for one of the clusters with bounded multiple random time delays and packet dropouts in wireless sensor and actor networks is investigated in this paper. A new multirate switching model is constructed to describe the feature of this single input multiple output linear system. According to the difficulty of controller design under multi-constraints in multirate switching model, this model can be converted to a Takagi-Sugeno fuzzy model. By designing a multirate parallel distributed compensation, a sufficient condition is established to ensure this closed-loop fuzzy control system to be globally exponentially stable. The solution of the multirate parallel distributed compensation gains can be obtained by solving an auxiliary convex optimisation problem. Finally, two numerical examples are given to show, compared with solving switching controller, multirate parallel distributed compensation can be obtained easily. Furthermore, it has stronger robust stability than arbitrary switching controller and single-rate parallel distributed compensation under the same conditions.

  13. Memory-induced mechanism for self-sustaining activity in networks

    NASA Astrophysics Data System (ADS)

    Allahverdyan, A. E.; Steeg, G. Ver; Galstyan, A.

    2015-12-01

    We study a mechanism of activity sustaining on networks inspired by a well-known model of neuronal dynamics. Our primary focus is the emergence of self-sustaining collective activity patterns, where no single node can stay active by itself, but the activity provided initially is sustained within the collective of interacting agents. In contrast to existing models of self-sustaining activity that are caused by (long) loops present in the network, here we focus on treelike structures and examine activation mechanisms that are due to temporal memory of the nodes. This approach is motivated by applications in social media, where long network loops are rare or absent. Our results suggest that under a weak behavioral noise, the nodes robustly split into several clusters, with partial synchronization of nodes within each cluster. We also study the randomly weighted version of the models where the nodes are allowed to change their connection strength (this can model attention redistribution) and show that it does facilitate the self-sustained activity.

  14. Sudden spreading of infections in an epidemic model with a finite seed fraction

    NASA Astrophysics Data System (ADS)

    Hasegawa, Takehisa; Nemoto, Koji

    2018-03-01

    We study a simple case of the susceptible-weakened-infected-removed model in regular random graphs in a situation where an epidemic starts from a finite fraction of initially infected nodes (seeds). Previous studies have shown that, assuming a single seed, this model exhibits a kind of discontinuous transition at a certain value of infection rate. Performing Monte Carlo simulations and evaluating approximate master equations, we find that the present model has two critical infection rates for the case with a finite seed fraction. At the first critical rate the system shows a percolation transition of clusters composed of removed nodes, and at the second critical rate, which is larger than the first one, a giant cluster suddenly grows and the order parameter jumps even though it has been already rising. Numerical evaluation of the master equations shows that such sudden epidemic spreading does occur if the degree of the underlying network is large and the seed fraction is small.

  15. Insulin Resistance: Regression and Clustering

    PubMed Central

    Yoon, Sangho; Assimes, Themistocles L.; Quertermous, Thomas; Hsiao, Chin-Fu; Chuang, Lee-Ming; Hwu, Chii-Min; Rajaratnam, Bala; Olshen, Richard A.

    2014-01-01

    In this paper we try to define insulin resistance (IR) precisely for a group of Chinese women. Our definition deliberately does not depend upon body mass index (BMI) or age, although in other studies, with particular random effects models quite different from models used here, BMI accounts for a large part of the variability in IR. We accomplish our goal through application of Gauss mixture vector quantization (GMVQ), a technique for clustering that was developed for application to lossy data compression. Defining data come from measurements that play major roles in medical practice. A precise statement of what the data are is in Section 1. Their family structures are described in detail. They concern levels of lipids and the results of an oral glucose tolerance test (OGTT). We apply GMVQ to residuals obtained from regressions of outcomes of an OGTT and lipids on functions of age and BMI that are inferred from the data. A bootstrap procedure developed for our family data supplemented by insights from other approaches leads us to believe that two clusters are appropriate for defining IR precisely. One cluster consists of women who are IR, and the other of women who seem not to be. Genes and other features are used to predict cluster membership. We argue that prediction with “main effects” is not satisfactory, but prediction that includes interactions may be. PMID:24887437

  16. Phase synchronization of bursting neurons in clustered small-world networks

    NASA Astrophysics Data System (ADS)

    Batista, C. A. S.; Lameu, E. L.; Batista, A. M.; Lopes, S. R.; Pereira, T.; Zamora-López, G.; Kurths, J.; Viana, R. L.

    2012-07-01

    We investigate the collective dynamics of bursting neurons on clustered networks. The clustered network model is composed of subnetworks, each of them presenting the so-called small-world property. This model can also be regarded as a network of networks. In each subnetwork a neuron is connected to other ones with regular as well as random connections, the latter with a given intracluster probability. Moreover, in a given subnetwork each neuron has an intercluster probability to be connected to the other subnetworks. The local neuron dynamics has two time scales (fast and slow) and is modeled by a two-dimensional map. In such small-world network the neuron parameters are chosen to be slightly different such that, if the coupling strength is large enough, there may be synchronization of the bursting (slow) activity. We give bounds for the critical coupling strength to obtain global burst synchronization in terms of the network structure, that is, the probabilities of intracluster and intercluster connections. We find that, as the heterogeneity in the network is reduced, the network global synchronizability is improved. We show that the transitions to global synchrony may be abrupt or smooth depending on the intercluster probability.

  17. A Highly Efficient Design Strategy for Regression with Outcome Pooling

    PubMed Central

    Mitchell, Emily M.; Lyles, Robert H.; Manatunga, Amita K.; Perkins, Neil J.; Schisterman, Enrique F.

    2014-01-01

    The potential for research involving biospecimens can be hindered by the prohibitive cost of performing laboratory assays on individual samples. To mitigate this cost, strategies such as randomly selecting a portion of specimens for analysis or randomly pooling specimens prior to performing laboratory assays may be employed. These techniques, while effective in reducing cost, are often accompanied by a considerable loss of statistical efficiency. We propose a novel pooling strategy based on the k-means clustering algorithm to reduce laboratory costs while maintaining a high level of statistical efficiency when predictor variables are measured on all subjects, but the outcome of interest is assessed in pools. We perform simulations motivated by the BioCycle study to compare this k-means pooling strategy with current pooling and selection techniques under simple and multiple linear regression models. While all of the methods considered produce unbiased estimates and confidence intervals with appropriate coverage, pooling under k-means clustering provides the most precise estimates, closely approximating results from the full data and losing minimal precision as the total number of pools decreases. The benefits of k-means clustering evident in the simulation study are then applied to an analysis of the BioCycle dataset. In conclusion, when the number of lab tests is limited by budget, pooling specimens based on k-means clustering prior to performing lab assays can be an effective way to save money with minimal information loss in a regression setting. PMID:25220822

  18. A highly efficient design strategy for regression with outcome pooling.

    PubMed

    Mitchell, Emily M; Lyles, Robert H; Manatunga, Amita K; Perkins, Neil J; Schisterman, Enrique F

    2014-12-10

    The potential for research involving biospecimens can be hindered by the prohibitive cost of performing laboratory assays on individual samples. To mitigate this cost, strategies such as randomly selecting a portion of specimens for analysis or randomly pooling specimens prior to performing laboratory assays may be employed. These techniques, while effective in reducing cost, are often accompanied by a considerable loss of statistical efficiency. We propose a novel pooling strategy based on the k-means clustering algorithm to reduce laboratory costs while maintaining a high level of statistical efficiency when predictor variables are measured on all subjects, but the outcome of interest is assessed in pools. We perform simulations motivated by the BioCycle study to compare this k-means pooling strategy with current pooling and selection techniques under simple and multiple linear regression models. While all of the methods considered produce unbiased estimates and confidence intervals with appropriate coverage, pooling under k-means clustering provides the most precise estimates, closely approximating results from the full data and losing minimal precision as the total number of pools decreases. The benefits of k-means clustering evident in the simulation study are then applied to an analysis of the BioCycle dataset. In conclusion, when the number of lab tests is limited by budget, pooling specimens based on k-means clustering prior to performing lab assays can be an effective way to save money with minimal information loss in a regression setting. Copyright © 2014 John Wiley & Sons, Ltd.

  19. Impact of non-uniform correlation structure on sample size and power in multiple-period cluster randomised trials.

    PubMed

    Kasza, J; Hemming, K; Hooper, R; Matthews, Jns; Forbes, A B

    2017-01-01

    Stepped wedge and cluster randomised crossover trials are examples of cluster randomised designs conducted over multiple time periods that are being used with increasing frequency in health research. Recent systematic reviews of both of these designs indicate that the within-cluster correlation is typically taken account of in the analysis of data using a random intercept mixed model, implying a constant correlation between any two individuals in the same cluster no matter how far apart in time they are measured: within-period and between-period intra-cluster correlations are assumed to be identical. Recently proposed extensions allow the within- and between-period intra-cluster correlations to differ, although these methods require that all between-period intra-cluster correlations are identical, which may not be appropriate in all situations. Motivated by a proposed intensive care cluster randomised trial, we propose an alternative correlation structure for repeated cross-sectional multiple-period cluster randomised trials in which the between-period intra-cluster correlation is allowed to decay depending on the distance between measurements. We present results for the variance of treatment effect estimators for varying amounts of decay, investigating the consequences of the variation in decay on sample size planning for stepped wedge, cluster crossover and multiple-period parallel-arm cluster randomised trials. We also investigate the impact of assuming constant between-period intra-cluster correlations instead of decaying between-period intra-cluster correlations. Our results indicate that in certain design configurations, including the one corresponding to the proposed trial, a correlation decay can have an important impact on variances of treatment effect estimators, and hence on sample size and power. An R Shiny app allows readers to interactively explore the impact of correlation decay.

  20. A null model for microbial diversification

    PubMed Central

    Straub, Timothy J.

    2017-01-01

    Whether prokaryotes (Bacteria and Archaea) are naturally organized into phenotypically and genetically cohesive units comparable to animal or plant species remains contested, frustrating attempts to estimate how many such units there might be, or to identify the ecological roles they play. Analyses of gene sequences in various closely related prokaryotic groups reveal that sequence diversity is typically organized into distinct clusters, and processes such as periodic selection and extensive recombination are understood to be drivers of cluster formation (“speciation”). However, observed patterns are rarely compared with those obtainable with simple null models of diversification under stochastic lineage birth and death and random genetic drift. Via a combination of simulations and analyses of core and phylogenetic marker genes, we show that patterns of diversity for the genera Escherichia, Neisseria, and Borrelia are generally indistinguishable from patterns arising under a null model. We suggest that caution should thus be taken in interpreting observed clustering as a result of selective evolutionary forces. Unknown forces do, however, appear to play a role in Helicobacter pylori, and some individual genes in all groups fail to conform to the null model. Taken together, we recommend the presented birth−death model as a null hypothesis in prokaryotic speciation studies. It is only when the real data are statistically different from the expectations under the null model that some speciation process should be invoked. PMID:28630293

  1. Generic Features of Tertiary Chromatin Structure as Detected in Natural Chromosomes

    PubMed Central

    Müller, Waltraud G.; Rieder, Dietmar; Kreth, Gregor; Cremer, Christoph; Trajanoski, Zlatko; McNally, James G.

    2004-01-01

    Knowledge of tertiary chromatin structure in mammalian interphase chromosomes is largely derived from artificial tandem arrays. In these model systems, light microscope images reveal fibers or beaded fibers after high-density targeting of transactivators to insertional domains spanning several megabases. These images of fibers have lent support to chromonema fiber models of tertiary structure. To assess the relevance of these studies to natural mammalian chromatin, we identified two different ∼400-kb regions on human chromosomes 6 and 22 and then examined light microscope images of interphase tertiary chromatin structure when the regions were transcriptionally active and inactive. When transcriptionally active, these natural chromosomal regions elongated, yielding images characterized by a series of adjacent puncta or “beads”, referred to hereafter as beaded images. These elongated structures required transcription for their maintenance. Thus, despite marked differences in the density and the mode of transactivation, the natural and artificial systems showed similarities, suggesting that beaded images are generic features of transcriptionally active tertiary chromatin. We show here, however, that these images do not necessarily favor chromonema fiber models but can also be explained by a radial-loop model or even a simple nucleosome affinity, random-chain model. Thus, light microscope images of tertiary structure cannot distinguish among competing models, although they do impose key constraints: chromatin must be clustered to yield beaded images and then packaged within each cluster to enable decondensation into adjacent clusters. PMID:15485905

  2. Diabetes Care Management Teams Did Not Reduce Utilization When Compared With Traditional Care: A Randomized Cluster Trial.

    PubMed

    Kearns, Patrick

    2017-10-01

    PURPOSE: Health services research evaluates redesign models for primary care. Care management is one alternative. Evaluation includes resource utilization as a criterion. Compare the impact of care-manager teams on resource utilization. The comparison includes entire panes of patients and the subset of patients with diabetes. DESIGN: Randomized, prospective, cohort study comparing change in utilization rates between groups, pre- and post-intervention. METHODOLOGY: Ten primary care physician panels in a safety-net setting. Ten physicians were randomized to either a care-management approach (Group 1) or a traditional approach (Group 2). Care managers focused on diabetes and the cardiovascular cluster of diseases. Analysis compared rates of hospitalization, 30-day readmission, emergency room visits, and urgent care visits. Analysis compared baseline rates to annual rates after a yearlong run-in for entire panels and the subset of patients with diabetes. RESULTS: Resource utilization showed no statistically significant change between baseline and Year 3 (P=.79). Emergency room visits and hospital readmission increased for both groups (P=.90), while hospital admissions and urgent care visits decreased (P=.73). Similarly, utilization was not significantly different for patients with diabetes (P=.69). CONCLUSIONS: A care-management team approach failed to improve resource utilization rates by entire panels and the subset of diabetic patients compared to traditional care. This reinforces the need for further evidentiary support for the care-management model's hypothesis in the safety net.

  3. A stochastic fault model. 2. Time-dependent case.

    USGS Publications Warehouse

    Andrews, D.J.

    1981-01-01

    A random model of fault motion in an earthquake is formulated by assuming that the slip velocity is a random function of position and time truncated at zero, so that it does not have negative values. This random function is chosen to be self-affine; that is, on change of length scale, the function is multiplied by a scale factor but is otherwise unchanged statistically. A snapshot of slip velocity at a given time resembles a cluster of islands with rough topography; the final slip function is a smoother island or cluster of islands. In the Fourier transform domain, shear traction on the fault equals the slip velocity times an impedance function. The fact that this impedance function has a pole at zero frequency implies that traction and slip velocity cannot have the same spectral dependence in space and time. To describe stress fluctuations of the order of 100 bars when smoothed over a length of kilometers and of the order of kilobars at the grain size, shear traction must have a one-dimensional power spectrum is space proportional to the reciprocal wave number. Then the one-dimensional power spectrum for the slip velocity is proportional to the reciprocal wave number squared and for slip to its cube. If slip velocity has the same power law spectrum in time as in space, then the spectrum of ground acceleration with be flat (white noise) both on the fault and in the far field.-Author

  4. Input clustering in the normal and learned circuits of adult barn owls.

    PubMed

    McBride, Thomas J; DeBello, William M

    2015-05-01

    Experience-dependent formation of synaptic input clusters can occur in juvenile brains. Whether this also occurs in adults is largely unknown. We previously reconstructed the normal and learned circuits of prism-adapted barn owls and found that changes in clustering of axo-dendritic contacts (putative synapses) predicted functional circuit strength. Here we asked whether comparable changes occurred in normal and prism-removed adults. Across all anatomical zones, no systematic differences in the primary metrics for within-branch or between-branch clustering were observed: 95-99% of contacts resided within clusters (<10-20 μm from nearest neighbor) regardless of circuit strength. Bouton volumes, a proxy measure of synaptic strength, were on average larger in the functionally strong zones, indicating that changes in synaptic efficacy contributed to the differences in circuit strength. Bootstrap analysis showed that the distribution of inter-contact distances strongly deviated from random not in the functionally strong zones but in those that had been strong during the sensitive period (60-250 d), indicating that clusters formed early in life were preserved regardless of current value. While cluster formation in juveniles appeared to require the production of new synapses, cluster formation in adults did not. In total, these results support a model in which high cluster dynamics in juveniles sculpt a potential connectivity map that is refined in adulthood. We propose that preservation of clusters in functionally weak adult circuits provides a storage mechanism for disused but potentially useful pathways. Copyright © 2015 Elsevier Inc. All rights reserved.

  5. Comparing selected morphological models of hydrated Nafion using large scale molecular dynamics simulations

    NASA Astrophysics Data System (ADS)

    Knox, Craig K.

    Experimental elucidation of the nanoscale structure of hydrated Nafion, the most popular polymer electrolyte or proton exchange membrane (PEM) to date, and its influence on macroscopic proton conductance is particularly challenging. While it is generally agreed that hydrated Nafion is organized into distinct hydrophilic domains or clusters within a hydrophobic matrix, the geometry and length scale of these domains continues to be debated. For example, at least half a dozen different domain shapes, ranging from spheres to cylinders, have been proposed based on experimental SAXS and SANS studies. Since the characteristic length scale of these domains is believed to be ˜2 to 5 nm, very large molecular dynamics (MD) simulations are needed to accurately probe the structure and morphology of these domains, especially their connectivity and percolation phenomena at varying water content. Using classical, all-atom MD with explicit hydronium ions, simulations have been performed to study the first-ever hydrated Nafion systems that are large enough (~2 million atoms in a ˜30 nm cell) to directly observe several hydrophilic domains at the molecular level. These systems consisted of six of the most significant and relevant morphological models of Nafion to-date: (1) the cluster-channel model of Gierke, (2) the parallel cylinder model of Schmidt-Rohr, (3) the local-order model of Dreyfus, (4) the lamellar model of Litt, (5) the rod network model of Kreuer, and (6) a 'random' model, commonly used in previous simulations, that does not directly assume any particular geometry, distribution, or morphology. These simulations revealed fast intercluster bridge formation and network percolation in all of the models. Sulfonates were found inside these bridges and played a significant role in percolation. Sulfonates also strongly aggregated around and inside clusters. Cluster surfaces were analyzed to study the hydrophilic-hydrophobic interface. Interfacial area and cluster volume significantly increased during the simulations, suggesting the need for morphological model refinement and improvement. Radial distribution functions and structure factors were calculated. All nonrandom models exhibited the characteristic experimental scattering peak, underscoring the insensitivity of this measurement to hydrophilic domain structure and highlighting the need for future work to clearly distinguish morphological models of Nafion.

  6. Dewetting and spreading transitions for active matter on random pinning substrates.

    PubMed

    Sándor, Cs; Libál, A; Reichhardt, C; Olson Reichhardt, C J

    2017-05-28

    We show that sterically interacting self-propelled disks in the presence of random pinning substrates exhibit transitions among a variety of different states. In particular, from a phase separated cluster state, the disks can spread out and homogeneously cover the substrate in what can be viewed as an example of an active matter wetting transition. We map the location of this transition as a function of activity, disk density, and substrate strength, and we also identify other phases including a cluster state, coexistence between a cluster and a labyrinth wetted phase, and a pinned liquid. Convenient measures of these phases include the cluster size, which dips at the wetting-dewetting transition, and the fraction of sixfold coordinated particles, which drops when dewetting occurs.

  7. The asymptotic behavior in a reversible random coagulation-fragmentation polymerization process with sub-exponential decay

    NASA Astrophysics Data System (ADS)

    Dong, Siqun; Zhao, Dianli

    2018-01-01

    This paper studies the subcritical, near-critical and supercritical asymptotic behavior of a reversible random coagulation-fragmentation polymerization process as N → ∞, with the number of distinct ways to form a k-clusters from k units satisfying f(k) =(1 + o (1)) cr-ke-kαk-β, where 0 < α < 1 and β > 0. When the cluster size is small, its distribution is proved to converge to the Gaussian distribution. For the medium clusters, its distribution will converge to Poisson distribution in supercritical stage, and no large clusters exist in this stage. Furthermore, the largest length of polymers of size N is of order ln N in the subcritical stage under α ⩽ 1 / 2.

  8. Activity Begins in Childhood (ABC) - inspiring healthy active behaviour in preschoolers: study protocol for a cluster randomized controlled trial.

    PubMed

    Adamo, Kristi B; Barrowman, Nick; Naylor, Patti Jean; Yaya, Sanni; Harvey, Alysha; Grattan, Kimberly P; Goldfield, Gary S

    2014-07-29

    Today's children are more overweight than previous generations and physical inactivity is a contributing factor. Modelling and promoting positive behaviour in the early years is imperative for the development of lifelong health habits. The social and physical environments where children spend their time have a powerful influence on behaviour. Since the majority of preschool children spend time in care outside of the home, this provides an ideal setting to examine the ability of an intervention to enhance movement skills and modify physical activity behaviour. This study aims to evaluate the efficacy of the Activity Begins in Childhood (ABC) intervention delivered in licensed daycare settings alone or in combination with a parent-driven home physical activity-promotion component to increase preschoolers' overall physical activity levels and, specifically, the time spent in moderate to vigorous physical activity. This study is a single site, three-arm, cluster-randomized controlled trial design with a daycare centre as the unit of measurement (clusters). All daycare centres in the National Capital region that serve children between the ages of 3 and 5, expressing an interest in receiving the ABC intervention will be invited to participate. Those who agree will be randomly assigned to one of three groups: i) ABC program delivered at a daycare centre only, ii) ABC program delivered at daycare with a home/parental education component, or iii) regular daycare curriculum. This study will recruit 18 daycare centres, 6 in each of the three groups. The intervention will last approximately 6 months, with baseline assessment prior to ABC implementation and follow-up assessments at 3 and 6 months. Physical activity is an acknowledged component of a healthy lifestyle and childhood experiences as it has an important impact on lifelong behaviour and health. Opportunities for physical activity and motor development in early childhood may, over the lifespan, influence the maintenance of a healthy body weight and reduce cardiovascular disease risk. If successful, the ABC program may be implemented in daycare centres as an effective way of increasing healthy activity behaviours of preschoolers. Current Controlled Trials: ISRCTN94022291. Registered in December 2012, first cluster randomized in April 2013.

  9. Implementation of a quantum random number generator based on the optimal clustering of photocounts

    NASA Astrophysics Data System (ADS)

    Balygin, K. A.; Zaitsev, V. I.; Klimov, A. N.; Kulik, S. P.; Molotkov, S. N.

    2017-10-01

    To implement quantum random number generators, it is fundamentally important to have a mathematically provable and experimentally testable process of measurements of a system from which an initial random sequence is generated. This makes sure that randomness indeed has a quantum nature. A quantum random number generator has been implemented with the use of the detection of quasi-single-photon radiation by a silicon photomultiplier (SiPM) matrix, which makes it possible to reliably reach the Poisson statistics of photocounts. The choice and use of the optimal clustering of photocounts for the initial sequence of photodetection events and a method of extraction of a random sequence of 0's and 1's, which is polynomial in the length of the sequence, have made it possible to reach a yield rate of 64 Mbit/s of the output certainly random sequence.

  10. Effect of health promotion and fluoride varnish on dental caries among Australian Aboriginal children: results from a community-randomized controlled trial*

    PubMed Central

    Slade, Gary D; Bailie, Ross S; Roberts-Thomson, Kaye; Leach, Amanda J; Raye, Iris; Endean, Colin; Simmons, Bruce; Morris, Peter

    2011-01-01

    Objectives We tested a dental health program in remote Aboriginal communities of Australia's Northern Territory, hypothesizing that it would reduce dental caries in preschool children. Methods In this 2-year, prospective, cluster-randomized, concurrent controlled, open trial of the dental health program compared to no such program, 30 communities were allocated at random to intervention and control groups. All residents aged 18–47 months were invited to participate. Twice per year for 2 years in the 15 intervention communities, fluoride varnish was applied to children's teeth, water consumption and daily tooth cleaning with toothpaste were advocated, dental health was promoted in community settings, and primary health care workers were trained in preventive dental care. Data from dental examinations at baseline and after 2 years were used to compute net dental caries increment per child (d3mfs). A multi-level statistical model compared d3mfs between intervention and control groups with adjustment for the clustered randomization design; four other models used additional variables for adjustment. Results At baseline, 666 children were examined; 543 of them (82%) were re-examined 2 years later. The adjusted d3mfs increment was significantly lower in the intervention group compared to the control group by an average of 3.0 surfaces per child (95% CI = 1.2, 4.9), a prevented fraction of 31%. Adjustment for additional variables yielded caries reductions ranging from 2.3 to 3.5 surfaces per child and prevented fractions of 24–36%. Conclusions These results corroborate findings from other studies where fluoride varnish was efficacious in preventing dental caries in young children. PMID:20707872

  11. Evaluating the Implementation of a School-Based Emotional Well-Being Programme: A Cluster Randomized Controlled Trial of Zippy's Friends for Children in Disadvantaged Primary Schools

    ERIC Educational Resources Information Center

    Clarke, Aleisha M.; Bunting, Brendan; Barry, Margaret M.

    2014-01-01

    Schools are recognized as one of the most important settings for promoting social and emotional well-being among children and adolescents. This clustered randomized controlled trial evaluated Zippy's Friends, an international school-based emotional well-being programme, with 766 children from designated disadvantaged schools. The purpose of this…

  12. Citywide cluster randomized trial to restore blighted vacant land and its effects on violence, crime, and fear

    Treesearch

    Charles C. Branas; Eugenia South; Michelle C. Kondo; Bernadette C. Hohl; Philippe Bourgois; Douglas J. Wiebe; John M. MacDonald

    2018-01-01

    Vacant and blighted urban land is a widespread and potentially risky environmental condition encountered by millions of people on a daily basis. About 15% of the land in US cities is deemed vacant or abandoned, an area roughly the size of Switzerland. In a citywide cluster randomized controlled trial, we investigated the effects of standardized, reproducible...

  13. The Long-Term Effectiveness of a Selective, Personality-Targeted Prevention Program in Reducing Alcohol Use and Related Harms: A Cluster Randomized Controlled Trial

    ERIC Educational Resources Information Center

    Newton, Nicola C.; Conrod, Patricia J.; Slade, Tim; Carragher, Natacha; Champion, Katrina E.; Barrett, Emma L.; Kelly, Erin V.; Nair, Natasha K.; Stapinski, Lexine; Teesson, Maree

    2016-01-01

    Background: This study investigated the long-term effectiveness of Preventure, a selective personality-targeted prevention program, in reducing the uptake of alcohol, harmful use of alcohol, and alcohol-related harms over a 3-year period. Methods: A cluster randomized controlled trial was conducted to assess the effectiveness of Preventure.…

  14. Reducing Tobacco Use among Low Socio-Economic Status Youth in Delhi, India: Outcomes from Project ACTIVITY, a Cluster Randomized Trial

    ERIC Educational Resources Information Center

    Harrell, Melissa B.; Arora, Monika; Bassi, Shalini; Gupta, Vinay K.; Perry, Cheryl L.; Reddy, K. Srinath

    2016-01-01

    To test the efficacy of an intervention to reduce tobacco use among youth (10-19 years old) in slum communities in Delhi, India. This community-based cluster-randomized trial included 14 slums composed of purposely built resettlement colonies and adjacent inhabitant-built Jhuggi Jhopris. Youth in the intervention received a 2 year…

  15. Cluster-Randomized Controlled Trial Evaluating the Effectiveness of Computer-Assisted Intervention Delivered by Educators for Children with Speech Sound Disorders

    ERIC Educational Resources Information Center

    McLeod, Sharynne; Baker, Elise; McCormack, Jane; Wren, Yvonne; Roulstone, Sue; Crowe, Kathryn; Masso, Sarah; White, Paul; Howland, Charlotte

    2017-01-01

    Purpose: The aim was to evaluate the effectiveness of computer-assisted input-based intervention for children with speech sound disorders (SSD). Method: The Sound Start Study was a cluster-randomized controlled trial. Seventy-nine early childhood centers were invited to participate, 45 were recruited, and 1,205 parents and educators of 4- and…

  16. Improving Elementary School Quality through the Use of a Social-Emotional and Character Development Program: A Matched-Pair, Cluster-Randomized, Controlled Trial in Hawai'i

    ERIC Educational Resources Information Center

    Snyder, Frank J.; Vuchinich, Samuel; Acock, Alan; Washburn, Isaac J.; Flay, Brian R.

    2012-01-01

    Background: School safety and quality affect student learning and success. This study examined the effects of a comprehensive elementary school-wide social-emotional and character education program, Positive Action, on teacher, parent, and student perceptions of school safety and quality utilizing a matched-pair, cluster-randomized, controlled…

  17. Inadequacy of ethical conduct and reporting of stepped wedge cluster randomized trials: Results from a systematic review.

    PubMed

    Taljaard, Monica; Hemming, Karla; Shah, Lena; Giraudeau, Bruno; Grimshaw, Jeremy M; Weijer, Charles

    2017-08-01

    Background/aims The use of the stepped wedge cluster randomized design is rapidly increasing. This design is commonly used to evaluate health policy and service delivery interventions. Stepped wedge cluster randomized trials have unique characteristics that complicate their ethical interpretation. The 2012 Ottawa Statement provides comprehensive guidance on the ethical design and conduct of cluster randomized trials, and the 2010 CONSORT extension for cluster randomized trials provides guidelines for reporting. Our aims were to assess the adequacy of the ethical conduct and reporting of stepped wedge trials to date, focusing on research ethics review and informed consent. Methods We conducted a systematic review of stepped wedge cluster randomized trials in health research published up to 2014 in English language journals. We extracted details of study intervention and data collection procedures, as well as reporting of research ethics review and informed consent. Two reviewers independently extracted data from each trial; discrepancies were resolved through discussion. We identified the presence of any research participants at the cluster level and the individual level. We assessed ethical conduct by tabulating reporting of research ethics review and informed consent against the presence of research participants. Results Of 32 identified stepped wedge trials, only 24 (75%) reported review by a research ethics committee, and only 16 (50%) reported informed consent from any research participants-yet, all trials included research participants at some level. In the subgroup of 20 trials with research participants at cluster level, only 4 (20%) reported informed consent from such participants; in 26 trials with individual-level research participants, only 15 (58%) reported their informed consent. Interventions (regardless of whether targeting cluster- or individual-level participants) were delivered at the group level in more than two-thirds of trials; nine trials (28%) had no identifiable data collected from any research participants. Overall, only three trials (9%) indicated that a waiver of consent had been granted by a research ethics committee. When considering the combined requirement of research ethics review and informed consent (or a waiver), only one in three studies were compliant. Conclusion The ethical conduct and reporting of key ethical protections in stepped wedge trials, namely, research ethics review and informed consent, are inadequate. We recommend that stepped wedge trials be classified as research and reviewed and approved by a research ethics committee. We also recommend that researchers appropriately identify research participants (which may include health professionals), seek informed consent or appeal to an ethics committee for a waiver of consent, and include explicit details of research ethics approval and informed consent in the trial report.

  18. Scalability of an Evidence-Based Adolescent Pregnancy Prevention Program: New Evidence From 5 Cluster-Randomized Evaluations of the Teen Outreach Program.

    PubMed

    Francis, Kimberly; Philliber, Susan; Walsh-Buhi, Eric R; Philliber, Ashley; Seshadri, Roopa; Daley, Ellen

    2016-09-01

    To determine if the Teen Outreach Program (TOP), a youth development and service learning program, can reduce sexual risk-taking behaviors compared with a business as usual or benign counterfactual. We synthesized results of 5 independent studies conducted in 5 geographically and ethnically diverse locations between 2011 and 2015 with 17 194 middle and high school students. Each study cluster-randomized classes, teachers, or schools to treatment or control groups and included the students enrolled in those clusters at baseline in an intent-to-treat analysis. Multilevel models tested impacts on recent sexual activity, recent unprotected sexual activity, and sexual initiation among the sexually inexperienced at baseline at approximately 1 and 2 years after baseline. Precision-weighted average effect sizes showed nonsignificant reductions of 1 percentage point or less in recent sexual activity (5 studies: -0.6; P = .32), recent unprotected sex (5 studies: -0.2; P = .76), and sexual initiation (4 studies: -1.1; P = .10) after 1 year. There was little evidence of the effectiveness of TOP in reducing sexual risk-taking behaviors. Results underscored the importance of continually evaluating evidence-based programs that have previously been shown to be effective.

  19. The effect of a physical activity intervention on preschoolers' fundamental motor skills - A cluster RCT.

    PubMed

    Wasenius, Niko S; Grattan, Kimberly P; Harvey, Alysha L J; Naylor, Patti-Jean; Goldfield, Gary S; Adamo, Kristi B

    2018-07-01

    To assess the effect of a physical activity intervention delivered in the childcare centres (CC), with or without a parent-driven home physical activity component, on children's fundamental motor skills (FMS). Six-month 3-arm cluster randomized controlled trial. Preschoolers were recruited from 18 licensed CC. CC were randomly assigned to a typical curriculum comparison group (COM), childcare intervention alone (CC), or childcare intervention with parental component (CC+HOME). FMS was measured with the Test of Gross Motor Development-2. Linear mixed models were performed at the level of the individual while accounting for clustering. Raw locomotor skills score increased significantly in the CC group (mean difference=2.5 units, 95% Confidence Intervals, CI, 1.0-4.1, p<0.001) and the CC+HOME group (mean difference=2.4 units, 95% CI, 0.8-4.0, p<0.001) compared to the COM group. No significant (p>0.05) between group differences were observed in the raw object control skills, sum of raw scores, or gross motor quotient. No significant sex differences were found in any of the measured outcomes. A physical activity intervention delivered in childcare with or without parents' involvement was effective in increasing locomotor skills in preschoolers. Copyright © 2017 Sports Medicine Australia. Published by Elsevier Ltd. All rights reserved.

  20. Patterns in the English language: phonological networks, percolation and assembly models

    NASA Astrophysics Data System (ADS)

    Stella, Massimo; Brede, Markus

    2015-05-01

    In this paper we provide a quantitative framework for the study of phonological networks (PNs) for the English language by carrying out principled comparisons to null models, either based on site percolation, randomization techniques, or network growth models. In contrast to previous work, we mainly focus on null models that reproduce lower order characteristics of the empirical data. We find that artificial networks matching connectivity properties of the English PN are exceedingly rare: this leads to the hypothesis that the word repertoire might have been assembled over time by preferentially introducing new words which are small modifications of old words. Our null models are able to explain the ‘power-law-like’ part of the degree distributions and generally retrieve qualitative features of the PN such as high clustering, high assortativity coefficient and small-world characteristics. However, the detailed comparison to expectations from null models also points out significant differences, suggesting the presence of additional constraints in word assembly. Key constraints we identify are the avoidance of large degrees, the avoidance of triadic closure and the avoidance of large non-percolating clusters.

  1. A critical analysis of high-redshift, massive, galaxy clusters. Part I

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hoyle, Ben; Jimenez, Raul; Verde, Licia

    2012-02-01

    We critically investigate current statistical tests applied to high redshift clusters of galaxies in order to test the standard cosmological model and describe their range of validity. We carefully compare a sample of high-redshift, massive, galaxy clusters with realistic Poisson sample simulations of the theoretical mass function, which include the effect of Eddington bias. We compare the observations and simulations using the following statistical tests: the distributions of ensemble and individual existence probabilities (in the > M, > z sense), the redshift distributions, and the 2d Kolmogorov-Smirnov test. Using seemingly rare clusters from Hoyle et al. (2011), and Jee etmore » al. (2011) and assuming the same survey geometry as in Jee et al. (2011, which is less conservative than Hoyle et al. 2011), we find that the ( > M, > z) existence probabilities of all clusters are fully consistent with ΛCDM. However assuming the same survey geometry, we use the 2d K-S test probability to show that the observed clusters are not consistent with being the least probable clusters from simulations at > 95% confidence, and are also not consistent with being a random selection of clusters, which may be caused by the non-trivial selection function and survey geometry. Tension can be removed if we examine only a X-ray selected sub sample, with simulations performed assuming a modified survey geometry.« less

  2. A Hybrid Spectral Clustering and Deep Neural Network Ensemble Algorithm for Intrusion Detection in Sensor Networks

    PubMed Central

    Ma, Tao; Wang, Fen; Cheng, Jianjun; Yu, Yang; Chen, Xiaoyun

    2016-01-01

    The development of intrusion detection systems (IDS) that are adapted to allow routers and network defence systems to detect malicious network traffic disguised as network protocols or normal access is a critical challenge. This paper proposes a novel approach called SCDNN, which combines spectral clustering (SC) and deep neural network (DNN) algorithms. First, the dataset is divided into k subsets based on sample similarity using cluster centres, as in SC. Next, the distance between data points in a testing set and the training set is measured based on similarity features and is fed into the deep neural network algorithm for intrusion detection. Six KDD-Cup99 and NSL-KDD datasets and a sensor network dataset were employed to test the performance of the model. These experimental results indicate that the SCDNN classifier not only performs better than backpropagation neural network (BPNN), support vector machine (SVM), random forest (RF) and Bayes tree models in detection accuracy and the types of abnormal attacks found. It also provides an effective tool of study and analysis of intrusion detection in large networks. PMID:27754380

  3. A Hybrid Spectral Clustering and Deep Neural Network Ensemble Algorithm for Intrusion Detection in Sensor Networks.

    PubMed

    Ma, Tao; Wang, Fen; Cheng, Jianjun; Yu, Yang; Chen, Xiaoyun

    2016-10-13

    The development of intrusion detection systems (IDS) that are adapted to allow routers and network defence systems to detect malicious network traffic disguised as network protocols or normal access is a critical challenge. This paper proposes a novel approach called SCDNN, which combines spectral clustering (SC) and deep neural network (DNN) algorithms. First, the dataset is divided into k subsets based on sample similarity using cluster centres, as in SC. Next, the distance between data points in a testing set and the training set is measured based on similarity features and is fed into the deep neural network algorithm for intrusion detection. Six KDD-Cup99 and NSL-KDD datasets and a sensor network dataset were employed to test the performance of the model. These experimental results indicate that the SCDNN classifier not only performs better than backpropagation neural network (BPNN), support vector machine (SVM), random forest (RF) and Bayes tree models in detection accuracy and the types of abnormal attacks found. It also provides an effective tool of study and analysis of intrusion detection in large networks.

  4. Effectiveness of a cognitive behavioural workbook for changing beliefs about antipsychotic polypharmacy: analysis from a cluster randomized controlled trial.

    PubMed

    Thompson, Andrew; Sullivan, Sarah; Barley, Maddi; Moore, Laurence; Rogers, Paul; Sipos, Attila; Harrison, Glynn

    2010-06-01

    Educational workbooks have been used in psychiatry to influence patient but not clinician behaviour. Targeted education interventions to change prescribing practice in other areas of medicine have only looked at changes in prescribing and not attitudes or beliefs related to the prescribing. We aimed to examine whether clinicians' beliefs about a common prescribing issue in psychiatry (antipsychotic polypharmacy prescription) changed alongside behaviour as a result of a complex intervention. Medical and nursing staff were recruited from 19 general adult psychiatry units in the south-west of the UK as part of a cluster randomized controlled trial. A questionnaire was used to assess beliefs on the prescribing of antipsychotic polypharmacy as a secondary outcome before and after completion of a cognitive behavioural 'self-help' style workbook (one part of a complex intervention). A factor analysis suggested three dimensions of the questionnaire that corresponded to predetermined themes. The data were analysed using a random-effects regression model (adjusting for clustering) controlling for possible confounders. There was a significant change in beliefs on both of the factors: antipsychotic polypharmacy (coefficient = -0.89, P < 0.01) and rapid tranquilization (coefficient = -0.68, P = 0.01) specifically targeted by the workbook. There was a modest but statistically significant change in antipsychotic polypharmacy prescribing (odds ratio 0.43, 95% confidence intervals 0.21-0.90). The workbook appeared to change staff beliefs about antipsychotic polypharmacy, but achieving substantial changes in clinician behaviour may require further exploration of other factors important in complex prescribing issues.

  5. Spatial point pattern analysis of human settlements and geographical associations in eastern coastal China - a case study.

    PubMed

    Zhang, Zhonghao; Xiao, Rui; Shortridge, Ashton; Wu, Jiaping

    2014-03-10

    Understanding the spatial point pattern of human settlements and their geographical associations are important for understanding the drivers of land use and land cover change and the relationship between environmental and ecological processes on one hand and cultures and lifestyles on the other. In this study, a Geographic Information System (GIS) approach, Ripley's K function and Monte Carlo simulation were used to investigate human settlement point patterns. Remotely sensed tools and regression models were employed to identify the effects of geographical determinants on settlement locations in the Wen-Tai region of eastern coastal China. Results indicated that human settlements displayed regular-random-cluster patterns from small to big scale. Most settlements located on the coastal plain presented either regular or random patterns, while those in hilly areas exhibited a clustered pattern. Moreover, clustered settlements were preferentially located at higher elevations with steeper slopes and south facing aspects than random or regular settlements. Regression showed that influences of topographic factors (elevation, slope and aspect) on settlement locations were stronger across hilly regions. This study demonstrated a new approach to analyzing the spatial patterns of human settlements from a wide geographical prospective. We argue that the spatial point patterns of settlements, in addition to the characteristics of human settlements, such as area, density and shape, should be taken into consideration in the future, and land planners and decision makers should pay more attention to city planning and management. Conceptual and methodological bridges linking settlement patterns to regional and site-specific geographical characteristics will be a key to human settlement studies and planning.

  6. Intravenous levetiracetam vs phenytoin for status epilepticus and cluster seizures: A prospective, randomized study.

    PubMed

    Gujjar, Arunodaya R; Nandhagopal, Ramachandiran; Jacob, Poovathoor C; Al-Hashim, Abdulhakeem; Al-Amrani, Khalfan; Ganguly, Shyam S; Al-Asmi, Abdullah

    2017-07-01

    Status Epilepticus (SE) is a common medical emergency carrying a high morbidity and mortality. Levetiracetam (LEV) is a novel anticonvulsant effective against varied seizures. Few prospective studies have addressed its use in SE. We aimed to examine the efficacy of intravenous LEV in controlling SE and cluster attacks of seizures (CS), in comparison with IV phenytoin (DPH), using a prospective, randomized study design. Adult patients with SE or CS, following an initial dose of IV benzodiazepine to control ongoing seizure, were randomized to receive either medication. Rates of seizure control over 24h, adverse effects and outcomes were compared. A logistic regression model was used to identify outcome predictors. 52 patients with SE and 63 with CS received either LEV or DPH. In the SE group, LEV was effective in18/22(82%) and DPH in 22/30(73.3%) patients in controlling seizures. Among patients with CS, LEV was effective in 31/38(81.6%) and DPH in 20/25(80%). With the use of LEV, DPH or both, SE and CS were controlled among 92% and 96% of patients respectively. Adverse events included hypotension (in 2 on DPH) and transient agitation (2 on LEV). IV Levetiracetam controls status epilepticus or cluster seizures with an efficacy comparable to that of phenytoin. Use of these two agents consecutively may control >90% of all such conditions without resort to anaesthetic agents. Further studies should explore its efficacy in larger cohorts of epileptic emergencies. Copyright © 2017 British Epilepsy Association. Published by Elsevier Ltd. All rights reserved.

  7. Cluster designs to assess the prevalence of acute malnutrition by lot quality assurance sampling: a validation study by computer simulation.

    PubMed

    Olives, Casey; Pagano, Marcello; Deitchler, Megan; Hedt, Bethany L; Egge, Kari; Valadez, Joseph J

    2009-04-01

    Traditional lot quality assurance sampling (LQAS) methods require simple random sampling to guarantee valid results. However, cluster sampling has been proposed to reduce the number of random starting points. This study uses simulations to examine the classification error of two such designs, a 67x3 (67 clusters of three observations) and a 33x6 (33 clusters of six observations) sampling scheme to assess the prevalence of global acute malnutrition (GAM). Further, we explore the use of a 67x3 sequential sampling scheme for LQAS classification of GAM prevalence. Results indicate that, for independent clusters with moderate intracluster correlation for the GAM outcome, the three sampling designs maintain approximate validity for LQAS analysis. Sequential sampling can substantially reduce the average sample size that is required for data collection. The presence of intercluster correlation can impact dramatically the classification error that is associated with LQAS analysis.

  8. Random Walk Quantum Clustering Algorithm Based on Space

    NASA Astrophysics Data System (ADS)

    Xiao, Shufen; Dong, Yumin; Ma, Hongyang

    2018-01-01

    In the random quantum walk, which is a quantum simulation of the classical walk, data points interacted when selecting the appropriate walk strategy by taking advantage of quantum-entanglement features; thus, the results obtained when the quantum walk is used are different from those when the classical walk is adopted. A new quantum walk clustering algorithm based on space is proposed by applying the quantum walk to clustering analysis. In this algorithm, data points are viewed as walking participants, and similar data points are clustered using the walk function in the pay-off matrix according to a certain rule. The walk process is simplified by implementing a space-combining rule. The proposed algorithm is validated by a simulation test and is proved superior to existing clustering algorithms, namely, Kmeans, PCA + Kmeans, and LDA-Km. The effects of some of the parameters in the proposed algorithm on its performance are also analyzed and discussed. Specific suggestions are provided.

  9. An order statistics approach to the halo model for galaxies

    NASA Astrophysics Data System (ADS)

    Paul, Niladri; Paranjape, Aseem; Sheth, Ravi K.

    2017-04-01

    We use the halo model to explore the implications of assuming that galaxy luminosities in groups are randomly drawn from an underlying luminosity function. We show that even the simplest of such order statistics models - one in which this luminosity function p(L) is universal - naturally produces a number of features associated with previous analyses based on the 'central plus Poisson satellites' hypothesis. These include the monotonic relation of mean central luminosity with halo mass, the lognormal distribution around this mean and the tight relation between the central and satellite mass scales. In stark contrast to observations of galaxy clustering; however, this model predicts no luminosity dependence of large-scale clustering. We then show that an extended version of this model, based on the order statistics of a halo mass dependent luminosity function p(L|m), is in much better agreement with the clustering data as well as satellite luminosities, but systematically underpredicts central luminosities. This brings into focus the idea that central galaxies constitute a distinct population that is affected by different physical processes than are the satellites. We model this physical difference as a statistical brightening of the central luminosities, over and above the order statistics prediction. The magnitude gap between the brightest and second brightest group galaxy is predicted as a by-product, and is also in good agreement with observations. We propose that this order statistics framework provides a useful language in which to compare the halo model for galaxies with more physically motivated galaxy formation models.

  10. Implementation of patient education at first and second dispensing of statins in Dutch community pharmacies: the sequel of a cluster randomized trial

    PubMed Central

    2011-01-01

    Background As a result of the previous part of this trial, many patients with cardiovascular disease were expected to receive a statin for the first time. In order to provide these patients with comprehensive information on statins, as recommended by professional guidance, education at first and second dispensing of statins had to be implemented. This study was designed to assess the effectiveness of an intensive implementation program targeted at pharmacy project assistants on the frequency of providing education at first dispensing (EAFD) and education at second dispensing (EASD) of statins in community pharmacies. Methods The participating community pharmacies were clustered on the basis of local collaboration, were numbered by a research assistant and subsequently an independent statistician performed a block randomization, in which the cluster size (number of pharmacies in each cluster) was balanced. The pharmacies in the control group received a written manual on the implementation of EAFD and EASD; the pharmacies in the intervention group received intensive support for the implementation. The impact of the intensive implementation program on the implementation process and on the primary outcomes was examined in a random coefficient logistic regression model, which took into account that patients were grouped within pharmacy clusters. Results Of the 37 pharmacies in the intervention group, 17 pharmacies (50%) provided EAFD and 12 pharmacies (35.3%) provided EASD compared to 14 pharmacies (45.2%, P = 0.715) and 12 pharmacies (38.7%, P = 0.899), respectively, of the 34 pharmacies in the control group. In the intervention group a total of 72 of 469 new statin users (15.4%) received education and 49 of 393 patients with a second statin prescription (12.5%) compared to 78 of 402 new users (19.4%, P = 0.944) and 35 of 342 patients with a second prescription (10.2%, P = 0.579) in the control group. Conclusion The intensive implementation program did not increase the frequency of providing EAFD and EASD of statins in community pharmacies. Trial Registration clinicaltrials.gov NCT00509717 PMID:22087850

  11. Model of large volumetric capacitance in graphene supercapacitors based on ion clustering

    NASA Astrophysics Data System (ADS)

    Skinner, Brian; Fogler, M. M.; Shklovskii, B. I.

    2011-12-01

    Electric double-layer supercapacitors (SCs) are promising devices for high-power energy storage based on the reversible absorption of ions into porous conducting electrodes. Graphene is a particularly good candidate for the electrode material in SCs due to its high conductivity and large surface area. In this paper, we consider SC electrodes made from a stack of graphene sheets with randomly inserted spacer molecules. We show that the large volumetric capacitances C≳100F/cm3 observed experimentally can be understood as a result of collective intercalation of ions into the graphene stack and the accompanying nonlinear screening by graphene electrons that renormalizes the charge of the ion clusters.

  12. A model of large volumetric capacitance in graphene supercapacitors based on ion clustering

    NASA Astrophysics Data System (ADS)

    Skinner, Brian; Fogler, Michael; Shklovskii, Boris

    2012-02-01

    Electric double layer supercapacitors are promising devices for high-power energy storage based on the reversible absorption of ions into porous, conducting electrodes. Graphene is a particularly good candidate for the electrode material in supercapacitors due to its high conductivity and large surface area. In this paper we consider supercapacitor electrodes made from a stack of graphene sheets with randomly-inserted ``spacer" molecules. We show that the large volumetric capacitances C > 100 F/cm^3 observed experimentally can be understood as a result of collective intercalation of ions into the graphene stack and the accompanying nonlinear screening by graphene electrons that renormalizes the charge of the ion clusters.

  13. The PULSAR Specialist Care protocol: a stepped-wedge cluster randomized control trial of a training intervention for community mental health teams in recovery-oriented practice.

    PubMed

    Shawyer, Frances; Enticott, Joanne C; Brophy, Lisa; Bruxner, Annie; Fossey, Ellie; Inder, Brett; Julian, John; Kakuma, Ritsuko; Weller, Penelope; Wilson-Evered, Elisabeth; Edan, Vrinda; Slade, Mike; Meadows, Graham N

    2017-05-08

    Recovery features strongly in Australian mental health policy; however, evidence is limited for the efficacy of recovery-oriented practice at the service level. This paper describes the Principles Unite Local Services Assisting Recovery (PULSAR) Specialist Care trial protocol for a recovery-oriented practice training intervention delivered to specialist mental health services staff. The primary aim is to evaluate whether adult consumers accessing services where staff have received the intervention report superior recovery outcomes compared to adult consumers accessing services where staff have not yet received the intervention. A qualitative sub-study aims to examine staff and consumer views on implementing recovery-oriented practice. A process evaluation sub-study aims to articulate important explanatory variables affecting the interventions rollout and outcomes. The mixed methods design incorporates a two-step stepped-wedge cluster randomized controlled trial (cRCT) examining cross-sectional data from three phases, and nested qualitative and process evaluation sub-studies. Participating specialist mental health care services in Melbourne, Victoria are divided into 14 clusters with half randomly allocated to receive the staff training in year one and half in year two. Research participants are consumers aged 18-75 years who attended the cluster within a previous three-month period either at baseline, 12 (step 1) or 24 months (step 2). In the two nested sub-studies, participation extends to cluster staff. The primary outcome is the Questionnaire about the Process of Recovery collected from 756 consumers (252 each at baseline, step 1, step 2). Secondary and other outcomes measuring well-being, service satisfaction and health economic impact are collected from a subset of 252 consumers (63 at baseline; 126 at step 1; 63 at step 2) via interviews. Interview-based longitudinal data are also collected 12 months apart from 88 consumers with a psychotic disorder diagnosis (44 at baseline, step 1; 44 at step 1, step 2). cRCT data will be analyzed using multilevel mixed-effects modelling to account for clustering and some repeated measures, supplemented by thematic analysis of qualitative interview data. The process evaluation will draw on qualitative, quantitative and documentary data. Findings will provide an evidence-base for the continued transformation of Australian mental health service frameworks toward recovery. Australian and New Zealand Clinical Trial Registry: ACTRN12614000957695 . Date registered: 8 September 2014.

  14. Coordinate based random effect size meta-analysis of neuroimaging studies.

    PubMed

    Tench, C R; Tanasescu, Radu; Constantinescu, C S; Auer, D P; Cottam, W J

    2017-06-01

    Low power in neuroimaging studies can make them difficult to interpret, and Coordinate based meta-analysis (CBMA) may go some way to mitigating this issue. CBMA has been used in many analyses to detect where published functional MRI or voxel-based morphometry studies testing similar hypotheses report significant summary results (coordinates) consistently. Only the reported coordinates and possibly t statistics are analysed, and statistical significance of clusters is determined by coordinate density. Here a method of performing coordinate based random effect size meta-analysis and meta-regression is introduced. The algorithm (ClusterZ) analyses both coordinates and reported t statistic or Z score, standardised by the number of subjects. Statistical significance is determined not by coordinate density, but by a random effects meta-analyses of reported effects performed cluster-wise using standard statistical methods and taking account of censoring inherent in the published summary results. Type 1 error control is achieved using the false cluster discovery rate (FCDR), which is based on the false discovery rate. This controls both the family wise error rate under the null hypothesis that coordinates are randomly drawn from a standard stereotaxic space, and the proportion of significant clusters that are expected under the null. Such control is necessary to avoid propagating and even amplifying the very issues motivating the meta-analysis in the first place. ClusterZ is demonstrated on both numerically simulated data and on real data from reports of grey matter loss in multiple sclerosis (MS) and syndromes suggestive of MS, and of painful stimulus in healthy controls. The software implementation is available to download and use freely. Copyright © 2017 Elsevier Inc. All rights reserved.

  15. A survey of scientific literacy to provide a foundation for designing science communication in Japan.

    PubMed

    Kawamoto, Shishin; Nakayama, Minoru; Saijo, Miki

    2013-08-01

    There are various definitions and survey methods for scientific literacy. Taking into consideration the contemporary significance of scientific literacy, we have defined it with an emphasis on its social aspects. To acquire the insights needed to design a form of science communication that will enhance the scientific literacy of each individual, we conducted a large-scale random survey within Japan of individuals older than 18 years, using a printed questionnaire. The data thus acquired were analyzed using factor analysis and cluster analysis to create a 3-factor/4-cluster model of people's interest and attitude toward science, technology and society and their resulting tendencies. Differences were found among the four clusters in terms of the three factors: scientific factor, social factor, and science-appreciating factor. We propose a plan for designing a form of science communication that is appropriate to this current status of scientific literacy in Japan.

  16. Sample size adjustments for varying cluster sizes in cluster randomized trials with binary outcomes analyzed with second-order PQL mixed logistic regression.

    PubMed

    Candel, Math J J M; Van Breukelen, Gerard J P

    2010-06-30

    Adjustments of sample size formulas are given for varying cluster sizes in cluster randomized trials with a binary outcome when testing the treatment effect with mixed effects logistic regression using second-order penalized quasi-likelihood estimation (PQL). Starting from first-order marginal quasi-likelihood (MQL) estimation of the treatment effect, the asymptotic relative efficiency of unequal versus equal cluster sizes is derived. A Monte Carlo simulation study shows this asymptotic relative efficiency to be rather accurate for realistic sample sizes, when employing second-order PQL. An approximate, simpler formula is presented to estimate the efficiency loss due to varying cluster sizes when planning a trial. In many cases sampling 14 per cent more clusters is sufficient to repair the efficiency loss due to varying cluster sizes. Since current closed-form formulas for sample size calculation are based on first-order MQL, planning a trial also requires a conversion factor to obtain the variance of the second-order PQL estimator. In a second Monte Carlo study, this conversion factor turned out to be 1.25 at most. (c) 2010 John Wiley & Sons, Ltd.

  17. Multiple Imputation in Two-Stage Cluster Samples Using The Weighted Finite Population Bayesian Bootstrap.

    PubMed

    Zhou, Hanzhi; Elliott, Michael R; Raghunathan, Trivellore E

    2016-06-01

    Multistage sampling is often employed in survey samples for cost and convenience. However, accounting for clustering features when generating datasets for multiple imputation is a nontrivial task, particularly when, as is often the case, cluster sampling is accompanied by unequal probabilities of selection, necessitating case weights. Thus, multiple imputation often ignores complex sample designs and assumes simple random sampling when generating imputations, even though failing to account for complex sample design features is known to yield biased estimates and confidence intervals that have incorrect nominal coverage. In this article, we extend a recently developed, weighted, finite-population Bayesian bootstrap procedure to generate synthetic populations conditional on complex sample design data that can be treated as simple random samples at the imputation stage, obviating the need to directly model design features for imputation. We develop two forms of this method: one where the probabilities of selection are known at the first and second stages of the design, and the other, more common in public use files, where only the final weight based on the product of the two probabilities is known. We show that this method has advantages in terms of bias, mean square error, and coverage properties over methods where sample designs are ignored, with little loss in efficiency, even when compared with correct fully parametric models. An application is made using the National Automotive Sampling System Crashworthiness Data System, a multistage, unequal probability sample of U.S. passenger vehicle crashes, which suffers from a substantial amount of missing data in "Delta-V," a key crash severity measure.

  18. Multiple Imputation in Two-Stage Cluster Samples Using The Weighted Finite Population Bayesian Bootstrap

    PubMed Central

    Zhou, Hanzhi; Elliott, Michael R.; Raghunathan, Trivellore E.

    2017-01-01

    Multistage sampling is often employed in survey samples for cost and convenience. However, accounting for clustering features when generating datasets for multiple imputation is a nontrivial task, particularly when, as is often the case, cluster sampling is accompanied by unequal probabilities of selection, necessitating case weights. Thus, multiple imputation often ignores complex sample designs and assumes simple random sampling when generating imputations, even though failing to account for complex sample design features is known to yield biased estimates and confidence intervals that have incorrect nominal coverage. In this article, we extend a recently developed, weighted, finite-population Bayesian bootstrap procedure to generate synthetic populations conditional on complex sample design data that can be treated as simple random samples at the imputation stage, obviating the need to directly model design features for imputation. We develop two forms of this method: one where the probabilities of selection are known at the first and second stages of the design, and the other, more common in public use files, where only the final weight based on the product of the two probabilities is known. We show that this method has advantages in terms of bias, mean square error, and coverage properties over methods where sample designs are ignored, with little loss in efficiency, even when compared with correct fully parametric models. An application is made using the National Automotive Sampling System Crashworthiness Data System, a multistage, unequal probability sample of U.S. passenger vehicle crashes, which suffers from a substantial amount of missing data in “Delta-V,” a key crash severity measure. PMID:29226161

  19. Review of Recent Methodological Developments in Group-Randomized Trials: Part 2-Analysis.

    PubMed

    Turner, Elizabeth L; Prague, Melanie; Gallis, John A; Li, Fan; Murray, David M

    2017-07-01

    In 2004, Murray et al. reviewed methodological developments in the design and analysis of group-randomized trials (GRTs). We have updated that review with developments in analysis of the past 13 years, with a companion article to focus on developments in design. We discuss developments in the topics of the earlier review (e.g., methods for parallel-arm GRTs, individually randomized group-treatment trials, and missing data) and in new topics, including methods to account for multiple-level clustering and alternative estimation methods (e.g., augmented generalized estimating equations, targeted maximum likelihood, and quadratic inference functions). In addition, we describe developments in analysis of alternative group designs (including stepped-wedge GRTs, network-randomized trials, and pseudocluster randomized trials), which require clustering to be accounted for in their design and analysis.

  20. Cluster-Glass Phase in Pyrochlore X Y Antiferromagnets with Quenched Disorder

    NASA Astrophysics Data System (ADS)

    Andrade, Eric C.; Hoyos, José A.; Rachel, Stephan; Vojta, Matthias

    2018-03-01

    We study the impact of quenched disorder (random exchange couplings or site dilution) on easy-plane pyrochlore antiferromagnets. In the clean system, order by disorder selects a magnetically ordered state from a classically degenerate manifold. In the presence of randomness, however, different orders can be chosen locally depending on details of the disorder configuration. Using a combination of analytical considerations and classical Monte Carlo simulations, we argue that any long-range-ordered magnetic state is destroyed beyond a critical level of randomness where the system breaks into magnetic domains due to random exchange anisotropies, becoming, therefore, a glass of spin clusters, in accordance with the available experimental data. These random anisotropies originate from off-diagonal exchange couplings in the microscopic Hamiltonian, establishing their relevance to other magnets with strong spin-orbit coupling.

Top