Nie, Z Q; Ou, Y Q; Zhuang, J; Qu, Y J; Mai, J Z; Chen, J M; Liu, X Q
2016-05-01
Conditional logistic regression analysis and unconditional logistic regression analysis are commonly used in case control study, but Cox proportional hazard model is often used in survival data analysis. Most literature only refer to main effect model, however, generalized linear model differs from general linear model, and the interaction was composed of multiplicative interaction and additive interaction. The former is only statistical significant, but the latter has biological significance. In this paper, macros was written by using SAS 9.4 and the contrast ratio, attributable proportion due to interaction and synergy index were calculated while calculating the items of logistic and Cox regression interactions, and the confidence intervals of Wald, delta and profile likelihood were used to evaluate additive interaction for the reference in big data analysis in clinical epidemiology and in analysis of genetic multiplicative and additive interactions.
Regression analysis of informative current status data with the additive hazards model.
Zhao, Shishun; Hu, Tao; Ma, Ling; Wang, Peijie; Sun, Jianguo
2015-04-01
This paper discusses regression analysis of current status failure time data arising from the additive hazards model in the presence of informative censoring. Many methods have been developed for regression analysis of current status data under various regression models if the censoring is noninformative, and also there exists a large literature on parametric analysis of informative current status data in the context of tumorgenicity experiments. In this paper, a semiparametric maximum likelihood estimation procedure is presented and in the method, the copula model is employed to describe the relationship between the failure time of interest and the censoring time. Furthermore, I-splines are used to approximate the nonparametric functions involved and the asymptotic consistency and normality of the proposed estimators are established. A simulation study is conducted and indicates that the proposed approach works well for practical situations. An illustrative example is also provided.
Feaster, Toby D.; Gotvald, Anthony J.; Weaver, J. Curtis
2014-01-01
Reliable estimates of the magnitude and frequency of floods are essential for the design of transportation and water-conveyance structures, flood-insurance studies, and flood-plain management. Such estimates are particularly important in densely populated urban areas. In order to increase the number of streamflow-gaging stations (streamgages) available for analysis, expand the geographical coverage that would allow for application of regional regression equations across State boundaries, and build on a previous flood-frequency investigation of rural U.S Geological Survey streamgages in the Southeast United States, a multistate approach was used to update methods for determining the magnitude and frequency of floods in urban and small, rural streams that are not substantially affected by regulation or tidal fluctuations in Georgia, South Carolina, and North Carolina. The at-site flood-frequency analysis of annual peak-flow data for urban and small, rural streams (through September 30, 2011) included 116 urban streamgages and 32 small, rural streamgages, defined in this report as basins draining less than 1 square mile. The regional regression analysis included annual peak-flow data from an additional 338 rural streamgages previously included in U.S. Geological Survey flood-frequency reports and 2 additional rural streamgages in North Carolina that were not included in the previous Southeast rural flood-frequency investigation for a total of 488 streamgages included in the urban and small, rural regression analysis. The at-site flood-frequency analyses for the urban and small, rural streamgages included the expected moments algorithm, which is a modification of the Bulletin 17B log-Pearson type III method for fitting the statistical distribution to the logarithms of the annual peak flows. Where applicable, the flood-frequency analysis also included low-outlier and historic information. Additionally, the application of a generalized Grubbs-Becks test allowed for the detection of multiple potentially influential low outliers. Streamgage basin characteristics were determined using geographical information system techniques. Initial ordinary least squares regression simulations reduced the number of basin characteristics on the basis of such factors as statistical significance, coefficient of determination, Mallow’s Cp statistic, and ease of measurement of the explanatory variable. Application of generalized least squares regression techniques produced final predictive (regression) equations for estimating the 50-, 20-, 10-, 4-, 2-, 1-, 0.5-, and 0.2-percent annual exceedance probability flows for urban and small, rural ungaged basins for three hydrologic regions (HR1, Piedmont–Ridge and Valley; HR3, Sand Hills; and HR4, Coastal Plain), which previously had been defined from exploratory regression analysis in the Southeast rural flood-frequency investigation. Because of the limited availability of urban streamgages in the Coastal Plain of Georgia, South Carolina, and North Carolina, additional urban streamgages in Florida and New Jersey were used in the regression analysis for this region. Including the urban streamgages in New Jersey allowed for the expansion of the applicability of the predictive equations in the Coastal Plain from 3.5 to 53.5 square miles. Average standard error of prediction for the predictive equations, which is a measure of the average accuracy of the regression equations when predicting flood estimates for ungaged sites, range from 25.0 percent for the 10-percent annual exceedance probability regression equation for the Piedmont–Ridge and Valley region to 73.3 percent for the 0.2-percent annual exceedance probability regression equation for the Sand Hills region.
Development of a User Interface for a Regression Analysis Software Tool
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred; Volden, Thomas R.
2010-01-01
An easy-to -use user interface was implemented in a highly automated regression analysis tool. The user interface was developed from the start to run on computers that use the Windows, Macintosh, Linux, or UNIX operating system. Many user interface features were specifically designed such that a novice or inexperienced user can apply the regression analysis tool with confidence. Therefore, the user interface s design minimizes interactive input from the user. In addition, reasonable default combinations are assigned to those analysis settings that influence the outcome of the regression analysis. These default combinations will lead to a successful regression analysis result for most experimental data sets. The user interface comes in two versions. The text user interface version is used for the ongoing development of the regression analysis tool. The official release of the regression analysis tool, on the other hand, has a graphical user interface that is more efficient to use. This graphical user interface displays all input file names, output file names, and analysis settings for a specific software application mode on a single screen which makes it easier to generate reliable analysis results and to perform input parameter studies. An object-oriented approach was used for the development of the graphical user interface. This choice keeps future software maintenance costs to a reasonable limit. Examples of both the text user interface and graphical user interface are discussed in order to illustrate the user interface s overall design approach.
Quantile regression for the statistical analysis of immunological data with many non-detects.
Eilers, Paul H C; Röder, Esther; Savelkoul, Huub F J; van Wijk, Roy Gerth
2012-07-07
Immunological parameters are hard to measure. A well-known problem is the occurrence of values below the detection limit, the non-detects. Non-detects are a nuisance, because classical statistical analyses, like ANOVA and regression, cannot be applied. The more advanced statistical techniques currently available for the analysis of datasets with non-detects can only be used if a small percentage of the data are non-detects. Quantile regression, a generalization of percentiles to regression models, models the median or higher percentiles and tolerates very high numbers of non-detects. We present a non-technical introduction and illustrate it with an implementation to real data from a clinical trial. We show that by using quantile regression, groups can be compared and that meaningful linear trends can be computed, even if more than half of the data consists of non-detects. Quantile regression is a valuable addition to the statistical methods that can be used for the analysis of immunological datasets with non-detects.
Kanada, Yoshikiyo; Sakurai, Hiroaki; Sugiura, Yoshito; Arai, Tomoaki; Koyama, Soichiro; Tanabe, Shigeo
2017-11-01
[Purpose] To create a regression formula in order to estimate 1RM for knee extensors, based on the maximal isometric muscle strength measured using a hand-held dynamometer and data regarding the body composition. [Subjects and Methods] Measurement was performed in 21 healthy males in their twenties to thirties. Single regression analysis was performed, with measurement values representing 1RM and the maximal isometric muscle strength as dependent and independent variables, respectively. Furthermore, multiple regression analysis was performed, with data regarding the body composition incorporated as another independent variable, in addition to the maximal isometric muscle strength. [Results] Through single regression analysis with the maximal isometric muscle strength as an independent variable, the following regression formula was created: 1RM (kg)=0.714 + 0.783 × maximal isometric muscle strength (kgf). On multiple regression analysis, only the total muscle mass was extracted. [Conclusion] A highly accurate regression formula to estimate 1RM was created based on both the maximal isometric muscle strength and body composition. Using a hand-held dynamometer and body composition analyzer, it was possible to measure these items in a short time, and obtain clinically useful results.
Analytics For Distracted Driver Behavior Modeling in Dilemma Zone
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Jan-Mou; Malikopoulos, Andreas; Thakur, Gautam
2014-01-01
In this paper, we present the results obtained and insights gained through the analysis of TRB contest data. We used exploratory analysis, regression, and clustering models for gaining insights into the driver behavior in a dilemma zone while driving under distraction. While simple exploratory analysis showed the distinguishing driver behavior patterns among different popu- lation groups in the dilemma zone, regression analysis showed statically signification relationships between groups of variables. In addition to analyzing the contest data, we have also looked into the possible impact of distracted driving on the fuel economy.
Experimental investigation of fuel regression rate in a HTPB based lab-scale hybrid rocket motor
NASA Astrophysics Data System (ADS)
Li, Xintian; Tian, Hui; Yu, Nanjia; Cai, Guobiao
2014-12-01
The fuel regression rate is an important parameter in the design process of the hybrid rocket motor. Additives in the solid fuel may have influences on the fuel regression rate, which will affect the internal ballistics of the motor. A series of firing experiments have been conducted on lab-scale hybrid rocket motors with 98% hydrogen peroxide (H2O2) oxidizer and hydroxyl terminated polybutadiene (HTPB) based fuels in this paper. An innovative fuel regression rate analysis method is established to diminish the errors caused by start and tailing stages in a short time firing test. The effects of the metal Mg, Al, aromatic hydrocarbon anthracene (C14H10), and carbon black (C) on the fuel regression rate are investigated. The fuel regression rate formulas of different fuel components are fitted according to the experiment data. The results indicate that the influence of C14H10 on the fuel regression rate of HTPB is not evident. However, the metal additives in the HTPB fuel can increase the fuel regression rate significantly.
Iterative Strain-Gage Balance Calibration Data Analysis for Extended Independent Variable Sets
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred
2011-01-01
A new method was developed that makes it possible to use an extended set of independent calibration variables for an iterative analysis of wind tunnel strain gage balance calibration data. The new method permits the application of the iterative analysis method whenever the total number of balance loads and other independent calibration variables is greater than the total number of measured strain gage outputs. Iteration equations used by the iterative analysis method have the limitation that the number of independent and dependent variables must match. The new method circumvents this limitation. It simply adds a missing dependent variable to the original data set by using an additional independent variable also as an additional dependent variable. Then, the desired solution of the regression analysis problem can be obtained that fits each gage output as a function of both the original and additional independent calibration variables. The final regression coefficients can be converted to data reduction matrix coefficients because the missing dependent variables were added to the data set without changing the regression analysis result for each gage output. Therefore, the new method still supports the application of the two load iteration equation choices that the iterative method traditionally uses for the prediction of balance loads during a wind tunnel test. An example is discussed in the paper that illustrates the application of the new method to a realistic simulation of temperature dependent calibration data set of a six component balance.
A primer for biomedical scientists on how to execute model II linear regression analysis.
Ludbrook, John
2012-04-01
1. There are two very different ways of executing linear regression analysis. One is Model I, when the x-values are fixed by the experimenter. The other is Model II, in which the x-values are free to vary and are subject to error. 2. I have received numerous complaints from biomedical scientists that they have great difficulty in executing Model II linear regression analysis. This may explain the results of a Google Scholar search, which showed that the authors of articles in journals of physiology, pharmacology and biochemistry rarely use Model II regression analysis. 3. I repeat my previous arguments in favour of using least products linear regression analysis for Model II regressions. I review three methods for executing ordinary least products (OLP) and weighted least products (WLP) regression analysis: (i) scientific calculator and/or computer spreadsheet; (ii) specific purpose computer programs; and (iii) general purpose computer programs. 4. Using a scientific calculator and/or computer spreadsheet, it is easy to obtain correct values for OLP slope and intercept, but the corresponding 95% confidence intervals (CI) are inaccurate. 5. Using specific purpose computer programs, the freeware computer program smatr gives the correct OLP regression coefficients and obtains 95% CI by bootstrapping. In addition, smatr can be used to compare the slopes of OLP lines. 6. When using general purpose computer programs, I recommend the commercial programs systat and Statistica for those who regularly undertake linear regression analysis and I give step-by-step instructions in the Supplementary Information as to how to use loss functions. © 2011 The Author. Clinical and Experimental Pharmacology and Physiology. © 2011 Blackwell Publishing Asia Pty Ltd.
Trend Analysis Using Microcomputers.
ERIC Educational Resources Information Center
Berger, Carl F.
A trend analysis statistical package and additional programs for the Apple microcomputer are presented. They illustrate strategies of data analysis suitable to the graphics and processing capabilities of the microcomputer. The programs analyze data sets using examples of: (1) analysis of variance with multiple linear regression; (2) exponential…
Regression with Small Data Sets: A Case Study using Code Surrogates in Additive Manufacturing
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kamath, C.; Fan, Y. J.
There has been an increasing interest in recent years in the mining of massive data sets whose sizes are measured in terabytes. While it is easy to collect such large data sets in some application domains, there are others where collecting even a single data point can be very expensive, so the resulting data sets have only tens or hundreds of samples. For example, when complex computer simulations are used to understand a scientific phenomenon, we want to run the simulation for many different values of the input parameters and analyze the resulting output. The data set relating the simulationmore » inputs and outputs is typically quite small, especially when each run of the simulation is expensive. However, regression techniques can still be used on such data sets to build an inexpensive \\surrogate" that could provide an approximate output for a given set of inputs. A good surrogate can be very useful in sensitivity analysis, uncertainty analysis, and in designing experiments. In this paper, we compare different regression techniques to determine how well they predict melt-pool characteristics in the problem domain of additive manufacturing. Our analysis indicates that some of the commonly used regression methods do perform quite well even on small data sets.« less
Specific factors for prenatal lead exposure in the border area of China.
Kawata, Kimiko; Li, Yan; Liu, Hao; Zhang, Xiao Qin; Ushijima, Hiroshi
2006-07-01
The objectives of this study are to examine the prevalence of increased blood lead concentrations in mothers and their umbilical cords, and to identify risk factors for prenatal lead exposure in Kunming city, Yunnan province, China. The study was conducted at two obstetrics departments, and 100 peripartum women were enrolled. The mean blood lead concentrations of the mothers and the umbilical cords were 67.3microg/l and 53.1microg/l, respectively. In multiple linear regression analysis, maternal occupational exposure, maternal consumption of homemade dehydrated vegetables and maternal habitation period in Kunming city were significantly associated with an increase of umbilical cord blood lead concentration. In addition, logistic regression analysis was used to assess the association of umbilical cord blood lead concentrations that possibly have adverse effects on brain development of newborns with each potential risk factor. Maternal frequent use of tableware with color patterns inside was significantly associated with higher cord blood lead concentration in addition to the three items in the multiple linear regression analysis. These points should be considered as specific recommendations for maternal and fetal lead exposure in this city.
Understanding poisson regression.
Hayat, Matthew J; Higgins, Melinda
2014-04-01
Nurse investigators often collect study data in the form of counts. Traditional methods of data analysis have historically approached analysis of count data either as if the count data were continuous and normally distributed or with dichotomization of the counts into the categories of occurred or did not occur. These outdated methods for analyzing count data have been replaced with more appropriate statistical methods that make use of the Poisson probability distribution, which is useful for analyzing count data. The purpose of this article is to provide an overview of the Poisson distribution and its use in Poisson regression. Assumption violations for the standard Poisson regression model are addressed with alternative approaches, including addition of an overdispersion parameter or negative binomial regression. An illustrative example is presented with an application from the ENSPIRE study, and regression modeling of comorbidity data is included for illustrative purposes. Copyright 2014, SLACK Incorporated.
ERIC Educational Resources Information Center
Cohen, Ayala; Nahum-Shani, Inbal; Doveh, Etti
2010-01-01
In their seminal paper, Edwards and Parry (1993) presented the polynomial regression as a better alternative to applying difference score in the study of congruence. Although this method is increasingly applied in congruence research, its complexity relative to other methods for assessing congruence (e.g., difference score methods) was one of the…
Structured functional additive regression in reproducing kernel Hilbert spaces.
Zhu, Hongxiao; Yao, Fang; Zhang, Hao Helen
2014-06-01
Functional additive models (FAMs) provide a flexible yet simple framework for regressions involving functional predictors. The utilization of data-driven basis in an additive rather than linear structure naturally extends the classical functional linear model. However, the critical issue of selecting nonlinear additive components has been less studied. In this work, we propose a new regularization framework for the structure estimation in the context of Reproducing Kernel Hilbert Spaces. The proposed approach takes advantage of the functional principal components which greatly facilitates the implementation and the theoretical analysis. The selection and estimation are achieved by penalized least squares using a penalty which encourages the sparse structure of the additive components. Theoretical properties such as the rate of convergence are investigated. The empirical performance is demonstrated through simulation studies and a real data application.
The ice age cycle and the deglaciations: an application of nonlinear regression modelling
NASA Astrophysics Data System (ADS)
Dalgleish, A. N.; Boulton, G. S.; Renshaw, E.
2000-03-01
We have applied the nonlinear regression technique known as additivity and variance stabilisation (AVAS) to time series which reflect Earth's climate over the last 600 ka. AVAS estimates a smooth, nonlinear transform for each variable, under the assumption of an additive model. The Earth's orbital parameters and insolation variations have been used as regression variables. Analysis of the contribution of each variable shows that the deglaciations are characterised by periods of increasing obliquity and perihelion approaching the vernal equinox, but not by any systematic change in eccentricity. The magnitude of insolation changes also plays no role. By approximating the transforms we can obtain a future prediction, with a glacial maximum at 60 ka AP, and a subsequent obliquity and precession forced deglaciation.
Regression Model Term Selection for the Analysis of Strain-Gage Balance Calibration Data
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred; Volden, Thomas R.
2010-01-01
The paper discusses the selection of regression model terms for the analysis of wind tunnel strain-gage balance calibration data. Different function class combinations are presented that may be used to analyze calibration data using either a non-iterative or an iterative method. The role of the intercept term in a regression model of calibration data is reviewed. In addition, useful algorithms and metrics originating from linear algebra and statistics are recommended that will help an analyst (i) to identify and avoid both linear and near-linear dependencies between regression model terms and (ii) to make sure that the selected regression model of the calibration data uses only statistically significant terms. Three different tests are suggested that may be used to objectively assess the predictive capability of the final regression model of the calibration data. These tests use both the original data points and regression model independent confirmation points. Finally, data from a simplified manual calibration of the Ames MK40 balance is used to illustrate the application of some of the metrics and tests to a realistic calibration data set.
Wang, Wen-Cheng; Cho, Wen-Chien; Chen, Yin-Jen
2014-01-01
It is estimated that mainland Chinese tourists travelling to Taiwan can bring annual revenues of 400 billion NTD to the Taiwan economy. Thus, how the Taiwanese Government formulates relevant measures to satisfy both sides is the focus of most concern. Taiwan must improve the facilities and service quality of its tourism industry so as to attract more mainland tourists. This paper conducted a questionnaire survey of mainland tourists and used grey relational analysis in grey mathematics to analyze the satisfaction performance of all satisfaction question items. The first eight satisfaction items were used as independent variables, and the overall satisfaction performance was used as a dependent variable for quantile regression model analysis to discuss the relationship between the dependent variable under different quantiles and independent variables. Finally, this study further discussed the predictive accuracy of the least mean regression model and each quantile regression model, as a reference for research personnel. The analysis results showed that other variables could also affect the overall satisfaction performance of mainland tourists, in addition to occupation and age. The overall predictive accuracy of quantile regression model Q0.25 was higher than that of the other three models. PMID:24574916
Wang, Wen-Cheng; Cho, Wen-Chien; Chen, Yin-Jen
2014-01-01
It is estimated that mainland Chinese tourists travelling to Taiwan can bring annual revenues of 400 billion NTD to the Taiwan economy. Thus, how the Taiwanese Government formulates relevant measures to satisfy both sides is the focus of most concern. Taiwan must improve the facilities and service quality of its tourism industry so as to attract more mainland tourists. This paper conducted a questionnaire survey of mainland tourists and used grey relational analysis in grey mathematics to analyze the satisfaction performance of all satisfaction question items. The first eight satisfaction items were used as independent variables, and the overall satisfaction performance was used as a dependent variable for quantile regression model analysis to discuss the relationship between the dependent variable under different quantiles and independent variables. Finally, this study further discussed the predictive accuracy of the least mean regression model and each quantile regression model, as a reference for research personnel. The analysis results showed that other variables could also affect the overall satisfaction performance of mainland tourists, in addition to occupation and age. The overall predictive accuracy of quantile regression model Q0.25 was higher than that of the other three models.
Ecologic regression analysis and the study of the influence of air quality on mortality.
Selvin, S; Merrill, D; Wong, L; Sacks, S T
1984-01-01
This presentation focuses entirely on the use and evaluation of regression analysis applied to ecologic data as a method to study the effects of ambient air pollution on mortality rates. Using extensive national data on mortality, air quality and socio-economic status regression analyses are used to study the influence of air quality on mortality. The analytic methods and data are selected in such a way that direct comparisons can be made with other ecologic regression studies of mortality and air quality. Analyses are performed by use of two types of geographic areas, age-specific mortality of both males and females and three pollutants (total suspended particulates, sulfur dioxide and nitrogen dioxide). The overall results indicate no persuasive evidence exists of a link between air quality and general mortality levels. Additionally, a lack of consistency between the present results and previous published work is noted. Overall, it is concluded that linear regression analysis applied to nationally collected ecologic data cannot be used to usefully infer a causal relationship between air quality and mortality which is in direct contradiction to other major published studies. PMID:6734568
Application of stepwise multiple regression techniques to inversion of Nimbus 'IRIS' observations.
NASA Technical Reports Server (NTRS)
Ohring, G.
1972-01-01
Exploratory studies with Nimbus-3 infrared interferometer-spectrometer (IRIS) data indicate that, in addition to temperature, such meteorological parameters as geopotential heights of pressure surfaces, tropopause pressure, and tropopause temperature can be inferred from the observed spectra with the use of simple regression equations. The technique of screening the IRIS spectral data by means of stepwise regression to obtain the best radiation predictors of meteorological parameters is validated. The simplicity of application of the technique and the simplicity of the derived linear regression equations - which contain only a few terms - suggest usefulness for this approach. Based upon the results obtained, suggestions are made for further development and exploitation of the stepwise regression analysis technique.
Neither fixed nor random: weighted least squares meta-regression.
Stanley, T D; Doucouliagos, Hristos
2017-03-01
Our study revisits and challenges two core conventional meta-regression estimators: the prevalent use of 'mixed-effects' or random-effects meta-regression analysis and the correction of standard errors that defines fixed-effects meta-regression analysis (FE-MRA). We show how and explain why an unrestricted weighted least squares MRA (WLS-MRA) estimator is superior to conventional random-effects (or mixed-effects) meta-regression when there is publication (or small-sample) bias that is as good as FE-MRA in all cases and better than fixed effects in most practical applications. Simulations and statistical theory show that WLS-MRA provides satisfactory estimates of meta-regression coefficients that are practically equivalent to mixed effects or random effects when there is no publication bias. When there is publication selection bias, WLS-MRA always has smaller bias than mixed effects or random effects. In practical applications, an unrestricted WLS meta-regression is likely to give practically equivalent or superior estimates to fixed-effects, random-effects, and mixed-effects meta-regression approaches. However, random-effects meta-regression remains viable and perhaps somewhat preferable if selection for statistical significance (publication bias) can be ruled out and when random, additive normal heterogeneity is known to directly affect the 'true' regression coefficient. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Straub, D.E.
1998-01-01
The streamflow-gaging station network in Ohio was evaluated for its effectiveness in providing regional streamflow information. The analysis involved application of the principles of generalized least squares regression between streamflow and climatic and basin characteristics. Regression equations were developed for three flow characteristics: (1) the instantaneous peak flow with a 100-year recurrence interval (P100), (2) the mean annual flow (Qa), and (3) the 7-day, 10-year low flow (7Q10). All active and discontinued gaging stations with 5 or more years of unregulated-streamflow data with respect to each flow characteristic were used to develop the regression equations. The gaging-station network was evaluated for the current (1996) condition of the network and estimated conditions of various network strategies if an additional 5 and 20 years of streamflow data were collected. Any active or discontinued gaging station with (1) less than 5 years of unregulated-streamflow record, (2) previously defined basin and climatic characteristics, and (3) the potential for collection of more unregulated-streamflow record were included in the network strategies involving the additional 5 and 20 years of data. The network analysis involved use of the regression equations, in combination with location, period of record, and cost of operation, to determine the contribution of the data for each gaging station to regional streamflow information. The contribution of each gaging station was based on a cost-weighted reduction of the mean square error (average sampling-error variance) associated with each regional estimating equation. All gaging stations included in the network analysis were then ranked according to their contribution to the regional information for each flow characteristic. The predictive ability of the regression equations developed from the gaging station network could be improved for all three flow characteristics with the collection of additional streamflow data. The addition of new gaging stations to the network would result in an even greater improvement of the accuracy of the regional regression equations. Typically, continued data collection at stations with unregulated streamflow for all flow conditions that had less than 11 years of record with drainage areas smaller than 200 square miles contributed the largest cost-weighted reduction to the average sampling-error variance of the regional estimating equations. The results of the network analyses can be used to prioritize the continued operation of active gaging stations or the reactivation of discontinued gaging stations if the objective is to maximize the regional information content in the streamflow-gaging station network.
Productivity: Vocational Education's Role. Information Series No. 223.
ERIC Educational Resources Information Center
Bolino, August C.
This paper's overiew of the relationship between vocational education and productivity includes the presentation of results from a multiple regression analysis of vocational education enrollments and various productivity indices. This tentative analysis contributes additional observations to the studies reviewed and offers pertinent suggestions…
Vidyasagar, Mathukumalli
2015-01-01
This article reviews several techniques from machine learning that can be used to study the problem of identifying a small number of features, from among tens of thousands of measured features, that can accurately predict a drug response. Prediction problems are divided into two categories: sparse classification and sparse regression. In classification, the clinical parameter to be predicted is binary, whereas in regression, the parameter is a real number. Well-known methods for both classes of problems are briefly discussed. These include the SVM (support vector machine) for classification and various algorithms such as ridge regression, LASSO (least absolute shrinkage and selection operator), and EN (elastic net) for regression. In addition, several well-established methods that do not directly fall into machine learning theory are also reviewed, including neural networks, PAM (pattern analysis for microarrays), SAM (significance analysis for microarrays), GSEA (gene set enrichment analysis), and k-means clustering. Several references indicative of the application of these methods to cancer biology are discussed.
Structured functional additive regression in reproducing kernel Hilbert spaces
Zhu, Hongxiao; Yao, Fang; Zhang, Hao Helen
2013-01-01
Summary Functional additive models (FAMs) provide a flexible yet simple framework for regressions involving functional predictors. The utilization of data-driven basis in an additive rather than linear structure naturally extends the classical functional linear model. However, the critical issue of selecting nonlinear additive components has been less studied. In this work, we propose a new regularization framework for the structure estimation in the context of Reproducing Kernel Hilbert Spaces. The proposed approach takes advantage of the functional principal components which greatly facilitates the implementation and the theoretical analysis. The selection and estimation are achieved by penalized least squares using a penalty which encourages the sparse structure of the additive components. Theoretical properties such as the rate of convergence are investigated. The empirical performance is demonstrated through simulation studies and a real data application. PMID:25013362
Analysis of Sting Balance Calibration Data Using Optimized Regression Models
NASA Technical Reports Server (NTRS)
Ulbrich, N.; Bader, Jon B.
2010-01-01
Calibration data of a wind tunnel sting balance was processed using a candidate math model search algorithm that recommends an optimized regression model for the data analysis. During the calibration the normal force and the moment at the balance moment center were selected as independent calibration variables. The sting balance itself had two moment gages. Therefore, after analyzing the connection between calibration loads and gage outputs, it was decided to choose the difference and the sum of the gage outputs as the two responses that best describe the behavior of the balance. The math model search algorithm was applied to these two responses. An optimized regression model was obtained for each response. Classical strain gage balance load transformations and the equations of the deflection of a cantilever beam under load are used to show that the search algorithm s two optimized regression models are supported by a theoretical analysis of the relationship between the applied calibration loads and the measured gage outputs. The analysis of the sting balance calibration data set is a rare example of a situation when terms of a regression model of a balance can directly be derived from first principles of physics. In addition, it is interesting to note that the search algorithm recommended the correct regression model term combinations using only a set of statistical quality metrics that were applied to the experimental data during the algorithm s term selection process.
Parental education predicts change in intelligence quotient after childhood epilepsy surgery.
Meekes, Joost; van Schooneveld, Monique M J; Braams, Olga B; Jennekens-Schinkel, Aag; van Rijen, Peter C; Hendriks, Marc P H; Braun, Kees P J; van Nieuwenhuizen, Onno
2015-04-01
To know whether change in the intelligence quotient (IQ) of children who undergo epilepsy surgery is associated with the educational level of their parents. Retrospective analysis of data obtained from a cohort of children who underwent epilepsy surgery between January 1996 and September 2010. We performed simple and multiple regression analyses to identify predictors associated with IQ change after surgery. In addition to parental education, six variables previously demonstrated to be associated with IQ change after surgery were included as predictors: age at surgery, duration of epilepsy, etiology, presurgical IQ, reduction of antiepileptic drugs, and seizure freedom. We used delta IQ (IQ 2 years after surgery minus IQ shortly before surgery) as the primary outcome variable, but also performed analyses with pre- and postsurgical IQ as outcome variables to support our findings. To validate the results we performed simple regression analysis with parental education as the predictor in specific subgroups. The sample for regression analysis included 118 children (60 male; median age at surgery 9.73 years). Parental education was significantly associated with delta IQ in simple regression analysis (p = 0.004), and also contributed significantly to postsurgical IQ in multiple regression analysis (p = 0.008). Additional analyses demonstrated that parental education made a unique contribution to prediction of delta IQ, that is, it could not be replaced by the illness-related variables. Subgroup analyses confirmed the association of parental education with IQ change after surgery for most groups. Children whose parents had higher education demonstrate on average a greater increase in IQ after surgery and a higher postsurgical--but not presurgical--IQ than children whose parents completed at most lower secondary education. Parental education--and perhaps other environmental variables--should be considered in the prognosis of cognitive function after childhood epilepsy surgery. Wiley Periodicals, Inc. © 2015 International League Against Epilepsy.
Factor analysis and multiple regression between topography and precipitation on Jeju Island, Korea
NASA Astrophysics Data System (ADS)
Um, Myoung-Jin; Yun, Hyeseon; Jeong, Chang-Sam; Heo, Jun-Haeng
2011-11-01
SummaryIn this study, new factors that influence precipitation were extracted from geographic variables using factor analysis, which allow for an accurate estimation of orographic precipitation. Correlation analysis was also used to examine the relationship between nine topographic variables from digital elevation models (DEMs) and the precipitation in Jeju Island. In addition, a spatial analysis was performed in order to verify the validity of the regression model. From the results of the correlation analysis, it was found that all of the topographic variables had a positive correlation with the precipitation. The relations between the variables also changed in accordance with a change in the precipitation duration. However, upon examining the correlation matrix, no significant relationship between the latitude and the aspect was found. According to the factor analysis, eight topographic variables (latitude being the exception) were found to have a direct influence on the precipitation. Three factors were then extracted from the eight topographic variables. By directly comparing the multiple regression model with the factors (model 1) to the multiple regression model with the topographic variables (model 3), it was found that model 1 did not violate the limits of statistical significance and multicollinearity. As such, model 1 was considered to be appropriate for estimating the precipitation when taking into account the topography. In the study of model 1, the multiple regression model using factor analysis was found to be the best method for estimating the orographic precipitation on Jeju Island.
Analysis of Sting Balance Calibration Data Using Optimized Regression Models
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert; Bader, Jon B.
2009-01-01
Calibration data of a wind tunnel sting balance was processed using a search algorithm that identifies an optimized regression model for the data analysis. The selected sting balance had two moment gages that were mounted forward and aft of the balance moment center. The difference and the sum of the two gage outputs were fitted in the least squares sense using the normal force and the pitching moment at the balance moment center as independent variables. The regression model search algorithm predicted that the difference of the gage outputs should be modeled using the intercept and the normal force. The sum of the two gage outputs, on the other hand, should be modeled using the intercept, the pitching moment, and the square of the pitching moment. Equations of the deflection of a cantilever beam are used to show that the search algorithm s two recommended math models can also be obtained after performing a rigorous theoretical analysis of the deflection of the sting balance under load. The analysis of the sting balance calibration data set is a rare example of a situation when regression models of balance calibration data can directly be derived from first principles of physics and engineering. In addition, it is interesting to see that the search algorithm recommended the same regression models for the data analysis using only a set of statistical quality metrics.
Meta-regression analysis of commensal and pathogenic Escherichia coli survival in soil and water.
Franz, Eelco; Schijven, Jack; de Roda Husman, Ana Maria; Blaak, Hetty
2014-06-17
The extent to which pathogenic and commensal E. coli (respectively PEC and CEC) can survive, and which factors predominantly determine the rate of decline, are crucial issues from a public health point of view. The goal of this study was to provide a quantitative summary of the variability in E. coli survival in soil and water over a broad range of individual studies and to identify the most important sources of variability. To that end, a meta-regression analysis on available literature data was conducted. The considerable variation in reported decline rates indicated that the persistence of E. coli is not easily predictable. The meta-analysis demonstrated that for soil and water, the type of experiment (laboratory or field), the matrix subtype (type of water and soil), and temperature were the main factors included in the regression analysis. A higher average decline rate in soil of PEC compared with CEC was observed. The regression models explained at best 57% of the variation in decline rate in soil and 41% of the variation in decline rate in water. This indicates that additional factors, not included in the current meta-regression analysis, are of importance but rarely reported. More complete reporting of experimental conditions may allow future inference on the global effects of these variables on the decline rate of E. coli.
Karabatsos, George
2017-02-01
Most of applied statistics involves regression analysis of data. In practice, it is important to specify a regression model that has minimal assumptions which are not violated by data, to ensure that statistical inferences from the model are informative and not misleading. This paper presents a stand-alone and menu-driven software package, Bayesian Regression: Nonparametric and Parametric Models, constructed from MATLAB Compiler. Currently, this package gives the user a choice from 83 Bayesian models for data analysis. They include 47 Bayesian nonparametric (BNP) infinite-mixture regression models; 5 BNP infinite-mixture models for density estimation; and 31 normal random effects models (HLMs), including normal linear models. Each of the 78 regression models handles either a continuous, binary, or ordinal dependent variable, and can handle multi-level (grouped) data. All 83 Bayesian models can handle the analysis of weighted observations (e.g., for meta-analysis), and the analysis of left-censored, right-censored, and/or interval-censored data. Each BNP infinite-mixture model has a mixture distribution assigned one of various BNP prior distributions, including priors defined by either the Dirichlet process, Pitman-Yor process (including the normalized stable process), beta (two-parameter) process, normalized inverse-Gaussian process, geometric weights prior, dependent Dirichlet process, or the dependent infinite-probits prior. The software user can mouse-click to select a Bayesian model and perform data analysis via Markov chain Monte Carlo (MCMC) sampling. After the sampling completes, the software automatically opens text output that reports MCMC-based estimates of the model's posterior distribution and model predictive fit to the data. Additional text and/or graphical output can be generated by mouse-clicking other menu options. This includes output of MCMC convergence analyses, and estimates of the model's posterior predictive distribution, for selected functionals and values of covariates. The software is illustrated through the BNP regression analysis of real data.
Lin, Feng-Chang; Zhu, Jun
2012-01-01
We develop continuous-time models for the analysis of environmental or ecological monitoring data such that subjects are observed at multiple monitoring time points across space. Of particular interest are additive hazards regression models where the baseline hazard function can take on flexible forms. We consider time-varying covariates and take into account spatial dependence via autoregression in space and time. We develop statistical inference for the regression coefficients via partial likelihood. Asymptotic properties, including consistency and asymptotic normality, are established for parameter estimates under suitable regularity conditions. Feasible algorithms utilizing existing statistical software packages are developed for computation. We also consider a simpler additive hazards model with homogeneous baseline hazard and develop hypothesis testing for homogeneity. A simulation study demonstrates that the statistical inference using partial likelihood has sound finite-sample properties and offers a viable alternative to maximum likelihood estimation. For illustration, we analyze data from an ecological study that monitors bark beetle colonization of red pines in a plantation of Wisconsin.
Spectroscopic analysis and control
Tate; , James D.; Reed, Christopher J.; Domke, Christopher H.; Le, Linh; Seasholtz, Mary Beth; Weber, Andy; Lipp, Charles
2017-04-18
Apparatus for spectroscopic analysis which includes a tunable diode laser spectrometer having a digital output signal and a digital computer for receiving the digital output signal from the spectrometer, the digital computer programmed to process the digital output signal using a multivariate regression algorithm. In addition, a spectroscopic method of analysis using such apparatus. Finally, a method for controlling an ethylene cracker hydrogenator.
Fenske, Nora; Burns, Jacob; Hothorn, Torsten; Rehfuess, Eva A.
2013-01-01
Background Most attempts to address undernutrition, responsible for one third of global child deaths, have fallen behind expectations. This suggests that the assumptions underlying current modelling and intervention practices should be revisited. Objective We undertook a comprehensive analysis of the determinants of child stunting in India, and explored whether the established focus on linear effects of single risks is appropriate. Design Using cross-sectional data for children aged 0–24 months from the Indian National Family Health Survey for 2005/2006, we populated an evidence-based diagram of immediate, intermediate and underlying determinants of stunting. We modelled linear, non-linear, spatial and age-varying effects of these determinants using additive quantile regression for four quantiles of the Z-score of standardized height-for-age and logistic regression for stunting and severe stunting. Results At least one variable within each of eleven groups of determinants was significantly associated with height-for-age in the 35% Z-score quantile regression. The non-modifiable risk factors child age and sex, and the protective factors household wealth, maternal education and BMI showed the largest effects. Being a twin or multiple birth was associated with dramatically decreased height-for-age. Maternal age, maternal BMI, birth order and number of antenatal visits influenced child stunting in non-linear ways. Findings across the four quantile and two logistic regression models were largely comparable. Conclusions Our analysis confirms the multifactorial nature of child stunting. It emphasizes the need to pursue a systems-based approach and to consider non-linear effects, and suggests that differential effects across the height-for-age distribution do not play a major role. PMID:24223839
Fenske, Nora; Burns, Jacob; Hothorn, Torsten; Rehfuess, Eva A
2013-01-01
Most attempts to address undernutrition, responsible for one third of global child deaths, have fallen behind expectations. This suggests that the assumptions underlying current modelling and intervention practices should be revisited. We undertook a comprehensive analysis of the determinants of child stunting in India, and explored whether the established focus on linear effects of single risks is appropriate. Using cross-sectional data for children aged 0-24 months from the Indian National Family Health Survey for 2005/2006, we populated an evidence-based diagram of immediate, intermediate and underlying determinants of stunting. We modelled linear, non-linear, spatial and age-varying effects of these determinants using additive quantile regression for four quantiles of the Z-score of standardized height-for-age and logistic regression for stunting and severe stunting. At least one variable within each of eleven groups of determinants was significantly associated with height-for-age in the 35% Z-score quantile regression. The non-modifiable risk factors child age and sex, and the protective factors household wealth, maternal education and BMI showed the largest effects. Being a twin or multiple birth was associated with dramatically decreased height-for-age. Maternal age, maternal BMI, birth order and number of antenatal visits influenced child stunting in non-linear ways. Findings across the four quantile and two logistic regression models were largely comparable. Our analysis confirms the multifactorial nature of child stunting. It emphasizes the need to pursue a systems-based approach and to consider non-linear effects, and suggests that differential effects across the height-for-age distribution do not play a major role.
Interrupted time series regression for the evaluation of public health interventions: a tutorial.
Bernal, James Lopez; Cummins, Steven; Gasparrini, Antonio
2017-02-01
Interrupted time series (ITS) analysis is a valuable study design for evaluating the effectiveness of population-level health interventions that have been implemented at a clearly defined point in time. It is increasingly being used to evaluate the effectiveness of interventions ranging from clinical therapy to national public health legislation. Whereas the design shares many properties of regression-based approaches in other epidemiological studies, there are a range of unique features of time series data that require additional methodological considerations. In this tutorial we use a worked example to demonstrate a robust approach to ITS analysis using segmented regression. We begin by describing the design and considering when ITS is an appropriate design choice. We then discuss the essential, yet often omitted, step of proposing the impact model a priori. Subsequently, we demonstrate the approach to statistical analysis including the main segmented regression model. Finally we describe the main methodological issues associated with ITS analysis: over-dispersion of time series data, autocorrelation, adjusting for seasonal trends and controlling for time-varying confounders, and we also outline some of the more complex design adaptations that can be used to strengthen the basic ITS design.
Interrupted time series regression for the evaluation of public health interventions: a tutorial
Bernal, James Lopez; Cummins, Steven; Gasparrini, Antonio
2017-01-01
Abstract Interrupted time series (ITS) analysis is a valuable study design for evaluating the effectiveness of population-level health interventions that have been implemented at a clearly defined point in time. It is increasingly being used to evaluate the effectiveness of interventions ranging from clinical therapy to national public health legislation. Whereas the design shares many properties of regression-based approaches in other epidemiological studies, there are a range of unique features of time series data that require additional methodological considerations. In this tutorial we use a worked example to demonstrate a robust approach to ITS analysis using segmented regression. We begin by describing the design and considering when ITS is an appropriate design choice. We then discuss the essential, yet often omitted, step of proposing the impact model a priori. Subsequently, we demonstrate the approach to statistical analysis including the main segmented regression model. Finally we describe the main methodological issues associated with ITS analysis: over-dispersion of time series data, autocorrelation, adjusting for seasonal trends and controlling for time-varying confounders, and we also outline some of the more complex design adaptations that can be used to strengthen the basic ITS design. PMID:27283160
A method for fitting regression splines with varying polynomial order in the linear mixed model.
Edwards, Lloyd J; Stewart, Paul W; MacDougall, James E; Helms, Ronald W
2006-02-15
The linear mixed model has become a widely used tool for longitudinal analysis of continuous variables. The use of regression splines in these models offers the analyst additional flexibility in the formulation of descriptive analyses, exploratory analyses and hypothesis-driven confirmatory analyses. We propose a method for fitting piecewise polynomial regression splines with varying polynomial order in the fixed effects and/or random effects of the linear mixed model. The polynomial segments are explicitly constrained by side conditions for continuity and some smoothness at the points where they join. By using a reparameterization of this explicitly constrained linear mixed model, an implicitly constrained linear mixed model is constructed that simplifies implementation of fixed-knot regression splines. The proposed approach is relatively simple, handles splines in one variable or multiple variables, and can be easily programmed using existing commercial software such as SAS or S-plus. The method is illustrated using two examples: an analysis of longitudinal viral load data from a study of subjects with acute HIV-1 infection and an analysis of 24-hour ambulatory blood pressure profiles.
Regression Analysis of Mixed Recurrent-Event and Panel-Count Data with Additive Rate Models
Zhu, Liang; Zhao, Hui; Sun, Jianguo; Leisenring, Wendy; Robison, Leslie L.
2015-01-01
Summary Event-history studies of recurrent events are often conducted in fields such as demography, epidemiology, medicine, and social sciences (Cook and Lawless, 2007; Zhao et al., 2011). For such analysis, two types of data have been extensively investigated: recurrent-event data and panel-count data. However, in practice, one may face a third type of data, mixed recurrent-event and panel-count data or mixed event-history data. Such data occur if some study subjects are monitored or observed continuously and thus provide recurrent-event data, while the others are observed only at discrete times and hence give only panel-count data. A more general situation is that each subject is observed continuously over certain time periods but only at discrete times over other time periods. There exists little literature on the analysis of such mixed data except that published by Zhu et al. (2013). In this paper, we consider the regression analysis of mixed data using the additive rate model and develop some estimating equation-based approaches to estimate the regression parameters of interest. Both finite sample and asymptotic properties of the resulting estimators are established, and the numerical studies suggest that the proposed methodology works well for practical situations. The approach is applied to a Childhood Cancer Survivor Study that motivated this study. PMID:25345405
Regression analysis of mixed recurrent-event and panel-count data with additive rate models.
Zhu, Liang; Zhao, Hui; Sun, Jianguo; Leisenring, Wendy; Robison, Leslie L
2015-03-01
Event-history studies of recurrent events are often conducted in fields such as demography, epidemiology, medicine, and social sciences (Cook and Lawless, 2007, The Statistical Analysis of Recurrent Events. New York: Springer-Verlag; Zhao et al., 2011, Test 20, 1-42). For such analysis, two types of data have been extensively investigated: recurrent-event data and panel-count data. However, in practice, one may face a third type of data, mixed recurrent-event and panel-count data or mixed event-history data. Such data occur if some study subjects are monitored or observed continuously and thus provide recurrent-event data, while the others are observed only at discrete times and hence give only panel-count data. A more general situation is that each subject is observed continuously over certain time periods but only at discrete times over other time periods. There exists little literature on the analysis of such mixed data except that published by Zhu et al. (2013, Statistics in Medicine 32, 1954-1963). In this article, we consider the regression analysis of mixed data using the additive rate model and develop some estimating equation-based approaches to estimate the regression parameters of interest. Both finite sample and asymptotic properties of the resulting estimators are established, and the numerical studies suggest that the proposed methodology works well for practical situations. The approach is applied to a Childhood Cancer Survivor Study that motivated this study. © 2014, The International Biometric Society.
Sperm Retrieval in Patients with Klinefelter Syndrome: A Skewed Regression Model Analysis.
Chehrazi, Mohammad; Rahimiforoushani, Abbas; Sabbaghian, Marjan; Nourijelyani, Keramat; Sadighi Gilani, Mohammad Ali; Hoseini, Mostafa; Vesali, Samira; Yaseri, Mehdi; Alizadeh, Ahad; Mohammad, Kazem; Samani, Reza Omani
2017-01-01
The most common chromosomal abnormality due to non-obstructive azoospermia (NOA) is Klinefelter syndrome (KS) which occurs in 1-1.72 out of 500-1000 male infants. The probability of retrieving sperm as the outcome could be asymmetrically different between patients with and without KS, therefore logistic regression analysis is not a well-qualified test for this type of data. This study has been designed to evaluate skewed regression model analysis for data collected from microsurgical testicular sperm extraction (micro-TESE) among azoospermic patients with and without non-mosaic KS syndrome. This cohort study compared the micro-TESE outcome between 134 men with classic KS and 537 men with NOA and normal karyotype who were referred to Royan Institute between 2009 and 2011. In addition to our main outcome, which was sperm retrieval, we also used logistic and skewed regression analyses to compare the following demographic and hormonal factors: age, level of follicle stimulating hormone (FSH), luteinizing hormone (LH), and testosterone between the two groups. A comparison of the micro-TESE between the KS and control groups showed a success rate of 28.4% (38/134) for the KS group and 22.2% (119/537) for the control group. In the KS group, a significantly difference (P<0.001) existed between testosterone levels for the successful sperm retrieval group (3.4 ± 0.48 mg/mL) compared to the unsuccessful sperm retrieval group (2.33 ± 0.23 mg/mL). The index for quasi Akaike information criterion (QAIC) had a goodness of fit of 74 for the skewed model which was lower than logistic regression (QAIC=85). According to the results, skewed regression is more efficient in estimating sperm retrieval success when the data from patients with KS are analyzed. This finding should be investigated by conducting additional studies with different data structures.
An Examination and Comparison of Airline and Navy Pilot Career Earnings
1986-03-01
RECEIVED ........ .............. 45 16. AIRLINE PILOT PROBATIONARY WAGES .... ........ 46 17. 1985 FAPA MAXIMUM PILOT WAGE ESTIMATES ..... 53 1 1983...tI% LIN PILOT WAGES REGRESSION EQUATIONS . 5 19. AVERAGE 1983 PILOT WAGES COMPUTED FROM REGRESSION ANALYSIS ...... ............. 56 20. FAPA MAXIMUM...Western N/A 1,200 1,500 Source: FAPA This establishes a wage "base" for pilots. In addition, a pilot who ilys more than average in one month may "bank
Application of artificial neural network to fMRI regression analysis.
Misaki, Masaya; Miyauchi, Satoru
2006-01-15
We used an artificial neural network (ANN) to detect correlations between event sequences and fMRI (functional magnetic resonance imaging) signals. The layered feed-forward neural network, given a series of events as inputs and the fMRI signal as a supervised signal, performed a non-linear regression analysis. This type of ANN is capable of approximating any continuous function, and thus this analysis method can detect any fMRI signals that correlated with corresponding events. Because of the flexible nature of ANNs, fitting to autocorrelation noise is a problem in fMRI analyses. We avoided this problem by using cross-validation and an early stopping procedure. The results showed that the ANN could detect various responses with different time courses. The simulation analysis also indicated an additional advantage of ANN over non-parametric methods in detecting parametrically modulated responses, i.e., it can detect various types of parametric modulations without a priori assumptions. The ANN regression analysis is therefore beneficial for exploratory fMRI analyses in detecting continuous changes in responses modulated by changes in input values.
Huang, Li-Shan; Myers, Gary J.; Davidson, Philip W.; Cox, Christopher; Xiao, Fenyuan; Thurston, Sally W.; Cernichiari, Elsa; Shamlaye, Conrad F.; Sloane-Reeves, Jean; Georger, Lesley; Clarkson, Thomas W.
2007-01-01
Studies of the association between prenatal methylmercury exposure from maternal fish consumption during pregnancy and neurodevelopmental test scores in the Seychelles Child Development Study have found no consistent pattern of associations through age nine years. The analyses for the most recent nine-year data examined the population effects of prenatal exposure, but did not address the possibility of non-homogeneous susceptibility. This paper presents a regression tree approach: covariate effects are treated nonlinearly and non-additively and non-homogeneous effects of prenatal methylmercury exposure are permitted among the covariate clusters identified by the regression tree. The approach allows us to address whether children in the lower or higher ends of the developmental spectrum differ in susceptibility to subtle exposure effects. Of twenty-one endpoints available at age nine years, we chose the Weschler Full Scale IQ and its associated covariates to construct the regression tree. The prenatal mercury effect in each of the nine resulting clusters was assessed linearly and non-homogeneously. In addition we reanalyzed five other nine-year endpoints that in the linear analysis has a two-tailed p-value <0.2 for the effect of prenatal exposure. In this analysis, motor proficiency and activity level improved significantly with increasing MeHg for 53% of the children who had an average home environment. Motor proficiency significantly decreased with increasing prenatal MeHg exposure in 7% of the children whose home environment was below average. The regression tree results support previous analyses of outcomes in this cohort. However, this analysis raises the intriguing possibility that an effect may be non-homogeneous among children with different backgrounds and IQ levels. PMID:17942158
Huang, Li-Shan; Myers, Gary J; Davidson, Philip W; Cox, Christopher; Xiao, Fenyuan; Thurston, Sally W; Cernichiari, Elsa; Shamlaye, Conrad F; Sloane-Reeves, Jean; Georger, Lesley; Clarkson, Thomas W
2007-11-01
Studies of the association between prenatal methylmercury exposure from maternal fish consumption during pregnancy and neurodevelopmental test scores in the Seychelles Child Development Study have found no consistent pattern of associations through age 9 years. The analyses for the most recent 9-year data examined the population effects of prenatal exposure, but did not address the possibility of non-homogeneous susceptibility. This paper presents a regression tree approach: covariate effects are treated non-linearly and non-additively and non-homogeneous effects of prenatal methylmercury exposure are permitted among the covariate clusters identified by the regression tree. The approach allows us to address whether children in the lower or higher ends of the developmental spectrum differ in susceptibility to subtle exposure effects. Of 21 endpoints available at age 9 years, we chose the Weschler Full Scale IQ and its associated covariates to construct the regression tree. The prenatal mercury effect in each of the nine resulting clusters was assessed linearly and non-homogeneously. In addition we reanalyzed five other 9-year endpoints that in the linear analysis had a two-tailed p-value <0.2 for the effect of prenatal exposure. In this analysis, motor proficiency and activity level improved significantly with increasing MeHg for 53% of the children who had an average home environment. Motor proficiency significantly decreased with increasing prenatal MeHg exposure in 7% of the children whose home environment was below average. The regression tree results support previous analyses of outcomes in this cohort. However, this analysis raises the intriguing possibility that an effect may be non-homogeneous among children with different backgrounds and IQ levels.
Baldi, F; Alencar, M M; Albuquerque, L G
2010-12-01
The objective of this work was to estimate covariance functions using random regression models on B-splines functions of animal age, for weights from birth to adult age in Canchim cattle. Data comprised 49,011 records on 2435 females. The model of analysis included fixed effects of contemporary groups, age of dam as quadratic covariable and the population mean trend taken into account by a cubic regression on orthogonal polynomials of animal age. Residual variances were modelled through a step function with four classes. The direct and maternal additive genetic effects, and animal and maternal permanent environmental effects were included as random effects in the model. A total of seventeen analyses, considering linear, quadratic and cubic B-splines functions and up to seven knots, were carried out. B-spline functions of the same order were considered for all random effects. Random regression models on B-splines functions were compared to a random regression model on Legendre polynomials and with a multitrait model. Results from different models of analyses were compared using the REML form of the Akaike Information criterion and Schwarz' Bayesian Information criterion. In addition, the variance components and genetic parameters estimated for each random regression model were also used as criteria to choose the most adequate model to describe the covariance structure of the data. A model fitting quadratic B-splines, with four knots or three segments for direct additive genetic effect and animal permanent environmental effect and two knots for maternal additive genetic effect and maternal permanent environmental effect, was the most adequate to describe the covariance structure of the data. Random regression models using B-spline functions as base functions fitted the data better than Legendre polynomials, especially at mature ages, but higher number of parameters need to be estimated with B-splines functions. © 2010 Blackwell Verlag GmbH.
Chen, Ling; Feng, Yanqin; Sun, Jianguo
2017-10-01
This paper discusses regression analysis of clustered failure time data, which occur when the failure times of interest are collected from clusters. In particular, we consider the situation where the correlated failure times of interest may be related to cluster sizes. For inference, we present two estimation procedures, the weighted estimating equation-based method and the within-cluster resampling-based method, when the correlated failure times of interest arise from a class of additive transformation models. The former makes use of the inverse of cluster sizes as weights in the estimating equations, while the latter can be easily implemented by using the existing software packages for right-censored failure time data. An extensive simulation study is conducted and indicates that the proposed approaches work well in both the situations with and without informative cluster size. They are applied to a dental study that motivated this study.
Optimizing methods for linking cinematic features to fMRI data.
Kauttonen, Janne; Hlushchuk, Yevhen; Tikka, Pia
2015-04-15
One of the challenges of naturalistic neurosciences using movie-viewing experiments is how to interpret observed brain activations in relation to the multiplicity of time-locked stimulus features. As previous studies have shown less inter-subject synchronization across viewers of random video footage than story-driven films, new methods need to be developed for analysis of less story-driven contents. To optimize the linkage between our fMRI data collected during viewing of a deliberately non-narrative silent film 'At Land' by Maya Deren (1944) and its annotated content, we combined the method of elastic-net regularization with the model-driven linear regression and the well-established data-driven independent component analysis (ICA) and inter-subject correlation (ISC) methods. In the linear regression analysis, both IC and region-of-interest (ROI) time-series were fitted with time-series of a total of 36 binary-valued and one real-valued tactile annotation of film features. The elastic-net regularization and cross-validation were applied in the ordinary least-squares linear regression in order to avoid over-fitting due to the multicollinearity of regressors, the results were compared against both the partial least-squares (PLS) regression and the un-regularized full-model regression. Non-parametric permutation testing scheme was applied to evaluate the statistical significance of regression. We found statistically significant correlation between the annotation model and 9 ICs out of 40 ICs. Regression analysis was also repeated for a large set of cubic ROIs covering the grey matter. Both IC- and ROI-based regression analyses revealed activations in parietal and occipital regions, with additional smaller clusters in the frontal lobe. Furthermore, we found elastic-net based regression more sensitive than PLS and un-regularized regression since it detected a larger number of significant ICs and ROIs. Along with the ISC ranking methods, our regression analysis proved a feasible method for ordering the ICs based on their functional relevance to the annotated cinematic features. The novelty of our method is - in comparison to the hypothesis-driven manual pre-selection and observation of some individual regressors biased by choice - in applying data-driven approach to all content features simultaneously. We found especially the combination of regularized regression and ICA useful when analyzing fMRI data obtained using non-narrative movie stimulus with a large set of complex and correlated features. Copyright © 2015. Published by Elsevier Inc.
Tukiendorf, Andrzej; Mansournia, Mohammad Ali; Wydmański, Jerzy; Wolny-Rokicka, Edyta
2017-04-01
Background: Clinical datasets for epithelial ovarian cancer brain metastatic patients are usually small in size. When adequate case numbers are lacking, resulting estimates of regression coefficients may demonstrate bias. One of the direct approaches to reduce such sparse-data bias is based on penalized estimation. Methods: A re- analysis of formerly reported hazard ratios in diagnosed patients was performed using penalized Cox regression with a popular SAS package providing additional software codes for a statistical computational procedure. Results: It was found that the penalized approach can readily diminish sparse data artefacts and radically reduce the magnitude of estimated regression coefficients. Conclusions: It was confirmed that classical statistical approaches may exaggerate regression estimates or distort study interpretations and conclusions. The results support the thesis that penalization via weak informative priors and data augmentation are the safest approaches to shrink sparse data artefacts frequently occurring in epidemiological research. Creative Commons Attribution License
NASA Technical Reports Server (NTRS)
Waller, M. C.
1976-01-01
An electro-optical device called an oculometer which tracks a subject's lookpoint as a time function has been used to collect data in a real-time simulation study of instrument landing system (ILS) approaches. The data describing the scanning behavior of a pilot during the instrument approaches have been analyzed by use of a stepwise regression analysis technique. A statistically significant correlation between pilot workload, as indicated by pilot ratings, and scanning behavior has been established. In addition, it was demonstrated that parameters derived from the scanning behavior data can be combined in a mathematical equation to provide a good representation of pilot workload.
Olson, Scott A.
2003-01-01
The stream-gaging network in New Hampshire was analyzed for its effectiveness in providing regional information on peak-flood flow, mean-flow, and low-flow frequency. The data available for analysis were from stream-gaging stations in New Hampshire and selected stations in adjacent States. The principles of generalized-least-squares regression analysis were applied to develop regional regression equations that relate streamflow-frequency characteristics to watershed characteristics. Regression equations were developed for (1) the instantaneous peak flow with a 100-year recurrence interval, (2) the mean-annual flow, and (3) the 7-day, 10-year low flow. Active and discontinued stream-gaging stations with 10 or more years of flow data were used to develop the regression equations. Each stream-gaging station in the network was evaluated and ranked on the basis of how much the data from that station contributed to the cost-weighted sampling-error component of the regression equation. The potential effect of data from proposed and new stream-gaging stations on the sampling error also was evaluated. The stream-gaging network was evaluated for conditions in water year 2000 and for estimated conditions under various network strategies if an additional 5 years and 20 years of streamflow data were collected. The effectiveness of the stream-gaging network in providing regional streamflow information could be improved for all three flow characteristics with the collection of additional flow data, both temporally and spatially. With additional years of data collection, the greatest reduction in the average sampling error of the regional regression equations was found for the peak- and low-flow characteristics. In general, additional data collection at stream-gaging stations with unregulated flow, relatively short-term record (less than 20 years), and drainage areas smaller than 45 square miles contributed the largest cost-weighted reduction to the average sampling error of the regional estimating equations. The results of the network analyses can be used to prioritize the continued operation of active stations, the reactivation of discontinued stations, or the activation of new stations to maximize the regional information content provided by the stream-gaging network. Final decisions regarding altering the New Hampshire stream-gaging network would require the consideration of the many uses of the streamflow data serving local, State, and Federal interests.
Ghosh, Sudipta; Dosaev, Tasbulat; Prakash, Jai; Livshits, Gregory
2017-04-01
The major aim of this study was to conduct comparative quantitative-genetic analysis of the body composition (BCP) and somatotype (STP) variation, as well as their correlations with blood pressure (BP) in two ethnically, culturally and geographically different populations: Santhal, indigenous ethnic group from India and Chuvash, indigenous population from Russia. Correspondently two pedigree-based samples were collected from 1,262 Santhal and1,558 Chuvash individuals, respectively. At the first stage of the study, descriptive statistics and a series of univariate regression analyses were calculated. Finally, multiple and multivariate regression (MMR) analyses, with BP measurements as dependent variables and age, sex, BCP and STP as independent variables were carried out in each sample separately. The significant and independent covariates of BP were identified and used for re-examination in pedigree-based variance decomposition analysis. Despite clear and significant differences between the populations in BCP/STP, both Santhal and Chuvash were found to be predominantly mesomorphic irrespective of their sex. According to MMR analyses variation of BP significantly depended on age and mesomorphic component in both samples, and in addition on sex, ectomorphy and fat mass index in Santhal and on fat free mass index in Chuvash samples, respectively. Additive genetic component contributes to a substantial proportion of blood pressure and body composition variance. Variance component analysis in addition to above mentioned results suggests that additive genetic factors influence BP and BCP/STP associations significantly. © 2017 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Abunama, Taher; Othman, Faridah
2017-06-01
Analysing the fluctuations of wastewater inflow rates in sewage treatment plants (STPs) is essential to guarantee a sufficient treatment of wastewater before discharging it to the environment. The main objectives of this study are to statistically analyze and forecast the wastewater inflow rates into the Bandar Tun Razak STP in Kuala Lumpur, Malaysia. A time series analysis of three years’ weekly influent data (156weeks) has been conducted using the Auto-Regressive Integrated Moving Average (ARIMA) model. Various combinations of ARIMA orders (p, d, q) have been tried to select the most fitted model, which was utilized to forecast the wastewater inflow rates. The linear regression analysis was applied to testify the correlation between the observed and predicted influents. ARIMA (3, 1, 3) model was selected with the highest significance R-square and lowest normalized Bayesian Information Criterion (BIC) value, and accordingly the wastewater inflow rates were forecasted to additional 52weeks. The linear regression analysis between the observed and predicted values of the wastewater inflow rates showed a positive linear correlation with a coefficient of 0.831.
Dinç, Erdal; Ozdemir, Abdil
2005-01-01
Multivariate chromatographic calibration technique was developed for the quantitative analysis of binary mixtures enalapril maleate (EA) and hydrochlorothiazide (HCT) in tablets in the presence of losartan potassium (LST). The mathematical algorithm of multivariate chromatographic calibration technique is based on the use of the linear regression equations constructed using relationship between concentration and peak area at the five-wavelength set. The algorithm of this mathematical calibration model having a simple mathematical content was briefly described. This approach is a powerful mathematical tool for an optimum chromatographic multivariate calibration and elimination of fluctuations coming from instrumental and experimental conditions. This multivariate chromatographic calibration contains reduction of multivariate linear regression functions to univariate data set. The validation of model was carried out by analyzing various synthetic binary mixtures and using the standard addition technique. Developed calibration technique was applied to the analysis of the real pharmaceutical tablets containing EA and HCT. The obtained results were compared with those obtained by classical HPLC method. It was observed that the proposed multivariate chromatographic calibration gives better results than classical HPLC.
MAGMA: Generalized Gene-Set Analysis of GWAS Data
de Leeuw, Christiaan A.; Mooij, Joris M.; Heskes, Tom; Posthuma, Danielle
2015-01-01
By aggregating data for complex traits in a biologically meaningful way, gene and gene-set analysis constitute a valuable addition to single-marker analysis. However, although various methods for gene and gene-set analysis currently exist, they generally suffer from a number of issues. Statistical power for most methods is strongly affected by linkage disequilibrium between markers, multi-marker associations are often hard to detect, and the reliance on permutation to compute p-values tends to make the analysis computationally very expensive. To address these issues we have developed MAGMA, a novel tool for gene and gene-set analysis. The gene analysis is based on a multiple regression model, to provide better statistical performance. The gene-set analysis is built as a separate layer around the gene analysis for additional flexibility. This gene-set analysis also uses a regression structure to allow generalization to analysis of continuous properties of genes and simultaneous analysis of multiple gene sets and other gene properties. Simulations and an analysis of Crohn’s Disease data are used to evaluate the performance of MAGMA and to compare it to a number of other gene and gene-set analysis tools. The results show that MAGMA has significantly more power than other tools for both the gene and the gene-set analysis, identifying more genes and gene sets associated with Crohn’s Disease while maintaining a correct type 1 error rate. Moreover, the MAGMA analysis of the Crohn’s Disease data was found to be considerably faster as well. PMID:25885710
MAGMA: generalized gene-set analysis of GWAS data.
de Leeuw, Christiaan A; Mooij, Joris M; Heskes, Tom; Posthuma, Danielle
2015-04-01
By aggregating data for complex traits in a biologically meaningful way, gene and gene-set analysis constitute a valuable addition to single-marker analysis. However, although various methods for gene and gene-set analysis currently exist, they generally suffer from a number of issues. Statistical power for most methods is strongly affected by linkage disequilibrium between markers, multi-marker associations are often hard to detect, and the reliance on permutation to compute p-values tends to make the analysis computationally very expensive. To address these issues we have developed MAGMA, a novel tool for gene and gene-set analysis. The gene analysis is based on a multiple regression model, to provide better statistical performance. The gene-set analysis is built as a separate layer around the gene analysis for additional flexibility. This gene-set analysis also uses a regression structure to allow generalization to analysis of continuous properties of genes and simultaneous analysis of multiple gene sets and other gene properties. Simulations and an analysis of Crohn's Disease data are used to evaluate the performance of MAGMA and to compare it to a number of other gene and gene-set analysis tools. The results show that MAGMA has significantly more power than other tools for both the gene and the gene-set analysis, identifying more genes and gene sets associated with Crohn's Disease while maintaining a correct type 1 error rate. Moreover, the MAGMA analysis of the Crohn's Disease data was found to be considerably faster as well.
Regression and multivariate models for predicting particulate matter concentration level.
Nazif, Amina; Mohammed, Nurul Izma; Malakahmad, Amirhossein; Abualqumboz, Motasem S
2018-01-01
The devastating health effects of particulate matter (PM 10 ) exposure by susceptible populace has made it necessary to evaluate PM 10 pollution. Meteorological parameters and seasonal variation increases PM 10 concentration levels, especially in areas that have multiple anthropogenic activities. Hence, stepwise regression (SR), multiple linear regression (MLR) and principal component regression (PCR) analyses were used to analyse daily average PM 10 concentration levels. The analyses were carried out using daily average PM 10 concentration, temperature, humidity, wind speed and wind direction data from 2006 to 2010. The data was from an industrial air quality monitoring station in Malaysia. The SR analysis established that meteorological parameters had less influence on PM 10 concentration levels having coefficient of determination (R 2 ) result from 23 to 29% based on seasoned and unseasoned analysis. While, the result of the prediction analysis showed that PCR models had a better R 2 result than MLR methods. The results for the analyses based on both seasoned and unseasoned data established that MLR models had R 2 result from 0.50 to 0.60. While, PCR models had R 2 result from 0.66 to 0.89. In addition, the validation analysis using 2016 data also recognised that the PCR model outperformed the MLR model, with the PCR model for the seasoned analysis having the best result. These analyses will aid in achieving sustainable air quality management strategies.
Lo, Benjamin W Y; Fukuda, Hitoshi; Angle, Mark; Teitelbaum, Jeanne; Macdonald, R Loch; Farrokhyar, Forough; Thabane, Lehana; Levine, Mitchell A H
2016-01-01
Classification and regression tree analysis involves the creation of a decision tree by recursive partitioning of a dataset into more homogeneous subgroups. Thus far, there is scarce literature on using this technique to create clinical prediction tools for aneurysmal subarachnoid hemorrhage (SAH). The classification and regression tree analysis technique was applied to the multicenter Tirilazad database (3551 patients) in order to create the decision-making algorithm. In order to elucidate prognostic subgroups in aneurysmal SAH, neurologic, systemic, and demographic factors were taken into account. The dependent variable used for analysis was the dichotomized Glasgow Outcome Score at 3 months. Classification and regression tree analysis revealed seven prognostic subgroups. Neurological grade, occurrence of post-admission stroke, occurrence of post-admission fever, and age represented the explanatory nodes of this decision tree. Split sample validation revealed classification accuracy of 79% for the training dataset and 77% for the testing dataset. In addition, the occurrence of fever at 1-week post-aneurysmal SAH is associated with increased odds of post-admission stroke (odds ratio: 1.83, 95% confidence interval: 1.56-2.45, P < 0.01). A clinically useful classification tree was generated, which serves as a prediction tool to guide bedside prognostication and clinical treatment decision making. This prognostic decision-making algorithm also shed light on the complex interactions between a number of risk factors in determining outcome after aneurysmal SAH.
Totton, Sarah C; Farrar, Ashley M; Wilkins, Wendy; Bucher, Oliver; Waddell, Lisa A; Wilhelm, Barbara J; McEwen, Scott A; Rajić, Andrijana
2012-10-01
Eating inappropriately prepared poultry meat is a major cause of foodborne salmonellosis. Our objectives were to determine the efficacy of feed and water additives (other than competitive exclusion and antimicrobials) on reducing Salmonella prevalence or concentration in broiler chickens using systematic review-meta-analysis and to explore sources of heterogeneity found in the meta-analysis through meta-regression. Six electronic databases were searched (Current Contents (1999-2009), Agricola (1924-2009), MEDLINE (1860-2009), Scopus (1960-2009), Centre for Agricultural Bioscience (CAB) (1913-2009), and CAB Global Health (1971-2009)), five topic experts were contacted, and the bibliographies of review articles and a topic-relevant textbook were manually searched to identify all relevant research. Study inclusion criteria comprised: English-language primary research investigating the effects of feed and water additives on the Salmonella prevalence or concentration in broiler chickens. Data extraction and study methodological assessment were conducted by two reviewers independently using pretested forms. Seventy challenge studies (n=910 unique treatment-control comparisons), seven controlled studies (n=154), and one quasi-experiment (n=1) met the inclusion criteria. Compared to an assumed control group prevalence of 44 of 1000 broilers, random-effects meta-analysis indicated that the Salmonella cecal colonization in groups with prebiotics (fructooligosaccharide, lactose, whey, dried milk, lactulose, lactosucrose, sucrose, maltose, mannanoligosaccharide) added to feed or water was 15 out of 1000 broilers; with lactose added to feed or water it was 10 out of 1000 broilers; with experimental chlorate product (ECP) added to feed or water it was 21 out of 1000. For ECP the concentration of Salmonella in the ceca was decreased by 0.61 log(10)cfu/g in the treated group compared to the control group. Significant heterogeneity (Cochran's Q-statistic p≤0.10) was observed among studies examining all organic acids (controlled or challenge experiments), butyric acid, formic acid, a formic/propionic acid mixture, fermented liquid feed, and D-mannose. Meta-regressions were conducted to examine the source of heterogeneity among studies. For prevalence outcomes, 36% and 60% of the total variance was within and between studies, respectively. For concentration outcomes, 39% and 33% of the total variance was within and between studies, respectively. Inadequate blinding and randomization was common, and no studies undergoing meta-analysis or meta-regression were conducted on a commercial farm. The strength of evidence of the effect of these additives was very low. Studies conducted under commercial conditions are needed to understand the potential benefit of these interventions for the poultry industry and to improve the strength of evidence of the effectiveness of these additives. Copyright © 2012 Elsevier B.V. All rights reserved.
Kinoshita, Shoji; Kakuda, Wataru; Momosaki, Ryo; Yamada, Naoki; Sugawara, Hidekazu; Watanabe, Shu; Abo, Masahiro
2015-05-01
Early rehabilitation for acute stroke patients is widely recommended. We tested the hypothesis that clinical outcome of stroke patients who receive early rehabilitation managed by board-certificated physiatrists (BCP) is generally better than that provided by other medical specialties. Data of stroke patients who underwent early rehabilitation in 19 acute hospitals between January 2005 and December 2013 were collected from the Japan Rehabilitation Database and analyzed retrospectively. Multivariate linear regression analysis using generalized estimating equations method was performed to assess the association between Functional Independence Measure (FIM) effectiveness and management provided by BCP in early rehabilitation. In addition, multivariate logistic regression analysis was also performed to assess the impact of management provided by BCP in acute phase on discharge destination. After setting the inclusion criteria, data of 3838 stroke patients were eligible for analysis. BCP provided early rehabilitation in 814 patients (21.2%). Both the duration of daily exercise time and the frequency of regular conferencing were significantly higher for patients managed by BCP than by other specialties. Although the mortality rate was not different, multivariate regression analysis showed that FIM effectiveness correlated significantly and positively with the management provided by BCP (coefficient, .35; 95% confidence interval [CI], .012-.059; P < .005). In addition, multivariate logistic analysis identified clinical management by BCP as a significant determinant of home discharge (odds ratio, 1.24; 95% CI, 1.08-1.44; P < .005). Our retrospective cohort study demonstrated that clinical management provided by BCP in early rehabilitation can lead to functional recovery of acute stroke. Copyright © 2015 National Stroke Association. Published by Elsevier Inc. All rights reserved.
Magnitude and Frequency of Floods for Urban and Small Rural Streams in Georgia, 2008
Gotvald, Anthony J.; Knaak, Andrew E.
2011-01-01
A study was conducted that updated methods for estimating the magnitude and frequency of floods in ungaged urban basins in Georgia that are not substantially affected by regulation or tidal fluctuations. Annual peak-flow data for urban streams from September 2008 were analyzed for 50 streamgaging stations (streamgages) in Georgia and 6 streamgages on adjacent urban streams in Florida and South Carolina having 10 or more years of data. Flood-frequency estimates were computed for the 56 urban streamgages by fitting logarithms of annual peak flows for each streamgage to a Pearson Type III distribution. Additionally, basin characteristics for the streamgages were computed by using a geographical information system and computer algorithms. Regional regression analysis, using generalized least-squares regression, was used to develop a set of equations for estimating flows with 50-, 20-, 10-, 4-, 2-, 1-, 0.5-, and 0.2-percent annual exceedance probabilities for ungaged urban basins in Georgia. In addition to the 56 urban streamgages, 171 rural streamgages were included in the regression analysis to maintain continuity between flood estimates for urban and rural basins as the basin characteristics pertaining to urbanization approach zero. Because 21 of the rural streamgages have drainage areas less than 1 square mile, the set of equations developed for this study can also be used for estimating small ungaged rural streams in Georgia. Flood-frequency estimates and basin characteristics for 227 streamgages were combined to form the final database used in the regional regression analysis. Four hydrologic regions were developed for Georgia. The final equations are functions of drainage area and percentage of impervious area for three of the regions and drainage area, percentage of developed land, and mean basin slope for the fourth region. Average standard errors of prediction for these regression equations range from 20.0 to 74.5 percent.
Multiple Imputation of a Randomly Censored Covariate Improves Logistic Regression Analysis.
Atem, Folefac D; Qian, Jing; Maye, Jacqueline E; Johnson, Keith A; Betensky, Rebecca A
2016-01-01
Randomly censored covariates arise frequently in epidemiologic studies. The most commonly used methods, including complete case and single imputation or substitution, suffer from inefficiency and bias. They make strong parametric assumptions or they consider limit of detection censoring only. We employ multiple imputation, in conjunction with semi-parametric modeling of the censored covariate, to overcome these shortcomings and to facilitate robust estimation. We develop a multiple imputation approach for randomly censored covariates within the framework of a logistic regression model. We use the non-parametric estimate of the covariate distribution or the semiparametric Cox model estimate in the presence of additional covariates in the model. We evaluate this procedure in simulations, and compare its operating characteristics to those from the complete case analysis and a survival regression approach. We apply the procedures to an Alzheimer's study of the association between amyloid positivity and maternal age of onset of dementia. Multiple imputation achieves lower standard errors and higher power than the complete case approach under heavy and moderate censoring and is comparable under light censoring. The survival regression approach achieves the highest power among all procedures, but does not produce interpretable estimates of association. Multiple imputation offers a favorable alternative to complete case analysis and ad hoc substitution methods in the presence of randomly censored covariates within the framework of logistic regression.
Selapa, N W; Nephawe, K A; Maiwashe, A; Norris, D
2012-02-08
The aim of this study was to estimate genetic parameters for body weights of individually fed beef bulls measured at centralized testing stations in South Africa using random regression models. Weekly body weights of Bonsmara bulls (N = 2919) tested between 1999 and 2003 were available for the analyses. The model included a fixed regression of the body weights on fourth-order orthogonal Legendre polynomials of the actual days on test (7, 14, 21, 28, 35, 42, 49, 56, 63, 70, 77, and 84) for starting age and contemporary group effects. Random regressions on fourth-order orthogonal Legendre polynomials of the actual days on test were included for additive genetic effects and additional uncorrelated random effects of the weaning-herd-year and the permanent environment of the animal. Residual effects were assumed to be independently distributed with heterogeneous variance for each test day. Variance ratios for additive genetic, permanent environment and weaning-herd-year for weekly body weights at different test days ranged from 0.26 to 0.29, 0.37 to 0.44 and 0.26 to 0.34, respectively. The weaning-herd-year was found to have a significant effect on the variation of body weights of bulls despite a 28-day adjustment period. Genetic correlations amongst body weights at different test days were high, ranging from 0.89 to 1.00. Heritability estimates were comparable to literature using multivariate models. Therefore, random regression model could be applied in the genetic evaluation of body weight of individually fed beef bulls in South Africa.
Lamont, Andrea E.; Vermunt, Jeroen K.; Van Horn, M. Lee
2016-01-01
Regression mixture models are increasingly used as an exploratory approach to identify heterogeneity in the effects of a predictor on an outcome. In this simulation study, we test the effects of violating an implicit assumption often made in these models – i.e., independent variables in the model are not directly related to latent classes. Results indicated that the major risk of failing to model the relationship between predictor and latent class was an increase in the probability of selecting additional latent classes and biased class proportions. Additionally, this study tests whether regression mixture models can detect a piecewise relationship between a predictor and outcome. Results suggest that these models are able to detect piecewise relations, but only when the relationship between the latent class and the predictor is included in model estimation. We illustrate the implications of making this assumption through a re-analysis of applied data examining heterogeneity in the effects of family resources on academic achievement. We compare previous results (which assumed no relation between independent variables and latent class) to the model where this assumption is lifted. Implications and analytic suggestions for conducting regression mixture based on these findings are noted. PMID:26881956
Indicators of suboptimal performance embedded in the Wechsler Memory Scale-Fourth Edition (WMS-IV).
Bouman, Zita; Hendriks, Marc P H; Schmand, Ben A; Kessels, Roy P C; Aldenkamp, Albert P
2016-01-01
Recognition and visual working memory tasks from the Wechsler Memory Scale-Fourth Edition (WMS-IV) have previously been documented as useful indicators for suboptimal performance. The present study examined the clinical utility of the Dutch version of the WMS-IV (WMS-IV-NL) for the identification of suboptimal performance using an analogue study design. The patient group consisted of 59 mixed-etiology patients; the experimental malingerers were 50 healthy individuals who were asked to simulate cognitive impairment as a result of a traumatic brain injury; the last group consisted of 50 healthy controls who were instructed to put forth full effort. Experimental malingerers performed significantly lower on all WMS-IV-NL tasks than did the patients and healthy controls. A binary logistic regression analysis was performed on the experimental malingerers and the patients. The first model contained the visual working memory subtests (Spatial Addition and Symbol Span) and the recognition tasks of the following subtests: Logical Memory, Verbal Paired Associates, Designs, Visual Reproduction. The results showed an overall classification rate of 78.4%, and only Spatial Addition explained a significant amount of variation (p < .001). Subsequent logistic regression analysis and receiver operating characteristic (ROC) analysis supported the discriminatory power of the subtest Spatial Addition. A scaled score cutoff of <4 produced 93% specificity and 52% sensitivity for detection of suboptimal performance. The WMS-IV-NL Spatial Addition subtest may provide clinically useful information for the detection of suboptimal performance.
Van Houtven, George; Powers, John; Jessup, Amber; Yang, Jui-Chen
2006-08-01
Many economists argue that willingness-to-pay (WTP) measures are most appropriate for assessing the welfare effects of health changes. Nevertheless, the health evaluation literature is still dominated by studies estimating nonmonetary health status measures (HSMs), which are often used to assess changes in quality-adjusted life years (QALYs). Using meta-regression analysis, this paper combines results from both WTP and HSM studies applied to acute morbidity, and it tests whether a systematic relationship exists between HSM and WTP estimates. We analyze over 230 WTP estimates from 17 different studies and find evidence that QALY-based estimates of illness severity--as measured by the Quality of Well-Being (QWB) Scale--are significant factors in explaining variation in WTP, as are changes in the duration of illness and the average income and age of the study populations. In addition, we test and reject the assumption of a constant WTP per QALY gain. We also demonstrate how the estimated meta-regression equations can serve as benefit transfer functions for policy analysis. By specifying the change in duration and severity of the acute illness and the characteristics of the affected population, we apply the regression functions to predict average WTP per case avoided. Copyright 2006 John Wiley & Sons, Ltd.
Application of linear regression analysis in accuracy assessment of rolling force calculations
NASA Astrophysics Data System (ADS)
Poliak, E. I.; Shim, M. K.; Kim, G. S.; Choo, W. Y.
1998-10-01
Efficient operation of the computational models employed in process control systems require periodical assessment of the accuracy of their predictions. Linear regression is proposed as a tool which allows separate systematic and random prediction errors from those related to measurements. A quantitative characteristic of the model predictive ability is introduced in addition to standard statistical tests for model adequacy. Rolling force calculations are considered as an example for the application. However, the outlined approach can be used to assess the performance of any computational model.
Do Business Communication Courses Improve Student Performance in Introductory Marketing?
ERIC Educational Resources Information Center
Marcal, Leah E.; Hennessey, Judith E.; Curren, Mary T.; Roberts, William W.
2005-01-01
In this study, the authors investigated whether completion of a business communications course improved student performance in an introductory marketing management course. Regression analysis indicated that students who completed the communications course received higher grades than the otherwise comparable students. In addition, marketing majors…
Korany, Mohamed A; Gazy, Azza A; Khamis, Essam F; Ragab, Marwa A A; Kamal, Miranda F
2018-06-01
This study outlines two robust regression approaches, namely least median of squares (LMS) and iteratively re-weighted least squares (IRLS) to investigate their application in instrument analysis of nutraceuticals (that is, fluorescence quenching of merbromin reagent upon lipoic acid addition). These robust regression methods were used to calculate calibration data from the fluorescence quenching reaction (∆F and F-ratio) under ideal or non-ideal linearity conditions. For each condition, data were treated using three regression fittings: Ordinary Least Squares (OLS), LMS and IRLS. Assessment of linearity, limits of detection (LOD) and quantitation (LOQ), accuracy and precision were carefully studied for each condition. LMS and IRLS regression line fittings showed significant improvement in correlation coefficients and all regression parameters for both methods and both conditions. In the ideal linearity condition, the intercept and slope changed insignificantly, but a dramatic change was observed for the non-ideal condition and linearity intercept. Under both linearity conditions, LOD and LOQ values after the robust regression line fitting of data were lower than those obtained before data treatment. The results obtained after statistical treatment indicated that the linearity ranges for drug determination could be expanded to lower limits of quantitation by enhancing the regression equation parameters after data treatment. Analysis results for lipoic acid in capsules, using both fluorimetric methods, treated by parametric OLS and after treatment by robust LMS and IRLS were compared for both linearity conditions. Copyright © 2018 John Wiley & Sons, Ltd.
Predictors of condom use and refusal among the population of Free State province in South Africa
2012-01-01
Background This study investigated the extent and predictors of condom use and condom refusal in the Free State province in South Africa. Methods Through a household survey conducted in the Free Sate province of South Africa, 5,837 adults were interviewed. Univariate and multivariate survey logistic regressions and classification trees (CT) were used for analysing two response variables ‘ever used condom’ and ‘ever refused condom’. Results Eighty-three per cent of the respondents had ever used condoms, of which 38% always used them; 61% used them during the last sexual intercourse and 9% had ever refused to use them. The univariate logistic regression models and CT analysis indicated that a strong predictor of condom use was its perceived need. In the CT analysis, this variable was followed in importance by ‘knowledge of correct use of condom’, condom availability, young age, being single and higher education. ‘Perceived need’ for condoms did not remain significant in the multivariate analysis after controlling for other variables. The strongest predictor of condom refusal, as shown by the CT, was shame associated with condoms followed by the presence of sexual risk behaviour, knowing one’s HIV status, older age and lacking knowledge of condoms (i.e., ability to prevent sexually transmitted diseases and pregnancy, availability, correct and consistent use and existence of female condoms). In the multivariate logistic regression, age was not significant for condom refusal while affordability and perceived need were additional significant variables. Conclusions The use of complementary modelling techniques such as CT in addition to logistic regressions adds to a better understanding of condom use and refusal. Further improvement in correct and consistent use of condoms will require targeted interventions. In addition to existing social marketing campaigns, tailored approaches should focus on establishing the perceived need for condom-use and improving skills for correct use. They should also incorporate interventions to reduce the shame associated with condoms and individual counselling of those likely to refuse condoms. PMID:22639964
Procedures for adjusting regional regression models of urban-runoff quality using local data
Hoos, A.B.; Sisolak, J.K.
1993-01-01
Statistical operations termed model-adjustment procedures (MAP?s) can be used to incorporate local data into existing regression models to improve the prediction of urban-runoff quality. Each MAP is a form of regression analysis in which the local data base is used as a calibration data set. Regression coefficients are determined from the local data base, and the resulting `adjusted? regression models can then be used to predict storm-runoff quality at unmonitored sites. The response variable in the regression analyses is the observed load or mean concentration of a constituent in storm runoff for a single storm. The set of explanatory variables used in the regression analyses is different for each MAP, but always includes the predicted value of load or mean concentration from a regional regression model. The four MAP?s examined in this study were: single-factor regression against the regional model prediction, P, (termed MAP-lF-P), regression against P,, (termed MAP-R-P), regression against P, and additional local variables (termed MAP-R-P+nV), and a weighted combination of P, and a local-regression prediction (termed MAP-W). The procedures were tested by means of split-sample analysis, using data from three cities included in the Nationwide Urban Runoff Program: Denver, Colorado; Bellevue, Washington; and Knoxville, Tennessee. The MAP that provided the greatest predictive accuracy for the verification data set differed among the three test data bases and among model types (MAP-W for Denver and Knoxville, MAP-lF-P and MAP-R-P for Bellevue load models, and MAP-R-P+nV for Bellevue concentration models) and, in many cases, was not clearly indicated by the values of standard error of estimate for the calibration data set. A scheme to guide MAP selection, based on exploratory data analysis of the calibration data set, is presented and tested. The MAP?s were tested for sensitivity to the size of a calibration data set. As expected, predictive accuracy of all MAP?s for the verification data set decreased as the calibration data-set size decreased, but predictive accuracy was not as sensitive for the MAP?s as it was for the local regression models.
A Neighborhood Analysis of Public Library Use in New York City
ERIC Educational Resources Information Center
Japzon, Andrea C.; Gong, Hongmian
2005-01-01
The use of 200 public libraries in New York City was analyzed according to their neighborhood characteristics. In addition to demographic, economic, and cultural factors traditionally considered, the social and spatial interactions within a neighborhood were related to public library use. Correlation and regression analyses were implemented for…
Artificial Neural Networks: A New Approach to Predicting Application Behavior.
ERIC Educational Resources Information Center
Gonzalez, Julie M. Byers; DesJardins, Stephen L.
2002-01-01
Applied the technique of artificial neural networks to predict which students were likely to apply to one research university. Compared the results to the traditional analysis tool, logistic regression modeling. Found that the addition of artificial intelligence models was a useful new tool for predicting student application behavior. (EV)
Ryberg, Karen R.
2007-01-01
This report presents the results of a study by the U.S. Geological Survey, done in cooperation with the North Dakota State Water Commission, to estimate water-quality constituent concentrations at seven sites on the Sheyenne River, N. Dak. Regression analysis of water-quality data collected in 1980-2006 was used to estimate concentrations for hardness, dissolved solids, calcium, magnesium, sodium, and sulfate. The explanatory variables examined for the regression relations were continuously monitored streamflow, specific conductance, and water temperature. For the conditions observed in 1980-2006, streamflow was a significant explanatory variable for some constituents. Specific conductance was a significant explanatory variable for all of the constituents, and water temperature was not a statistically significant explanatory variable for any of the constituents in this study. The regression relations were evaluated using common measures of variability, including R2, the proportion of variability in the estimated constituent concentration explained by the explanatory variables and regression equation. R2 values ranged from 0.784 for calcium to 0.997 for dissolved solids. The regression relations also were evaluated by calculating the median relative percentage difference (RPD) between measured constituent concentration and the constituent concentration estimated by the regression equations. Median RPDs ranged from 1.7 for dissolved solids to 11.5 for sulfate. The regression relations also may be used to estimate daily constituent loads. The relations should be monitored for change over time, especially at sites 2 and 3 which have a short period of record. In addition, caution should be used when the Sheyenne River is affected by ice or when upstream sites are affected by isolated storm runoff. Almost all of the outliers and highly influential samples removed from the analysis were made during periods when the Sheyenne River might be affected by ice.
Esserman, Denise A.; Moore, Charity G.; Roth, Mary T.
2009-01-01
Older community dwelling adults often take multiple medications for numerous chronic diseases. Non-adherence to these medications can have a large public health impact. Therefore, the measurement and modeling of medication adherence in the setting of polypharmacy is an important area of research. We apply a variety of different modeling techniques (standard linear regression; weighted linear regression; adjusted linear regression; naïve logistic regression; beta-binomial (BB) regression; generalized estimating equations (GEE)) to binary medication adherence data from a study in a North Carolina based population of older adults, where each medication an individual was taking was classified as adherent or non-adherent. In addition, through simulation we compare these different methods based on Type I error rates, bias, power, empirical 95% coverage, and goodness of fit. We find that estimation and inference using GEE is robust to a wide variety of scenarios and we recommend using this in the setting of polypharmacy when adherence is dichotomously measured for multiple medications per person. PMID:20414358
Dinç, Erdal; Ertekin, Zehra Ceren
2016-01-01
An application of parallel factor analysis (PARAFAC) and three-way partial least squares (3W-PLS1) regression models to ultra-performance liquid chromatography-photodiode array detection (UPLC-PDA) data with co-eluted peaks in the same wavelength and time regions was described for the multicomponent quantitation of hydrochlorothiazide (HCT) and olmesartan medoxomil (OLM) in tablets. Three-way dataset of HCT and OLM in their binary mixtures containing telmisartan (IS) as an internal standard was recorded with a UPLC-PDA instrument. Firstly, the PARAFAC algorithm was applied for the decomposition of three-way UPLC-PDA data into the chromatographic, spectral and concentration profiles to quantify the concerned compounds. Secondly, 3W-PLS1 approach was subjected to the decomposition of a tensor consisting of three-way UPLC-PDA data into a set of triads to build 3W-PLS1 regression for the analysis of the same compounds in samples. For the proposed three-way analysis methods in the regression and prediction steps, the applicability and validity of PARAFAC and 3W-PLS1 models were checked by analyzing the synthetic mixture samples, inter-day and intra-day samples, and standard addition samples containing HCT and OLM. Two different three-way analysis methods, PARAFAC and 3W-PLS1, were successfully applied to the quantitative estimation of the solid dosage form containing HCT and OLM. Regression and prediction results provided from three-way analysis were compared with those obtained by traditional UPLC method. Copyright © 2015 Elsevier B.V. All rights reserved.
[Risk factors for anorexia in children].
Liu, Wei-Xiao; Lang, Jun-Feng; Zhang, Qin-Feng
2016-11-01
To investigate the risk factors for anorexia in children, and to reduce the prevalence of anorexia in children. A questionnaire survey and a case-control study were used to collect the general information of 150 children with anorexia (case group) and 150 normal children (control group). Univariate analysis and multivariate logistic stepwise regression analysis were performed to identify the risk factors for anorexia in children. The results of the univariate analysis showed significant differences between the case and control groups in the age in months when supplementary food were added, feeding pattern, whether they liked meat, vegetables and salty food, whether they often took snacks and beverages, whether they liked to play while eating, and whether their parents asked them to eat food on time (P<0.05). The results of the multivariate logistic regression analysis showed that late addition of supplementary food (OR=5.408), high frequency of taking snacks and/or drinks (OR=11.813), and eating while playing (OR=6.654) were major risk factors for anorexia in children. Liking of meat (OR=0.093) and vegetables (OR=0.272) and eating on time required by parents (OR=0.079) were protective factors against anorexia in children. Timely addition of supplementary food, a proper diet, and development of children's proper eating and living habits can reduce the incidence of anorexia in children.
Knol, Mirjam J; van der Tweel, Ingeborg; Grobbee, Diederick E; Numans, Mattijs E; Geerlings, Mirjam I
2007-10-01
To determine the presence of interaction in epidemiologic research, typically a product term is added to the regression model. In linear regression, the regression coefficient of the product term reflects interaction as departure from additivity. However, in logistic regression it refers to interaction as departure from multiplicativity. Rothman has argued that interaction estimated as departure from additivity better reflects biologic interaction. So far, literature on estimating interaction on an additive scale using logistic regression only focused on dichotomous determinants. The objective of the present study was to provide the methods to estimate interaction between continuous determinants and to illustrate these methods with a clinical example. and results From the existing literature we derived the formulas to quantify interaction as departure from additivity between one continuous and one dichotomous determinant and between two continuous determinants using logistic regression. Bootstrapping was used to calculate the corresponding confidence intervals. To illustrate the theory with an empirical example, data from the Utrecht Health Project were used, with age and body mass index as risk factors for elevated diastolic blood pressure. The methods and formulas presented in this article are intended to assist epidemiologists to calculate interaction on an additive scale between two variables on a certain outcome. The proposed methods are included in a spreadsheet which is freely available at: http://www.juliuscenter.nl/additive-interaction.xls.
On the use of log-transformation vs. nonlinear regression for analyzing biological power laws.
Xiao, Xiao; White, Ethan P; Hooten, Mevin B; Durham, Susan L
2011-10-01
Power-law relationships are among the most well-studied functional relationships in biology. Recently the common practice of fitting power laws using linear regression (LR) on log-transformed data has been criticized, calling into question the conclusions of hundreds of studies. It has been suggested that nonlinear regression (NLR) is preferable, but no rigorous comparison of these two methods has been conducted. Using Monte Carlo simulations, we demonstrate that the error distribution determines which method performs better, with NLR better characterizing data with additive, homoscedastic, normal error and LR better characterizing data with multiplicative, heteroscedastic, lognormal error. Analysis of 471 biological power laws shows that both forms of error occur in nature. While previous analyses based on log-transformation appear to be generally valid, future analyses should choose methods based on a combination of biological plausibility and analysis of the error distribution. We provide detailed guidelines and associated computer code for doing so, including a model averaging approach for cases where the error structure is uncertain.
Lee, Seung Hee; Jang, Hyung Suk; Yang, Young Hee
2016-10-01
This study was done to investigate factors influencing successful aging in middle-aged women. A convenience sample of 103 middle-aged women was selected from the community. Data were collected using a structured questionnaire and analyzed using descriptive statistics, two-sample t-test, one-way ANOVA, Kruskal Wallis test, Pearson correlations, Spearman correlations and multiple regression analysis with the SPSS/WIN 22.0 program. Results of regression analysis showed that significant factors influencing successful aging were post-traumatic growth and social support. This regression model explained 48% of the variance in successful aging. Findings show that the concept 'post-traumatic growth' is an important factor influencing successful aging in middle-aged women. In addition, social support from friends/co-workers had greater influence on successful aging than social support from family. Thus, we need to consider the positive impact of post-traumatic growth and increase the chances of social participation in a successful aging program for middle-aged women.
Kuesten, Carla; Bi, Jian
2018-06-03
Conventional drivers of liking analysis was extended with a time dimension into temporal drivers of liking (TDOL) based on functional data analysis methodology and non-additive models for multiple-attribute time-intensity (MATI) data. The non-additive models, which consider both direct effects and interaction effects of attributes to consumer overall liking, include Choquet integral and fuzzy measure in the multi-criteria decision-making, and linear regression based on variance decomposition. Dynamics of TDOL, i.e., the derivatives of the relative importance functional curves were also explored. Well-established R packages 'fda', 'kappalab' and 'relaimpo' were used in the paper for developing TDOL. Applied use of these methods shows that the relative importance of MATI curves offers insights for understanding the temporal aspects of consumer liking for fruit chews.
Test data analysis for concentrating photovoltaic arrays
NASA Astrophysics Data System (ADS)
Maish, A. B.; Cannon, J. E.
A test data analysis approach for use with steady state efficiency measurements taken on concentrating photovoltaic arrays is presented. The analysis procedures can be used to identify based and erroneous data. The steps involved in analyzing the test data are screening the data, developing coefficients for the performance equation, analyzing statistics to ensure adequacy of the regression fit to the data, and plotting the data. In addition, this paper analyzes the sources and magnitudes of precision and bias errors that affect measurement accuracy are analyzed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seong W. Lee
2004-10-01
The systematic tests of the gasifier simulator on the clean thermocouple were completed in this reporting period. Within the systematic tests on the clean thermocouple, five (5) factors were considered as the experimental parameters including air flow rate, water flow rate, fine dust particle amount, ammonia addition and high/low frequency device (electric motor). The fractional factorial design method was used in the experiment design with sixteen (16) data sets of readings. Analysis of Variances (ANOVA) was applied to the results from systematic tests. The ANOVA results show that the un-balanced motor vibration frequency did not have the significant impact onmore » the temperature changes in the gasifier simulator. For the fine dust particles testing, the amount of fine dust particles has significant impact to the temperature measurements in the gasifier simulator. The effects of the air and water on the temperature measurements show the same results as reported in the previous report. The ammonia concentration was included as an experimental parameter for the reducing environment in this reporting period. The ammonia concentration does not seem to be a significant factor on the temperature changes. The linear regression analysis was applied to the temperature reading with five (5) factors. The accuracy of the linear regression is relatively low, which is less than 10% accuracy. Nonlinear regression was also conducted to the temperature reading with the same factors. Since the experiments were designed in two (2) levels, the nonlinear regression is not very effective with the dataset (16 readings). An extra central point test was conducted. With the data of the center point testing, the accuracy of the nonlinear regression is much better than the linear regression.« less
Rębacz-Maron, Ewa; Parafiniuk, Mirosław
2014-01-01
The aim of this paper was to examine the extent to which socioeconomic factors, anthropological data and somatic indices influenced the results of spirometric measurements (FEV1 and FVC) in Tanzanian youth. The population studied were young black Bantu men aged 12.8-24.0 years. Analysis was performed for the whole data set (n = 255), as well as separately for two age groups: under 17.5 years (n = 168) and 17.5 + (n = 87). A backward stepwise multiple regression analysis was performed for FEV1 and FVC as dependent variables on socioeconomic and anthropometric data. Multiple regression analysis for the whole group revealed that the socioeconomic and anthropometric data under analysis accounted for 38% of the variation in FEV1. In addition the analysis demonstrated that 34% of the variation in FVC could be accounted for by the variables used in the regression. A significant impact in explaining the variability of FVC was exhibited by the thorax mobility, financial situation of the participants and Pignet-Verwaecka Index. Analysis of the data indicates the significant role of selected socio-economic factors on the development of the biological specimens investigated. There were no perceptible pathologies, and the results can be treated as a credible interpretation of the influence exerted by the environment in which the teenagers under study grew up.
Leite, Fábio R M; Nascimento, Gustavo G; Demarco, Flávio F; Gomes, Brenda P F A; Pucci, Cesar R; Martinho, Frederico C
2015-05-01
This systematic review and meta-regression analysis aimed to calculate a combined prevalence estimate and evaluate the prevalence of different Treponema species in primary and secondary endodontic infections, including symptomatic and asymptomatic cases. The MEDLINE/PubMed, Embase, Scielo, Web of Knowledge, and Scopus databases were searched without starting date restriction up to and including March 2014. Only reports in English were included. The selected literature was reviewed by 2 authors and classified as suitable or not to be included in this review. Lists were compared, and, in case of disagreements, decisions were made after a discussion based on inclusion and exclusion criteria. A pooled prevalence of Treponema species in endodontic infections was estimated. Additionally, a meta-regression analysis was performed. Among the 265 articles identified in the initial search, only 51 were included in the final analysis. The studies were classified into 2 different groups according to the type of endodontic infection and whether it was an exclusively primary/secondary study (n = 36) or a primary/secondary comparison (n = 15). The pooled prevalence of Treponema species was 41.5% (95% confidence interval, 35.9-47.0). In the multivariate model of meta-regression analysis, primary endodontic infections (P < .001), acute apical abscess, symptomatic apical periodontitis (P < .001), and concomitant presence of 2 or more species (P = .028) explained the heterogeneity regarding the prevalence rates of Treponema species. Our findings suggest that Treponema species are important pathogens involved in endodontic infections, particularly in cases of primary and acute infections. Copyright © 2015 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.
Fossum, Kenneth D.; O'Day, Christie M.; Wilson, Barbara J.; Monical, Jim E.
2001-01-01
Stormwater and streamflow in Maricopa County were monitored to (1) describe the physical, chemical, and toxicity characteristics of stormwater from areas having different land uses, (2) describe the physical, chemical, and toxicity characteristics of streamflow from areas that receive urban stormwater, and (3) estimate constituent loads in stormwater. Urban stormwater and streamflow had similar ranges in most constituent concentrations. The mean concentration of dissolved solids in urban stormwater was lower than in streamflow from the Salt River and Indian Bend Wash. Urban stormwater, however, had a greater chemical oxygen demand and higher concentrations of most nutrients. Mean seasonal loads and mean annual loads of 11 constituents and volumes of runoff were estimated for municipalities in the metropolitan Phoenix area, Arizona, by adjusting regional regression equations of loads. This adjustment procedure uses the original regional regression equation and additional explanatory variables that were not included in the original equation. The adjusted equations had standard errors that ranged from 161 to 196 percent. The large standard errors of the prediction result from the large variability of the constituent concentration data used in the regression analysis. Adjustment procedures produced unsatisfactory results for nine of the regressions?suspended solids, dissolved solids, total phosphorus, dissolved phosphorus, total recoverable cadmium, total recoverable copper, total recoverable lead, total recoverable zinc, and storm runoff. These equations had no consistent direction of bias and no other additional explanatory variables correlated with the observed loads. A stepwise-multiple regression or a three-variable regression (total storm rainfall, drainage area, and impervious area) and local data were used to develop local regression equations for these nine constituents. These equations had standard errors from 15 to 183 percent.
Sando, Roy; Sando, Steven K.; McCarthy, Peter M.; Dutton, DeAnn M.
2016-04-05
The U.S. Geological Survey (USGS), in cooperation with the Montana Department of Natural Resources and Conservation, completed a study to update methods for estimating peak-flow frequencies at ungaged sites in Montana based on peak-flow data at streamflow-gaging stations through water year 2011. The methods allow estimation of peak-flow frequencies (that is, peak-flow magnitudes, in cubic feet per second, associated with annual exceedance probabilities of 66.7, 50, 42.9, 20, 10, 4, 2, 1, 0.5, and 0.2 percent) at ungaged sites. The annual exceedance probabilities correspond to 1.5-, 2-, 2.33-, 5-, 10-, 25-, 50-, 100-, 200-, and 500-year recurrence intervals, respectively.Regional regression analysis is a primary focus of Chapter F of this Scientific Investigations Report, and regression equations for estimating peak-flow frequencies at ungaged sites in eight hydrologic regions in Montana are presented. The regression equations are based on analysis of peak-flow frequencies and basin characteristics at 537 streamflow-gaging stations in or near Montana and were developed using generalized least squares regression or weighted least squares regression.All of the data used in calculating basin characteristics that were included as explanatory variables in the regression equations were developed for and are available through the USGS StreamStats application (http://water.usgs.gov/osw/streamstats/) for Montana. StreamStats is a Web-based geographic information system application that was created by the USGS to provide users with access to an assortment of analytical tools that are useful for water-resource planning and management. The primary purpose of the Montana StreamStats application is to provide estimates of basin characteristics and streamflow characteristics for user-selected ungaged sites on Montana streams. The regional regression equations presented in this report chapter can be conveniently solved using the Montana StreamStats application.Selected results from this study were compared with results of previous studies. For most hydrologic regions, the regression equations reported for this study had lower mean standard errors of prediction (in percent) than the previously reported regression equations for Montana. The equations presented for this study are considered to be an improvement on the previously reported equations primarily because this study (1) included 13 more years of peak-flow data; (2) included 35 more streamflow-gaging stations than previous studies; (3) used a detailed geographic information system (GIS)-based definition of the regulation status of streamflow-gaging stations, which allowed better determination of the unregulated peak-flow records that are appropriate for use in the regional regression analysis; (4) included advancements in GIS and remote-sensing technologies, which allowed more convenient calculation of basin characteristics and investigation of many more candidate basin characteristics; and (5) included advancements in computational and analytical methods, which allowed more thorough and consistent data analysis.This report chapter also presents other methods for estimating peak-flow frequencies at ungaged sites. Two methods for estimating peak-flow frequencies at ungaged sites located on the same streams as streamflow-gaging stations are described. Additionally, envelope curves relating maximum recorded annual peak flows to contributing drainage area for each of the eight hydrologic regions in Montana are presented and compared to a national envelope curve. In addition to providing general information on characteristics of large peak flows, the regional envelope curves can be used to assess the reasonableness of peak-flow frequency estimates determined using the regression equations.
NASA Astrophysics Data System (ADS)
McCulley, Jonathan M.
This research investigates the application of additive manufacturing techniques for fabricating hybrid rocket fuel grains composed of porous Acrylonitrile-butadiene-styrene impregnated with paraffin wax. The digitally manufactured ABS substrate provides mechanical support for the paraffin fuel material and serves as an additional fuel component. The embedded paraffin provides an enhanced fuel regression rate while having no detrimental effect on the thermodynamic burn properties of the fuel grain. Multiple fuel grains with various ABS-to-Paraffin mass ratios were fabricated and burned with nitrous oxide. Analytical predictions for end-to-end motor performance and fuel regression are compared against static test results. Baseline fuel grain regression calculations use an enthalpy balance energy analysis with the material and thermodynamic properties based on the mean paraffin/ABS mass fractions within the fuel grain. In support of these analytical comparisons, a novel method for propagating the fuel port burn surface was developed. In this modeling approach the fuel cross section grid is modeled as an image with white pixels representing the fuel and black pixels representing empty or burned grid cells.
Ruuska, Salla; Hämäläinen, Wilhelmiina; Kajava, Sari; Mughal, Mikaela; Matilainen, Pekka; Mononen, Jaakko
2018-03-01
The aim of the present study was to evaluate empirically confusion matrices in device validation. We compared the confusion matrix method to linear regression and error indices in the validation of a device measuring feeding behaviour of dairy cattle. In addition, we studied how to extract additional information on classification errors with confusion probabilities. The data consisted of 12 h behaviour measurements from five dairy cows; feeding and other behaviour were detected simultaneously with a device and from video recordings. The resulting 216 000 pairs of classifications were used to construct confusion matrices and calculate performance measures. In addition, hourly durations of each behaviour were calculated and the accuracy of measurements was evaluated with linear regression and error indices. All three validation methods agreed when the behaviour was detected very accurately or inaccurately. Otherwise, in the intermediate cases, the confusion matrix method and error indices produced relatively concordant results, but the linear regression method often disagreed with them. Our study supports the use of confusion matrix analysis in validation since it is robust to any data distribution and type of relationship, it makes a stringent evaluation of validity, and it offers extra information on the type and sources of errors. Copyright © 2018 Elsevier B.V. All rights reserved.
Nateghi, Roshanak; Guikema, Seth D; Quiring, Steven M
2011-12-01
This article compares statistical methods for modeling power outage durations during hurricanes and examines the predictive accuracy of these methods. Being able to make accurate predictions of power outage durations is valuable because the information can be used by utility companies to plan their restoration efforts more efficiently. This information can also help inform customers and public agencies of the expected outage times, enabling better collective response planning, and coordination of restoration efforts for other critical infrastructures that depend on electricity. In the long run, outage duration estimates for future storm scenarios may help utilities and public agencies better allocate risk management resources to balance the disruption from hurricanes with the cost of hardening power systems. We compare the out-of-sample predictive accuracy of five distinct statistical models for estimating power outage duration times caused by Hurricane Ivan in 2004. The methods compared include both regression models (accelerated failure time (AFT) and Cox proportional hazard models (Cox PH)) and data mining techniques (regression trees, Bayesian additive regression trees (BART), and multivariate additive regression splines). We then validate our models against two other hurricanes. Our results indicate that BART yields the best prediction accuracy and that it is possible to predict outage durations with reasonable accuracy. © 2011 Society for Risk Analysis.
40 CFR 86.1823-01 - Durability demonstration procedures for exhaust emissions.
Code of Federal Regulations, 2011 CFR
2011-07-01
... (including both hardware and software) must be installed and operating for the entire mileage accumulation... decimal places) from the regression analysis; the result shall be rounded to three-decimal places of... less than one shall be changed to one for the purposes of this paragraph. (2) An additive DF will be...
Salary Equity Studies: The State of the Art. ASHE Annual Meeting 1982 Paper.
ERIC Educational Resources Information Center
Hengstler, Dennis D.; And Others
The strengths and weaknesses of various methodologies in conducting salary equity studies are examined. Particular attention is paid to the problems of identifying appropriate matches in the paired-comparison approach and to the sample, predictor and decision-rule problems associated with the regression analysis approach. In addition, highlights…
ERIC Educational Resources Information Center
Fu, Ching-Sheue
2015-01-01
In this study, the participants comprised 385 preschool teachers. The relationship among their emotional labor, Job Involvement, and psychological capital were examined using hierarchical regression analysis. In addition, whether psychological capital exerted a mediating effect on Job Involvement was investigated. The results show that "deep…
ERIC Educational Resources Information Center
Astrom, Raven L.; Wadsworth, Sally J.; Olson, Richard K.; Willcutt, Erik G.; DeFries, John C.
2012-01-01
Reading performance data from 254 pairs of identical (MZ) and 420 pairs of fraternal (DZ) twins, 8.0 to 20.0 years of age, were subjected to multiple regression analyses. An extension of the DeFries-Fulker (DF) analysis (DeFries & Fulker, 1985, 1988) that facilitated inclusion of data from 303 of their nontwin siblings was employed. In addition to…
Peak oxygen consumption measured during the stair-climbing test in lung resection candidates.
Brunelli, Alessandro; Xiumé, Francesco; Refai, Majed; Salati, Michele; Di Nunzio, Luca; Pompili, Cecilia; Sabbatini, Armando
2010-01-01
The stair-climbing test is commonly used in the preoperative evaluation of lung resection candidates, but it is difficult to standardize and provides little physiologic information on the performance. To verify the association between the altitude and the V(O2peak) measured during the stair-climbing test. 109 consecutive candidates for lung resection performed a symptom-limited stair-climbing test with direct breath-by-breath measurement of V(O2peak) by a portable gas analyzer. Stepwise logistic regression and bootstrap analyses were used to verify the association of several perioperative variables with a V(O2peak) <15 ml/kg/min. Subsequently, multiple regression analysis was also performed to develop an equation to estimate V(O2peak) from stair-climbing parameters and other patient-related variables. 56% of patients climbing <14 m had a V(O2peak) <15 ml/kg/min, whereas 98% of those climbing >22 m had a V(O2peak) >15 ml/kg/min. The altitude reached at stair-climbing test resulted in the only significant predictor of a V(O2peak) <15 ml/kg/min after logistic regression analysis. Multiple regression analysis yielded an equation to estimate V(O2peak) factoring altitude (p < 0.0001), speed of ascent (p = 0.005) and body mass index (p = 0.0008). There was an association between altitude and V(O2peak) measured during the stair-climbing test. Most of the patients climbing more than 22 m are able to generate high values of V(O2peak) and can proceed to surgery without any additional tests. All others need to be referred for a formal cardiopulmonary exercise test. In addition, we were able to generate an equation to estimate V(O2peak), which could assist in streamlining the preoperative workup and could be used across different settings to standardize this test. Copyright (c) 2010 S. Karger AG, Basel.
Application of logistic regression to case-control association studies involving two causative loci.
North, Bernard V; Curtis, David; Sham, Pak C
2005-01-01
Models in which two susceptibility loci jointly influence the risk of developing disease can be explored using logistic regression analysis. Comparison of likelihoods of models incorporating different sets of disease model parameters allows inferences to be drawn regarding the nature of the joint effect of the loci. We have simulated case-control samples generated assuming different two-locus models and then analysed them using logistic regression. We show that this method is practicable and that, for the models we have used, it can be expected to allow useful inferences to be drawn from sample sizes consisting of hundreds of subjects. Interactions between loci can be explored, but interactive effects do not exactly correspond with classical definitions of epistasis. We have particularly examined the issue of the extent to which it is helpful to utilise information from a previously identified locus when investigating a second, unknown locus. We show that for some models conditional analysis can have substantially greater power while for others unconditional analysis can be more powerful. Hence we conclude that in general both conditional and unconditional analyses should be performed when searching for additional loci.
Regression Analysis of Mixed Panel Count Data with Dependent Terminal Events
Yu, Guanglei; Zhu, Liang; Li, Yang; Sun, Jianguo; Robison, Leslie L.
2017-01-01
Event history studies are commonly conducted in many fields and a great deal of literature has been established for the analysis of the two types of data commonly arising from these studies: recurrent event data and panel count data. The former arises if all study subjects are followed continuously, while the latter means that each study subject is observed only at discrete time points. In reality, a third type of data, a mixture of the two types of the data above, may occur and furthermore, as with the first two types of the data, there may exist a dependent terminal event, which may preclude the occurrences of recurrent events of interest. This paper discusses regression analysis of mixed recurrent event and panel count data in the presence of a terminal event and an estimating equation-based approach is proposed for estimation of regression parameters of interest. In addition, the asymptotic properties of the proposed estimator are established and a simulation study conducted to assess the finite-sample performance of the proposed method suggests that it works well in practical situations. Finally the methodology is applied to a childhood cancer study that motivated this study. PMID:28098397
Development of high performance hybrid rocket fuels
NASA Astrophysics Data System (ADS)
Zaseck, Christopher R.
In this document I discuss paraffin fuel combustion and investigate the effects of additives on paraffin entrainment and regression. In general, hybrid rockets offer an economical and safe alternative to standard liquid and solid rockets. However, slow polymeric fuel regression and low combustion efficiency have limited the commercial use of hybrid rockets. Paraffin is a fast burning fuel that has received significant attention in the 2000's and 2010's as a replacement for standard fuels. Paraffin regresses three to four times faster than polymeric fuels due to the entrainment of a surface melt layer. However, further regression rate enhancement over the base paraffin fuel is necessary for widespread hybrid rocket adoption. I use a small scale opposed flow burner to investigate the effect of additives on the combustion of paraffin. Standard additives such as aluminum combust above the flame zone where sufficient oxidizer levels are present. As a result no heat is generated below the flame itself. In small scale opposed burner experiments the effect of limited heat feedback is apparent. Aluminum in particular does not improve the regression of paraffin in the opposed burner. The lack of heat feedback from additive combustion limits the applicability of the opposed burner. In turn, the results obtained in the opposed burner with metal additive loaded hybrid fuels do not match results from hybrid rocket experiments. In addition, nano-scale aluminum increases melt layer viscosity and greatly slows the regression of paraffin in the opposed flow burner. However, the reactive additives improve the regression rate of paraffin in the opposed burner where standard metals do not. At 5 wt.% mechanically activated titanium and carbon (Ti-C) improves the regression rate of paraffin by 47% in the opposed burner. The mechanically activated Ti C likely reacts in or near the melt layer and provides heat feedback below the flame region that results in faster opposed burner regression. In order to examine paraffin/additive combustion in a motor environment, I conducted experiments on well characterized aluminum based additives. In particular, I investigate the influence of aluminum, unpassivated aluminum, milled aluminum/polytetrafluoroethylene (PTFE), and aluminum hydride on the performance of paraffin fuels for hybrid rocket propulsion. I use an optically accessible combustor to examine the performance of the fuel mixtures in terms of characteristic velocity efficiency and regression rate. Each combustor test consumes a 12.7 cm long, 1.9 cm diameter fuel strand under 160 kg/m 2s of oxygen at up to 1.4 MPa. The experimental results indicate that the addition of 5 wt.% 30 mum or 80 nm aluminum to paraffin increases the regression rate by approximately 15% compared to neat paraffin grains. At higher aluminum concentrations and nano-scale particles sizes, the increased melt layer viscosity causes slower regression. Alane and Al/PTFE at 12.5 wt.% increase the regression of paraffin by 21% and 32% respectively. Finally, an aging study indicates that paraffin can protect air and moisture sensitive particles from oxidation. The opposed burner and aluminum/paraffin hybrid rocket experiments show that additives can alter bulk fuel properties, such as viscosity, that regulate entrainment. The general effect of melt layer properties on the entrainment and regression rate of paraffin is not well understood. Improved understanding of how solid additives affect the properties and regression of paraffin is essential to maximize performance. In this document I investigate the effect of melt layer properties on paraffin regression using inert additives. Tests are performed in the optical cylindrical combustor at ˜1 MPa under a gaseous oxygen mass flux of ˜160 kg/m2s. The experiments indicate that the regression rate is proportional to mu0.08rho 0.38kappa0.82. In addition, I explore how to predict fuel viscosity, thermal conductivity, and density prior to testing. Mechanically activated Ti-C and Al/PTFE are examined in the optical combustor. I examine the effect of the reactivity by altering the mill time for the Ti-C and Al/PTFE particles. Mechanical activation of both Ti-C and Al/PTFE improve the regression rate of paraffin more than the unmilled additives. At 12.5 wt.% Al/PTFE milled for 40 minutes regresses 12% faster than the unmilled fuel. Similarly, at 12.5 wt.% 7.5 minute milled Ti C regresses 7% faster than unmilled Ti-C. The reactive particles increase heat transfer to the fuel surface and improve regression. The composition of the combustion products are examined using a particle catcher system in conjunction with visible light and electron microscopy. The exhaust products indicate that the mechanical activation of the Al/PTFE particles cause microexplosions that reduce exhaust particle size. However, the composition of the mechanically activated Al/PTFE products do not indicate more complete combustion. In addition, the mechanically activated and unmilled Ti-C showed no difference in exhaust products.
Qualitative Analysis of Dairy and Powder Milk Using Laser-Induced Breakdown Spectroscopy (LIBS).
Alfarraj, Bader A; Sanghapi, Herve K; Bhatt, Chet R; Yueh, Fang Y; Singh, Jagdish P
2018-01-01
Laser-induced breakdown spectroscopy (LIBS) technique was used to compare various types of commercial milk products. Laser-induced breakdown spectroscopy spectra were investigated for the determination of the elemental composition of soy and rice milk powder, dairy milk, and lactose-free dairy milk. The analysis was performed using radiative transitions. Atomic emissions from Ca, K, Na, and Mg lines observed in LIBS spectra of dairy milk were compared. In addition, proteins and fat level in milks can be determined using molecular emissions such as CN bands. Ca concentrations were calculated to be 2.165 ± 0.203 g/L in 1% of dairy milk fat samples and 2.809 ± 0.172 g/L in 2% of dairy milk fat samples using the standard addition method (SAM) with LIBS spectra. Univariate and multivariate statistical analysis methods showed that the contents of major mineral elements were higher in lactose-free dairy milk than those in dairy milk. The principal component analysis (PCA) method was used to discriminate four milk samples depending on their mineral elements concentration. In addition, proteins and fat level in dairy milks were determined using molecular emissions such as CN band. We applied partial least squares regression (PLSR) and simple linear regression (SLR) models to predict levels of milk fat in dairy milk samples. The PLSR model was successfully used to predict levels of milk fat in dairy milk sample with the relative accuracy (RA%) less than 6.62% using CN (0,0) band.
Cost analysis of incidental durotomy in spine surgery.
Nandyala, Sreeharsha V; Elboghdady, Islam M; Marquez-Lara, Alejandro; Noureldin, Mohamed N B; Sankaranarayanan, Sriram; Singh, Kern
2014-08-01
Retrospective database analysis. To characterize the consequences of an incidental durotomy with regard to perioperative complications and total hospital costs. There is a paucity of data regarding how an incidental durotomy and its associated complications may relate to total hospital costs. The Nationwide Inpatient Sample database was queried from 2008 to 2011. Patients who underwent cervical or lumbar decompression and/or fusion procedures were identified, stratified by approach, and separated into cohorts based on a documented intraoperative incidental durotomy. Patient demographics, comorbidities (Charlson Comorbidity Index), length of hospital stay, perioperative outcomes, and costs were assessed. Analysis of covariance and multivariate linear regression were used to assess the adjusted mean costs of hospitalization as a function of durotomy. The incidental durotomy rate in cervical and lumbar spine surgery is 0.4% and 2.9%, respectively. Patients with an incidental durotomy incurred a longer hospitalization and a greater incidence of perioperative complications including hematoma and neurological injury (P < 0.001). Regression analysis demonstrated that a cervical durotomy and its postoperative sequelae contributed an additional adjusted $7638 (95% confidence interval, 6489-8787; P < 0.001) to the total hospital costs. Similarly, lumbar durotomy contributed an additional adjusted $2412 (95% confidence interval, 1920-2902; P < 0.001) to the total hospital costs. The approach-specific procedural groups demonstrated similar discrepancies in the mean total hospital costs as a function of durotomy. This analysis of the Nationwide Inpatient Sample database demonstrates that incidental durotomies increase hospital resource utilization and costs. In addition, it seems that a cervical durotomy and its associated complications carry a greater financial burden than a lumbar durotomy. Further studies are warranted to investigate the long-term financial implications of incidental durotomies in spine surgery and to reduce the costs associated with this complication. 3.
Coates, Laura C; Walsh, Jessica; Haroon, Muhammad; FitzGerald, Oliver; Aslam, Tariq; Al Balushi, Farida; Burden, A D; Burden-Teh, Esther; Caperon, Anna R; Cerio, Rino; Chattopadhyay, Chandrabhusan; Chinoy, Hector; Goodfield, Mark J D; Kay, Lesley; Kelly, Stephen; Kirkham, Bruce W; Lovell, Christopher R; Marzo-Ortega, Helena; McHugh, Neil; Murphy, Ruth; Reynolds, Nick J; Smith, Catherine H; Stewart, Elizabeth J C; Warren, Richard B; Waxman, Robin; Wilson, Hilary E; Helliwell, Philip S
2014-09-01
Several questionnaires have been developed to screen for psoriatic arthritis (PsA), but head-to-head studies have found limitations. This study aimed to develop new questionnaires encompassing the most discriminative questions from existing instruments. Data from the CONTEST study, a head-to-head comparison of 3 existing questionnaires, were used to identify items with a Youden index score of ≥0.1. These were combined using 4 approaches: CONTEST (simple additions of questions), CONTESTw (weighting using logistic regression), CONTESTjt (addition of a joint manikin), and CONTESTtree (additional questions identified by classification and regression tree [CART] analysis). These candidate questionnaires were tested in independent data sets. Twelve individual questions with a Youden index score of ≥0.1 were identified, but 4 of these were excluded due to duplication and redundancy. Weighting for 2 of these questions was included in CONTESTw. Receiver operating characteristic (ROC) curve analysis showed that involvement in 6 joint areas on the manikin was predictive of PsA for inclusion in CONTESTjt. CART analysis identified a further 5 questions for inclusion in CONTESTtree. CONTESTtree was not significant on ROC curve analysis and discarded. The other 3 questionnaires were significant in all data sets, although CONTESTw was slightly inferior to the others in the validation data sets. Potential cut points for referral were also discussed. Of 4 candidate questionnaires combining existing discriminatory items to identify PsA in people with psoriasis, 3 were found to be significant on ROC curve analysis. Testing in independent data sets identified 2 questionnaires (CONTEST and CONTESTjt) that should be pursued for further prospective testing. Copyright © 2014 by the American College of Rheumatology.
Hill, Andrew; Loh, Po-Ru; Bharadwaj, Ragu B.; Pons, Pascal; Shang, Jingbo; Guinan, Eva; Lakhani, Karim; Kilty, Iain
2017-01-01
Abstract Background: The association of differing genotypes with disease-related phenotypic traits offers great potential to both help identify new therapeutic targets and support stratification of patients who would gain the greatest benefit from specific drug classes. Development of low-cost genotyping and sequencing has made collecting large-scale genotyping data routine in population and therapeutic intervention studies. In addition, a range of new technologies is being used to capture numerous new and complex phenotypic descriptors. As a result, genotype and phenotype datasets have grown exponentially. Genome-wide association studies associate genotypes and phenotypes using methods such as logistic regression. As existing tools for association analysis limit the efficiency by which value can be extracted from increasing volumes of data, there is a pressing need for new software tools that can accelerate association analyses on large genotype-phenotype datasets. Results: Using open innovation (OI) and contest-based crowdsourcing, the logistic regression analysis in a leading, community-standard genetics software package (PLINK 1.07) was substantially accelerated. OI allowed us to do this in <6 months by providing rapid access to highly skilled programmers with specialized, difficult-to-find skill sets. Through a crowd-based contest a combination of computational, numeric, and algorithmic approaches was identified that accelerated the logistic regression in PLINK 1.07 by 18- to 45-fold. Combining contest-derived logistic regression code with coarse-grained parallelization, multithreading, and associated changes to data initialization code further developed through distributed innovation, we achieved an end-to-end speedup of 591-fold for a data set size of 6678 subjects by 645 863 variants, compared to PLINK 1.07's logistic regression. This represents a reduction in run time from 4.8 hours to 29 seconds. Accelerated logistic regression code developed in this project has been incorporated into the PLINK2 project. Conclusions: Using iterative competition-based OI, we have developed a new, faster implementation of logistic regression for genome-wide association studies analysis. We present lessons learned and recommendations on running a successful OI process for bioinformatics. PMID:28327993
Hill, Andrew; Loh, Po-Ru; Bharadwaj, Ragu B; Pons, Pascal; Shang, Jingbo; Guinan, Eva; Lakhani, Karim; Kilty, Iain; Jelinsky, Scott A
2017-05-01
The association of differing genotypes with disease-related phenotypic traits offers great potential to both help identify new therapeutic targets and support stratification of patients who would gain the greatest benefit from specific drug classes. Development of low-cost genotyping and sequencing has made collecting large-scale genotyping data routine in population and therapeutic intervention studies. In addition, a range of new technologies is being used to capture numerous new and complex phenotypic descriptors. As a result, genotype and phenotype datasets have grown exponentially. Genome-wide association studies associate genotypes and phenotypes using methods such as logistic regression. As existing tools for association analysis limit the efficiency by which value can be extracted from increasing volumes of data, there is a pressing need for new software tools that can accelerate association analyses on large genotype-phenotype datasets. Using open innovation (OI) and contest-based crowdsourcing, the logistic regression analysis in a leading, community-standard genetics software package (PLINK 1.07) was substantially accelerated. OI allowed us to do this in <6 months by providing rapid access to highly skilled programmers with specialized, difficult-to-find skill sets. Through a crowd-based contest a combination of computational, numeric, and algorithmic approaches was identified that accelerated the logistic regression in PLINK 1.07 by 18- to 45-fold. Combining contest-derived logistic regression code with coarse-grained parallelization, multithreading, and associated changes to data initialization code further developed through distributed innovation, we achieved an end-to-end speedup of 591-fold for a data set size of 6678 subjects by 645 863 variants, compared to PLINK 1.07's logistic regression. This represents a reduction in run time from 4.8 hours to 29 seconds. Accelerated logistic regression code developed in this project has been incorporated into the PLINK2 project. Using iterative competition-based OI, we have developed a new, faster implementation of logistic regression for genome-wide association studies analysis. We present lessons learned and recommendations on running a successful OI process for bioinformatics. © The Author 2017. Published by Oxford University Press.
Poisson Regression Analysis of Illness and Injury Surveillance Data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Frome E.L., Watkins J.P., Ellis E.D.
2012-12-12
The Department of Energy (DOE) uses illness and injury surveillance to monitor morbidity and assess the overall health of the work force. Data collected from each participating site include health events and a roster file with demographic information. The source data files are maintained in a relational data base, and are used to obtain stratified tables of health event counts and person time at risk that serve as the starting point for Poisson regression analysis. The explanatory variables that define these tables are age, gender, occupational group, and time. Typical response variables of interest are the number of absences duemore » to illness or injury, i.e., the response variable is a count. Poisson regression methods are used to describe the effect of the explanatory variables on the health event rates using a log-linear main effects model. Results of fitting the main effects model are summarized in a tabular and graphical form and interpretation of model parameters is provided. An analysis of deviance table is used to evaluate the importance of each of the explanatory variables on the event rate of interest and to determine if interaction terms should be considered in the analysis. Although Poisson regression methods are widely used in the analysis of count data, there are situations in which over-dispersion occurs. This could be due to lack-of-fit of the regression model, extra-Poisson variation, or both. A score test statistic and regression diagnostics are used to identify over-dispersion. A quasi-likelihood method of moments procedure is used to evaluate and adjust for extra-Poisson variation when necessary. Two examples are presented using respiratory disease absence rates at two DOE sites to illustrate the methods and interpretation of the results. In the first example the Poisson main effects model is adequate. In the second example the score test indicates considerable over-dispersion and a more detailed analysis attributes the over-dispersion to extra-Poisson variation. The R open source software environment for statistical computing and graphics is used for analysis. Additional details about R and the data that were used in this report are provided in an Appendix. Information on how to obtain R and utility functions that can be used to duplicate results in this report are provided.« less
Covariance functions for body weight from birth to maturity in Nellore cows.
Boligon, A A; Mercadante, M E Z; Forni, S; Lôbo, R B; Albuquerque, L G
2010-03-01
The objective of this study was to estimate (co)variance functions using random regression models on Legendre polynomials for the analysis of repeated measures of BW from birth to adult age. A total of 82,064 records from 8,145 females were analyzed. Different models were compared. The models included additive direct and maternal effects, and animal and maternal permanent environmental effects as random terms. Contemporary group and dam age at calving (linear and quadratic effect) were included as fixed effects, and orthogonal Legendre polynomials of animal age (cubic regression) were considered as random covariables. Eight models with polynomials of third to sixth order were used to describe additive direct and maternal effects, and animal and maternal permanent environmental effects. Residual effects were modeled using 1 (i.e., assuming homogeneity of variances across all ages) or 5 age classes. The model with 5 classes was the best to describe the trajectory of residuals along the growth curve. The model including fourth- and sixth-order polynomials for additive direct and animal permanent environmental effects, respectively, and third-order polynomials for maternal genetic and maternal permanent environmental effects were the best. Estimates of (co)variance obtained with the multi-trait and random regression models were similar. Direct heritability estimates obtained with the random regression models followed a trend similar to that obtained with the multi-trait model. The largest estimates of maternal heritability were those of BW taken close to 240 d of age. In general, estimates of correlation between BW from birth to 8 yr of age decreased with increasing distance between ages.
A Comparative Study of Pairwise Learning Methods Based on Kernel Ridge Regression.
Stock, Michiel; Pahikkala, Tapio; Airola, Antti; De Baets, Bernard; Waegeman, Willem
2018-06-12
Many machine learning problems can be formulated as predicting labels for a pair of objects. Problems of that kind are often referred to as pairwise learning, dyadic prediction, or network inference problems. During the past decade, kernel methods have played a dominant role in pairwise learning. They still obtain a state-of-the-art predictive performance, but a theoretical analysis of their behavior has been underexplored in the machine learning literature. In this work we review and unify kernel-based algorithms that are commonly used in different pairwise learning settings, ranging from matrix filtering to zero-shot learning. To this end, we focus on closed-form efficient instantiations of Kronecker kernel ridge regression. We show that independent task kernel ridge regression, two-step kernel ridge regression, and a linear matrix filter arise naturally as a special case of Kronecker kernel ridge regression, implying that all these methods implicitly minimize a squared loss. In addition, we analyze universality, consistency, and spectral filtering properties. Our theoretical results provide valuable insights into assessing the advantages and limitations of existing pairwise learning methods.
Detection of epistatic effects with logic regression and a classical linear regression model.
Malina, Magdalena; Ickstadt, Katja; Schwender, Holger; Posch, Martin; Bogdan, Małgorzata
2014-02-01
To locate multiple interacting quantitative trait loci (QTL) influencing a trait of interest within experimental populations, usually methods as the Cockerham's model are applied. Within this framework, interactions are understood as the part of the joined effect of several genes which cannot be explained as the sum of their additive effects. However, if a change in the phenotype (as disease) is caused by Boolean combinations of genotypes of several QTLs, this Cockerham's approach is often not capable to identify them properly. To detect such interactions more efficiently, we propose a logic regression framework. Even though with the logic regression approach a larger number of models has to be considered (requiring more stringent multiple testing correction) the efficient representation of higher order logic interactions in logic regression models leads to a significant increase of power to detect such interactions as compared to a Cockerham's approach. The increase in power is demonstrated analytically for a simple two-way interaction model and illustrated in more complex settings with simulation study and real data analysis.
A stream-gaging network analysis for the 7-day, 10-year annual low flow in New Hampshire streams
Flynn, Robert H.
2003-01-01
The 7-day, 10-year (7Q10) low-flow-frequency statistic is a widely used measure of surface-water availability in New Hampshire. Regression equations and basin-characteristic digital data sets were developed to help water-resource managers determine surface-water resources during periods of low flow in New Hampshire streams. These regression equations and data sets were developed to estimate streamflow statistics for the annual and seasonal low-flow-frequency, and period-of-record and seasonal period-of-record flow durations. generalized-least-squares (GLS) regression methods were used to develop the annual 7Q10 low-flow-frequency regression equation from 60 continuous-record stream-gaging stations in New Hampshire and in neighboring States. In the regression equation, the dependent variables were the annual 7Q10 flows at the 60 stream-gaging stations. The independent (or predictor) variables were objectively selected characteristics of the drainage basins that contribute flow to those stations. In contrast to ordinary-least-squares (OLS) regression analysis, GLS-developed estimating equations account for differences in length of record and spatial correlations among the flow-frequency statistics at the various stations.A total of 93 measurable drainage-basin characteristics were candidate independent variables. On the basis of several statistical parameters that were used to evaluate which combination of basin characteristics contribute the most to the predictive power of the equations, three drainage-basin characteristics were determined to be statistically significant predictors of the annual 7Q10: (1) total drainage area, (2) mean summer stream-gaging station precipitation from 1961 to 90, and (3) average mean annual basinwide temperature from 1961 to 1990.To evaluate the effectiveness of the stream-gaging network in providing regional streamflow data for the annual 7Q10, the computer program GLSNET (generalized-least-squares NETwork) was used to analyze the network by application of GLS regression between streamflow and the climatic and basin characteristics of the drainage basin upstream from each stream-gaging station. Improvement to the predictive ability of the regression equations developed for the network analyses is measured by the reduction in the average sampling-error variance, and can be achieved by collecting additional streamflow data at existing stations. The predictive ability of the regression equations is enhanced even further with the addition of new stations to the network. Continued data collection at unregulated stream-gaging stations with less than 14 years of record resulted in the greatest cost-weighted reduction to the average sampling-error variance of the annual 7Q10 regional regression equation. The addition of new stations in basins with underrepresented values for the independent variables of the total drainage area, average mean annual basinwide temperature, or mean summer stream-gaging station precipitation in the annual 7Q10 regression equation yielded a much greater cost-weighted reduction to the average sampling-error variance than when more data were collected at existing unregulated stations. To maximize the regional information obtained from the stream-gaging network for the annual 7Q10, ranking of the streamflow data can be used to determine whether an active station should be continued or if a new or discontinued station should be activated for streamflow data collection. Thus, this network analysis can help determine the costs and benefits of continuing the operation of a particular station or activating a new station at another location to predict the 7Q10 at ungaged stream reaches. The decision to discontinue an existing station or activate a new station, however, must also consider its contribution to other water-resource analyses such as flood management, water quality, or trends in land use or climatic change.
HIV-related ocular microangiopathic syndrome and color contrast sensitivity.
Geier, S A; Hammel, G; Bogner, J R; Kronawitter, U; Berninger, T; Goebel, F D
1994-06-01
Color vision deficits in patients with acquired immunodeficiency syndrome (AIDS) or human immunodeficiency virus (HIV) disease were reported, and a retinal pathogenic mechanism was proposed. The purpose of this study was to evaluate the association of color vision deficits with HIV-related retinal microangiopathy. A computer graphics system was used to measure protan, deutan, and tritan color contrast sensitivity (CCS) thresholds in 60 HIV-infected patients. Retinal microangiopathy was measured by counting the number of cotton-wool spots, and conjunctival blood-flow sludging was determined. Additional predictors were CD4+ count, age, time on aerosolized pentamidine, time on zidovudine, and Walter Reed staging. The relative influence of each predictor was calculated by stepwise multiple regression analysis (inclusion criterion; incremental P value = < 0.05) using data for the right eyes (RE). The results were validated by using data for the left eyes (LE) and both eyes (BE). The only included predictors in multiple regression analyses for the RE were number of cotton-wool spots (tritan: R = .70; deutan: R = .46; and protan: R = .58; P < .0001 for all axes) and age (tritan: increment of R [Ri] = .05, P = .002; deutan: Ri = .10, P = .004; and protan: Ri = .05, P = .002). The predictors time on zidovudine (Ri = .05, P = .002) and Walter Reed staging (Ri = .03, P = .01) were additionally included in multiple regression analysis for tritan LE. The results for deutan LE were comparable to those for the RE. In the analysis for protan LE, the only included predictor was number of cotton-wool spots. In the analyses for BE, no further predictors were included. The predictors Walter Reed staging and CD4+ count showed a significant association with all three criteria in univariate analysis. Additionally, tritan CCS was significantly associated with conjunctival blood-flow sludging. CCS deficits in patients with HIV disease are primarily associated with the number of cotton-wool spots. Results of this study are in accordance with the hypothesis that CCS deficits are in a relevant part caused by neuroretinal damage secondary to HIV-related microangiopathy.
Principal component regression analysis with SPSS.
Liu, R X; Kuang, J; Gong, Q; Hou, X L
2003-06-01
The paper introduces all indices of multicollinearity diagnoses, the basic principle of principal component regression and determination of 'best' equation method. The paper uses an example to describe how to do principal component regression analysis with SPSS 10.0: including all calculating processes of the principal component regression and all operations of linear regression, factor analysis, descriptives, compute variable and bivariate correlations procedures in SPSS 10.0. The principal component regression analysis can be used to overcome disturbance of the multicollinearity. The simplified, speeded up and accurate statistical effect is reached through the principal component regression analysis with SPSS.
Patient satisfaction in Dental Healthcare Centers.
Ali, Dena A
2016-01-01
This study aimed to (1) measure the degree of patient satisfaction among the clinical and nonclinical dental services offered at specialty dental centers and (2) investigate the factors associated with the degree of overall satisfaction. Four hundred and ninety-seven participants from five dental centers were recruited for this study. Each participant completed a self-administered questionnaire to measure patient satisfaction with clinical and nonclinical dental services. Analysis of variance, t-tests, a general linear model, and stepwise regression analysis was applied. The respondents were generally satisfied, but internal differences were observed. The exhibited highest satisfaction with the dentists' performance, followed by the dental assistants' services, and the lowest satisfaction with the center's physical appearance and accessibility. Females, participants with less than a bachelor's degree, and younger individuals were more satisfied with the clinical and nonclinical dental services. The stepwise regression analysis revealed that the coefficient of determination (R (2)) was 40.4%. The patient satisfaction with the performance of the dentists explained 42.6% of the overall satisfaction, whereas their satisfaction with the clinical setting explained 31.5% of the overall satisfaction. Additional improvements with regard to the accessibility and physical appearance of the dental centers are needed. In addition, interventions regarding accessibility, particularly when booking an appointment, are required.
Online Statistical Modeling (Regression Analysis) for Independent Responses
NASA Astrophysics Data System (ADS)
Made Tirta, I.; Anggraeni, Dian; Pandutama, Martinus
2017-06-01
Regression analysis (statistical analmodelling) are among statistical methods which are frequently needed in analyzing quantitative data, especially to model relationship between response and explanatory variables. Nowadays, statistical models have been developed into various directions to model various type and complex relationship of data. Rich varieties of advanced and recent statistical modelling are mostly available on open source software (one of them is R). However, these advanced statistical modelling, are not very friendly to novice R users, since they are based on programming script or command line interface. Our research aims to developed web interface (based on R and shiny), so that most recent and advanced statistical modelling are readily available, accessible and applicable on web. We have previously made interface in the form of e-tutorial for several modern and advanced statistical modelling on R especially for independent responses (including linear models/LM, generalized linier models/GLM, generalized additive model/GAM and generalized additive model for location scale and shape/GAMLSS). In this research we unified them in the form of data analysis, including model using Computer Intensive Statistics (Bootstrap and Markov Chain Monte Carlo/ MCMC). All are readily accessible on our online Virtual Statistics Laboratory. The web (interface) make the statistical modeling becomes easier to apply and easier to compare them in order to find the most appropriate model for the data.
Maurya, Rakesh Kumar; Saxena, Mohit Raj; Rai, Piyush; Bhardwaj, Aashish
2018-05-01
Currently, diesel engines are more preferred over gasoline engines due to their higher torque output and fuel economy. However, diesel engines confront major challenge of meeting the future stringent emission norms (especially soot particle emissions) while maintaining the same fuel economy. In this study, nanosize range soot particle emission characteristics of a stationary (non-road) diesel engine have been experimentally investigated. Experiments are conducted at a constant speed of 1500 rpm for three compression ratios and nozzle opening pressures at different engine loads. In-cylinder pressure history for 2000 consecutive engine cycles is recorded and averaged data is used for analysis of combustion characteristics. An electrical mobility-based fast particle sizer is used for analyzing particle size and mass distributions of engine exhaust particles at different test conditions. Soot particle distribution from 5 to 1000 nm was recorded. Results show that total particle concentration decreases with an increase in engine operating loads. Moreover, the addition of butanol in the diesel fuel leads to the reduction in soot particle concentration. Regression analysis was also conducted to derive a correlation between combustion parameters and particle number emissions for different compression ratios. Regression analysis shows a strong correlation between cylinder pressure-based combustion parameters and particle number emission.
Patient satisfaction in Dental Healthcare Centers
Ali, Dena A.
2016-01-01
Objectives: This study aimed to (1) measure the degree of patient satisfaction among the clinical and nonclinical dental services offered at specialty dental centers and (2) investigate the factors associated with the degree of overall satisfaction. Materials and Methods: Four hundred and ninety-seven participants from five dental centers were recruited for this study. Each participant completed a self-administered questionnaire to measure patient satisfaction with clinical and nonclinical dental services. Analysis of variance, t-tests, a general linear model, and stepwise regression analysis was applied. Results: The respondents were generally satisfied, but internal differences were observed. The exhibited highest satisfaction with the dentists’ performance, followed by the dental assistants’ services, and the lowest satisfaction with the center's physical appearance and accessibility. Females, participants with less than a bachelor's degree, and younger individuals were more satisfied with the clinical and nonclinical dental services. The stepwise regression analysis revealed that the coefficient of determination (R2) was 40.4%. The patient satisfaction with the performance of the dentists explained 42.6% of the overall satisfaction, whereas their satisfaction with the clinical setting explained 31.5% of the overall satisfaction. Conclusion: Additional improvements with regard to the accessibility and physical appearance of the dental centers are needed. In addition, interventions regarding accessibility, particularly when booking an appointment, are required. PMID:27403045
Stuart P. Cottrell; Alan R. Graefe
1995-01-01
This paper examines predictors of boater behavior in a specific behavior situation, namely the percentage of raw sewage discharged from recreational vessels in a sanitation pumpout facility on the Chesapeake Bay. Results of a multiple regression analysis show knowledge predicts behavior in specific issue situations. In addition, the more specific the...
ERIC Educational Resources Information Center
Maschi, Tina
2006-01-01
This study examined how the cumulative (additive) versus differential (individual) effects of trauma influenced male delinquency. Using a comprehensive measure of trauma, a secondary data analysis was conducted on a nationally representative sample of male youths between the ages of 12 and 17. Logistic regression analyses revealed that all three…
Faculty Salary Equity: Issues in Regression Model Selection. AIR 1992 Annual Forum Paper.
ERIC Educational Resources Information Center
Moore, Nelle
This paper discusses the determination of college faculty salary inequity and identifies the areas in which human judgment must be used in order to conduct a statistical analysis of salary equity. In addition, it provides some informed guidelines for making those judgments. The paper provides a framework for selecting salary equity models, based…
Accounting for Slipping and Other False Negatives in Logistic Models of Student Learning
ERIC Educational Resources Information Center
MacLellan, Christopher J.; Liu, Ran; Koedinger, Kenneth R.
2015-01-01
Additive Factors Model (AFM) and Performance Factors Analysis (PFA) are two popular models of student learning that employ logistic regression to estimate parameters and predict performance. This is in contrast to Bayesian Knowledge Tracing (BKT) which uses a Hidden Markov Model formalism. While all three models tend to make similar predictions,…
Online Courses Assessment through Measuring and Archetyping of Usage Data
ERIC Educational Resources Information Center
Kazanidis, Ioannis; Theodosiou, Theodosios; Petasakis, Ioannis; Valsamidis, Stavros
2016-01-01
Database files and additional log files of Learning Management Systems (LMSs) contain an enormous volume of data which usually remain unexploited. A new methodology is proposed in order to analyse these data both on the level of both the courses and the learners. Specifically, "regression analysis" is proposed as a first step in the…
ERIC Educational Resources Information Center
Daberkow, Kevin S.; Lin, Wei
2012-01-01
Nearly half a century of lottery scholarship has measured lottery tax incidence predominantly through either the Suits Index or regression analysis. The present study builds on historic lottery tax burden measurement to present a comprehensive set of tools to determine the tax incidence of individual games in addition to determining which lottery…
Wang, Peijie; Zhao, Hui; Sun, Jianguo
2016-12-01
Interval-censored failure time data occur in many fields such as demography, economics, medical research, and reliability and many inference procedures on them have been developed (Sun, 2006; Chen, Sun, and Peace, 2012). However, most of the existing approaches assume that the mechanism that yields interval censoring is independent of the failure time of interest and it is clear that this may not be true in practice (Zhang et al., 2007; Ma, Hu, and Sun, 2015). In this article, we consider regression analysis of case K interval-censored failure time data when the censoring mechanism may be related to the failure time of interest. For the problem, an estimated sieve maximum-likelihood approach is proposed for the data arising from the proportional hazards frailty model and for estimation, a two-step procedure is presented. In the addition, the asymptotic properties of the proposed estimators of regression parameters are established and an extensive simulation study suggests that the method works well. Finally, we apply the method to a set of real interval-censored data that motivated this study. © 2016, The International Biometric Society.
Caballero, Daniel; Antequera, Teresa; Caro, Andrés; Ávila, María Del Mar; G Rodríguez, Pablo; Perez-Palacios, Trinidad
2017-07-01
Magnetic resonance imaging (MRI) combined with computer vision techniques have been proposed as an alternative or complementary technique to determine the quality parameters of food in a non-destructive way. The aim of this work was to analyze the sensory attributes of dry-cured loins using this technique. For that, different MRI acquisition sequences (spin echo, gradient echo and turbo 3D), algorithms for MRI analysis (GLCM, NGLDM, GLRLM and GLCM-NGLDM-GLRLM) and predictive data mining techniques (multiple linear regression and isotonic regression) were tested. The correlation coefficient (R) and mean absolute error (MAE) were used to validate the prediction results. The combination of spin echo, GLCM and isotonic regression produced the most accurate results. In addition, the MRI data from dry-cured loins seems to be more suitable than the data from fresh loins. The application of predictive data mining techniques on computational texture features from the MRI data of loins enables the determination of the sensory traits of dry-cured loins in a non-destructive way. © 2016 Society of Chemical Industry. © 2016 Society of Chemical Industry.
[In vitro testing of yeast resistance to antimycotic substances].
Potel, J; Arndt, K
1982-01-01
Investigations have been carried out in order to clarify the antibiotic susceptibility determination of yeasts. 291 yeast strains of different species were tested for sensitivity to 7 antimycotics: amphotericin B, flucytosin, nystatin, pimaricin, clotrimazol, econazol and miconazol. Additionally to the evaluation of inhibition zone diameters and MIC-values the influence of pH was examined. 1. The dependence of inhibition zone diameters upon pH-values varies due to the antimycotic tested. For standardizing purposes the pH 6.0 is proposed; moreover, further experimental parameters, such as nutrient composition, agar depth, cell density, incubation time and -temperature, have to be normed. 2. The relation between inhibition zone size and logarythmic MIC does not fit a linear regression analysis when all species are considered together. Therefore regression functions have to be calculated selecting the individual species. In case of the antimycotics amphotericin B, nystatin and pimaricin the low scattering of the MIC-values does not allow regression analysis. 3. A quantitative susceptibility determination of yeasts--particularly to the fungistatical substances with systemic applicability, flucytosin and miconazol, -- is advocated by the results of the MIC-tests.
On the use of log-transformation vs. nonlinear regression for analyzing biological power laws
Xiao, X.; White, E.P.; Hooten, M.B.; Durham, S.L.
2011-01-01
Power-law relationships are among the most well-studied functional relationships in biology. Recently the common practice of fitting power laws using linear regression (LR) on log-transformed data has been criticized, calling into question the conclusions of hundreds of studies. It has been suggested that nonlinear regression (NLR) is preferable, but no rigorous comparison of these two methods has been conducted. Using Monte Carlo simulations, we demonstrate that the error distribution determines which method performs better, with NLR better characterizing data with additive, homoscedastic, normal error and LR better characterizing data with multiplicative, heteroscedastic, lognormal error. Analysis of 471 biological power laws shows that both forms of error occur in nature. While previous analyses based on log-transformation appear to be generally valid, future analyses should choose methods based on a combination of biological plausibility and analysis of the error distribution. We provide detailed guidelines and associated computer code for doing so, including a model averaging approach for cases where the error structure is uncertain. ?? 2011 by the Ecological Society of America.
The status of diabetes control in Kurdistan province, west of Iran.
Esmailnasab, Nader; Afkhamzadeh, Abdorrahim; Roshani, Daem; Moradi, Ghobad
2013-09-17
Based on some estimation more than two million peoples in Iran are affected by Type 2 diabetes. The present study was designed to evaluate the status of diabetes control among Type 2 diabetes patients in Kurdistan, west of Iran and its associated factors. In our cross sectional study conducted in 2010, 411 Type 2 diabetes patients were randomly recruited from Sanandaj, Capital of Kurdistan. Chi square test was used in univariate analysis to address the association between HgAlc and FBS status and other variables. The significant results from Univariate analysis were entered in multivariate analysis and multinomial logistic regression model. In 38% of patients, FBS was in normal range (70-130) and in 47% HgA1c was <7% which is normal range for HgA1c. In univariate analysis, FBS level was associated with educational levels (P=0.001), referral style (P=0.001), referral time (P=0.009), and insulin injection (P=0.016). In addition, HgA1c had a relationship with sex (P=0.023), age (P=0.035), education (P=0.001), referral style (P=0.001), and insulin injection (P=0.008). After using multinomial logistic regression for significant results of univariate analysis, it was found that FBS was significantly associated with referral style. In addition HgA1c was significantly associated with referral style and Insulin injection. Although some of patients were under the coverage of specialized cares, but their diabetes were not properly controlled.
NASA Technical Reports Server (NTRS)
Barrett, C. A.
1985-01-01
Multiple linear regression analysis was used to determine an equation for estimating hot corrosion attack for a series of Ni base cast turbine alloys. The U transform (i.e., 1/sin (% A/100) to the 1/2) was shown to give the best estimate of the dependent variable, y. A complete second degree equation is described for the centered" weight chemistries for the elements Cr, Al, Ti, Mo, W, Cb, Ta, and Co. In addition linear terms for the minor elements C, B, and Zr were added for a basic 47 term equation. The best reduced equation was determined by the stepwise selection method with essentially 13 terms. The Cr term was found to be the most important accounting for 60 percent of the explained variability hot corrosion attack.
A Selective Review of Group Selection in High-Dimensional Models
Huang, Jian; Breheny, Patrick; Ma, Shuangge
2013-01-01
Grouping structures arise naturally in many statistical modeling problems. Several methods have been proposed for variable selection that respect grouping structure in variables. Examples include the group LASSO and several concave group selection methods. In this article, we give a selective review of group selection concerning methodological developments, theoretical properties and computational algorithms. We pay particular attention to group selection methods involving concave penalties. We address both group selection and bi-level selection methods. We describe several applications of these methods in nonparametric additive models, semiparametric regression, seemingly unrelated regressions, genomic data analysis and genome wide association studies. We also highlight some issues that require further study. PMID:24174707
NASA Astrophysics Data System (ADS)
Tan, C. H.; Matjafri, M. Z.; Lim, H. S.
2015-10-01
This paper presents the prediction models which analyze and compute the CO2 emission in Malaysia. Each prediction model for CO2 emission will be analyzed based on three main groups which is transportation, electricity and heat production as well as residential buildings and commercial and public services. The prediction models were generated using data obtained from World Bank Open Data. Best subset method will be used to remove irrelevant data and followed by multi linear regression to produce the prediction models. From the results, high R-square (prediction) value was obtained and this implies that the models are reliable to predict the CO2 emission by using specific data. In addition, the CO2 emissions from these three groups are forecasted using trend analysis plots for observation purpose.
Regression Analysis by Example. 5th Edition
ERIC Educational Resources Information Center
Chatterjee, Samprit; Hadi, Ali S.
2012-01-01
Regression analysis is a conceptually simple method for investigating relationships among variables. Carrying out a successful application of regression analysis, however, requires a balance of theoretical results, empirical rules, and subjective judgment. "Regression Analysis by Example, Fifth Edition" has been expanded and thoroughly…
Protective Effect of HLA-DQB1 Alleles Against Alloimmunization in Patients with Sickle Cell Disease
Tatari-Calderone, Zohreh; Gordish-Dressman, Heather; Fasano, Ross; Riggs, Michael; Fortier, Catherine; Andrew; Campbell, D.; Charron, Dominique; Gordeuk, Victor R.; Luban, Naomi L.C.; Vukmanovic, Stanislav; Tamouza, Ryad
2015-01-01
Background Alloimmunization or the development of alloantibodies to Red Blood Cell (RBC) antigens is considered one of the major complications after RBC transfusions in patients with sickle cell disease (SCD) and can lead to both acute and delayed hemolytic reactions. It has been suggested that polymorphisms in HLA genes, may play a role in alloimmunization. We conducted a retrospective study analyzing the influence of HLA-DRB1 and DQB1 genetic diversity on RBC-alloimmunization. Study design Two-hundred four multi-transfused SCD patients with and without RBC-alloimmunization were typed at low/medium resolution by PCR-SSO, using IMGT-HLA Database. HLA-DRB1 and DQB1 allele frequencies were analyzed using logistic regression models, and global p-value was calculated using multiple logistic regression. Results While only trends towards associations between HLA-DR diversity and alloimmunization were observed, analysis of HLA-DQ showed that HLA-DQ2 (p=0.02), -DQ3 (p=0.02) and -DQ5 (p=0.01) alleles were significantly higher in non-alloimmunized patients, likely behaving as protective alleles. In addition, multiple logistic regression analysis showed both HLA-DQ2/6 (p=0.01) and HLA-DQ5/5 (p=0.03) combinations constitute additional predictor of protective status. Conclusion Our data suggest that particular HLA-DQ alleles influence the clinical course of RBC transfusion in patients with SCD, which could pave the way towards predictive strategies. PMID:26476208
Seligman, D A; Pullinger, A G
2000-01-01
Confusion about the relationship of occlusion to temporomandibular disorders (TMD) persists. This study attempted to identify occlusal and attrition factors plus age that would characterize asymptomatic normal female subjects. A total of 124 female patients with intracapsular TMD were compared with 47 asymptomatic female controls for associations to 9 occlusal factors, 3 attrition severity measures, and age using classification tree, multiple stepwise logistic regression, and univariate analyses. Models were tested for accuracy (sensitivity and specificity) and total contribution to the variance. The classification tree model had 4 terminal nodes that used only anterior attrition and age. "Normals" were mainly characterized by low attrition levels, whereas patients had higher attrition and tended to be younger. The tree model was only moderately useful (sensitivity 63%, specificity 94%) in predicting normals. The logistic regression model incorporated unilateral posterior crossbite and mediotrusive attrition severity in addition to the 2 factors in the tree, but was slightly less accurate than the tree (sensitivity 51%, specificity 90%). When only occlusal factors were considered in the analysis, normals were additionally characterized by a lack of anterior open bite, smaller overjet, and smaller RCP-ICP slides. The log likelihood accounted for was similar for both the tree (pseudo R(2) = 29.38%; mean deviance = 0.95) and the multiple logistic regression (Cox Snell R(2) = 30.3%, mean deviance = 0.84) models. The occlusal and attrition factors studied were only moderately useful in differentiating normals from TMD patients.
Curcic, Marijana; Buha, Aleksandra; Stankovic, Sanja; Milovanovic, Vesna; Bulat, Zorica; Đukić-Ćosić, Danijela; Antonijević, Evica; Vučinić, Slavica; Matović, Vesna; Antonijevic, Biljana
2017-02-01
The objective of this study was to assess toxicity of Cd and BDE-209 mixture on haematological parameters in subacutely exposed rats and to determine the presence and type of interactions between these two chemicals using multiple factorial regression analysis. Furthermore, for the assessment of interaction type, an isobologram based methodology was applied and compared with multiple factorial regression analysis. Chemicals were given by oral gavage to the male Wistar rats weighing 200-240g for 28days. Animals were divided in 16 groups (8/group): control vehiculum group, three groups of rats were treated with 2.5, 7.5 or 15mg Cd/kg/day. These doses were chosen on the bases of literature data and reflect relatively high Cd environmental exposure, three groups of rats were treated with 1000, 2000 or 4000mg BDE-209/kg/bw/day, doses proved to induce toxic effects in rats. Furthermore, nine groups of animals were treated with different mixtures of Cd and BDE-209 containing doses of Cd and BDE-209 stated above. Blood samples were taken at the end of experiment and red blood cells, white blood cells and platelets counts were determined. For interaction assessment multiple factorial regression analysis and fitted isobologram approach were used. In this study, we focused on multiple factorial regression analysis as a method for interaction assessment. We also investigated the interactions between Cd and BDE-209 by the derived model for the description of the obtained fitted isobologram curves. Current study indicated that co-exposure to Cd and BDE-209 can result in significant decrease in RBC count, increase in WBC count and decrease in PLT count, when compared with controls. Multiple factorial regression analysis used for the assessment of interactions type between Cd and BDE-209 indicated synergism for the effect on RBC count and no interactions i.e. additivity for the effects on WBC and PLT counts. On the other hand, isobologram based approach showed slight antagonism for the effects on RBC and WBC while no interactions were proved for the joint effect on PLT count. These results confirm that the assessment of interactions between chemicals in the mixture greatly depends on the concept or method used for this evaluation. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Guisan, Antoine; Edwards, T.C.; Hastie, T.
2002-01-01
An important statistical development of the last 30 years has been the advance in regression analysis provided by generalized linear models (GLMs) and generalized additive models (GAMs). Here we introduce a series of papers prepared within the framework of an international workshop entitled: Advances in GLMs/GAMs modeling: from species distribution to environmental management, held in Riederalp, Switzerland, 6-11 August 2001. We first discuss some general uses of statistical models in ecology, as well as provide a short review of several key examples of the use of GLMs and GAMs in ecological modeling efforts. We next present an overview of GLMs and GAMs, and discuss some of their related statistics used for predictor selection, model diagnostics, and evaluation. Included is a discussion of several new approaches applicable to GLMs and GAMs, such as ridge regression, an alternative to stepwise selection of predictors, and methods for the identification of interactions by a combined use of regression trees and several other approaches. We close with an overview of the papers and how we feel they advance our understanding of their application to ecological modeling. ?? 2002 Elsevier Science B.V. All rights reserved.
Lee, Soo Yee; Mediani, Ahmed; Maulidiani, Maulidiani; Khatib, Alfi; Ismail, Intan Safinar; Zawawi, Norhasnida; Abas, Faridah
2018-01-01
Neptunia oleracea is a plant consumed as a vegetable and which has been used as a folk remedy for several diseases. Herein, two regression models (partial least squares, PLS; and random forest, RF) in a metabolomics approach were compared and applied to the evaluation of the relationship between phenolics and bioactivities of N. oleracea. In addition, the effects of different extraction conditions on the phenolic constituents were assessed by pattern recognition analysis. Comparison of the PLS and RF showed that RF exhibited poorer generalization and hence poorer predictive performance. Both the regression coefficient of PLS and the variable importance of RF revealed that quercetin and kaempferol derivatives, caffeic acid and vitexin-2-O-rhamnoside were significant towards the tested bioactivities. Furthermore, principal component analysis (PCA) and partial least squares-discriminant analysis (PLS-DA) results showed that sonication and absolute ethanol are the preferable extraction method and ethanol ratio, respectively, to produce N. oleracea extracts with high phenolic levels and therefore high DPPH scavenging and α-glucosidase inhibitory activities. Both PLS and RF are useful regression models in metabolomics studies. This work provides insight into the performance of different multivariate data analysis tools and the effects of different extraction conditions on the extraction of desired phenolics from plants. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry.
Enhancing the estimation of blood pressure using pulse arrival time and two confounding factors.
Baek, Hyun Jae; Kim, Ko Keun; Kim, Jung Soo; Lee, Boreom; Park, Kwang Suk
2010-02-01
A new method of blood pressure (BP) estimation using multiple regression with pulse arrival time (PAT) and two confounding factors was evaluated in clinical and unconstrained monitoring situations. For the first analysis with clinical data, electrocardiogram (ECG), photoplethysmogram (PPG) and invasive BP signals were obtained by a conventional patient monitoring device during surgery. In the second analysis, ECG, PPG and non-invasive BP were measured using systems developed to obtain data under conditions in which the subject was not constrained. To enhance the performance of BP estimation methods, heart rate (HR) and arterial stiffness were considered as confounding factors in regression analysis. The PAT and HR were easily extracted from ECG and PPG signals. For arterial stiffness, the duration from the maximum derivative point to the maximum of the dicrotic notch in the PPG signal, a parameter called TDB, was employed. In two experiments that normally cause BP variation, the correlation between measured BP and the estimated BP was investigated. Multiple-regression analysis with the two confounding factors improved correlation coefficients for diastolic blood pressure and systolic blood pressure to acceptable confidence levels, compared to existing methods that consider PAT only. In addition, reproducibility for the proposed method was determined using constructed test sets. Our results demonstrate that non-invasive, non-intrusive BP estimation can be obtained using methods that can be applied in both clinical and daily healthcare situations.
Yang, Shun-hua; Zhang, Hai-tao; Guo, Long; Ren, Yan
2015-06-01
Relative elevation and stream power index were selected as auxiliary variables based on correlation analysis for mapping soil organic matter. Geographically weighted regression Kriging (GWRK) and regression Kriging (RK) were used for spatial interpolation of soil organic matter and compared with ordinary Kriging (OK), which acts as a control. The results indicated that soil or- ganic matter was significantly positively correlated with relative elevation whilst it had a significantly negative correlation with stream power index. Semivariance analysis showed that both soil organic matter content and its residuals (including ordinary least square regression residual and GWR resi- dual) had strong spatial autocorrelation. Interpolation accuracies by different methods were esti- mated based on a data set of 98 validation samples. Results showed that the mean error (ME), mean absolute error (MAE) and root mean square error (RMSE) of RK were respectively 39.2%, 17.7% and 20.6% lower than the corresponding values of OK, with a relative-improvement (RI) of 20.63. GWRK showed a similar tendency, having its ME, MAE and RMSE to be respectively 60.6%, 23.7% and 27.6% lower than those of OK, with a RI of 59.79. Therefore, both RK and GWRK significantly improved the accuracy of OK interpolation of soil organic matter due to their in- corporation of auxiliary variables. In addition, GWRK performed obviously better than RK did in this study, and its improved performance should be attributed to the consideration of sample spatial locations.
Parametrically Guided Generalized Additive Models with Application to Mergers and Acquisitions Data
Fan, Jianqing; Maity, Arnab; Wang, Yihui; Wu, Yichao
2012-01-01
Generalized nonparametric additive models present a flexible way to evaluate the effects of several covariates on a general outcome of interest via a link function. In this modeling framework, one assumes that the effect of each of the covariates is nonparametric and additive. However, in practice, often there is prior information available about the shape of the regression functions, possibly from pilot studies or exploratory analysis. In this paper, we consider such situations and propose an estimation procedure where the prior information is used as a parametric guide to fit the additive model. Specifically, we first posit a parametric family for each of the regression functions using the prior information (parametric guides). After removing these parametric trends, we then estimate the remainder of the nonparametric functions using a nonparametric generalized additive model, and form the final estimates by adding back the parametric trend. We investigate the asymptotic properties of the estimates and show that when a good guide is chosen, the asymptotic variance of the estimates can be reduced significantly while keeping the asymptotic variance same as the unguided estimator. We observe the performance of our method via a simulation study and demonstrate our method by applying to a real data set on mergers and acquisitions. PMID:23645976
Parametrically Guided Generalized Additive Models with Application to Mergers and Acquisitions Data.
Fan, Jianqing; Maity, Arnab; Wang, Yihui; Wu, Yichao
2013-01-01
Generalized nonparametric additive models present a flexible way to evaluate the effects of several covariates on a general outcome of interest via a link function. In this modeling framework, one assumes that the effect of each of the covariates is nonparametric and additive. However, in practice, often there is prior information available about the shape of the regression functions, possibly from pilot studies or exploratory analysis. In this paper, we consider such situations and propose an estimation procedure where the prior information is used as a parametric guide to fit the additive model. Specifically, we first posit a parametric family for each of the regression functions using the prior information (parametric guides). After removing these parametric trends, we then estimate the remainder of the nonparametric functions using a nonparametric generalized additive model, and form the final estimates by adding back the parametric trend. We investigate the asymptotic properties of the estimates and show that when a good guide is chosen, the asymptotic variance of the estimates can be reduced significantly while keeping the asymptotic variance same as the unguided estimator. We observe the performance of our method via a simulation study and demonstrate our method by applying to a real data set on mergers and acquisitions.
An example of complex modelling in dentistry using Markov chain Monte Carlo (MCMC) simulation.
Helfenstein, Ulrich; Menghini, Giorgio; Steiner, Marcel; Murati, Francesca
2002-09-01
In the usual regression setting one regression line is computed for a whole data set. In a more complex situation, each person may be observed for example at several points in time and thus a regression line might be calculated for each person. Additional complexities, such as various forms of errors in covariables may make a straightforward statistical evaluation difficult or even impossible. During recent years methods have been developed allowing convenient analysis of problems where the data and the corresponding models show these and many other forms of complexity. The methodology makes use of a Bayesian approach and Markov chain Monte Carlo (MCMC) simulations. The methods allow the construction of increasingly elaborate models by building them up from local sub-models. The essential structure of the models can be represented visually by directed acyclic graphs (DAG). This attractive property allows communication and discussion of the essential structure and the substantial meaning of a complex model without needing algebra. After presentation of the statistical methods an example from dentistry is presented in order to demonstrate their application and use. The dataset of the example had a complex structure; each of a set of children was followed up over several years. The number of new fillings in permanent teeth had been recorded at several ages. The dependent variables were markedly different from the normal distribution and could not be transformed to normality. In addition, explanatory variables were assumed to be measured with different forms of error. Illustration of how the corresponding models can be estimated conveniently via MCMC simulation, in particular, 'Gibbs sampling', using the freely available software BUGS is presented. In addition, how the measurement error may influence the estimates of the corresponding coefficients is explored. It is demonstrated that the effect of the independent variable on the dependent variable may be markedly underestimated if the measurement error is not taken into account ('regression dilution bias'). Markov chain Monte Carlo methods may be of great value to dentists in allowing analysis of data sets which exhibit a wide range of different forms of complexity.
Padula, William V; Mishra, Manish K; Weaver, Christopher D; Yilmaz, Taygan; Splaine, Mark E
2012-06-01
To demonstrate complementary results of regression and statistical process control (SPC) chart analyses for hospital-acquired pressure ulcers (HAPUs), and identify possible links between changes and opportunities for improvement between hospital microsystems and macrosystems. Ordinary least squares and panel data regression of retrospective hospital billing data, and SPC charts of prospective patient records for a US tertiary-care facility (2004-2007). A prospective cohort of hospital inpatients at risk for HAPUs was the study population. There were 337 HAPU incidences hospital wide among 43 844 inpatients. A probit regression model predicted the correlation of age, gender and length of stay on HAPU incidence (pseudo R(2)=0.096). Panel data analysis determined that for each additional day in the hospital, there was a 0.28% increase in the likelihood of HAPU incidence. A p-chart of HAPU incidence showed a mean incidence rate of 1.17% remaining in statistical control. A t-chart showed the average time between events for the last 25 HAPUs was 13.25 days. There was one 57-day period between two incidences during the observation period. A p-chart addressing Braden scale assessments showed that 40.5% of all patients were risk stratified for HAPUs upon admission. SPC charts complement standard regression analysis. SPC amplifies patient outcomes at the microsystem level and is useful for guiding quality improvement. Macrosystems should monitor effective quality improvement initiatives in microsystems and aid the spread of successful initiatives to other microsystems, followed by system-wide analysis with regression. Although HAPU incidence in this study is below the national mean, there is still room to improve HAPU incidence in this hospital setting since 0% incidence is theoretically achievable. Further assessment of pressure ulcer incidence could illustrate improvement in the quality of care and prevent HAPUs.
Regression analysis of mixed panel count data with dependent terminal events.
Yu, Guanglei; Zhu, Liang; Li, Yang; Sun, Jianguo; Robison, Leslie L
2017-05-10
Event history studies are commonly conducted in many fields, and a great deal of literature has been established for the analysis of the two types of data commonly arising from these studies: recurrent event data and panel count data. The former arises if all study subjects are followed continuously, while the latter means that each study subject is observed only at discrete time points. In reality, a third type of data, a mixture of the two types of the data earlier, may occur and furthermore, as with the first two types of the data, there may exist a dependent terminal event, which may preclude the occurrences of recurrent events of interest. This paper discusses regression analysis of mixed recurrent event and panel count data in the presence of a terminal event and an estimating equation-based approach is proposed for estimation of regression parameters of interest. In addition, the asymptotic properties of the proposed estimator are established, and a simulation study conducted to assess the finite-sample performance of the proposed method suggests that it works well in practical situations. Finally, the methodology is applied to a childhood cancer study that motivated this study. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
Tangen, C M; Koch, G G
1999-03-01
In the randomized clinical trial setting, controlling for covariates is expected to produce variance reduction for the treatment parameter estimate and to adjust for random imbalances of covariates between the treatment groups. However, for the logistic regression model, variance reduction is not obviously obtained. This can lead to concerns about the assumptions of the logistic model. We introduce a complementary nonparametric method for covariate adjustment. It provides results that are usually compatible with expectations for analysis of covariance. The only assumptions required are based on randomization and sampling arguments. The resulting treatment parameter is a (unconditional) population average log-odds ratio that has been adjusted for random imbalance of covariates. Data from a randomized clinical trial are used to compare results from the traditional maximum likelihood logistic method with those from the nonparametric logistic method. We examine treatment parameter estimates, corresponding standard errors, and significance levels in models with and without covariate adjustment. In addition, we discuss differences between unconditional population average treatment parameters and conditional subpopulation average treatment parameters. Additional features of the nonparametric method, including stratified (multicenter) and multivariate (multivisit) analyses, are illustrated. Extensions of this methodology to the proportional odds model are also made.
Park, Ji Nam; Han, Mi Ah; Park, Jong; Ryu, So Yeon
2016-04-14
The aim of this study was to analyze the association between general working conditions and depressive symptoms among Korean employees. The target population of the study was native employees nationwide who were at least 15 years old, and 50,032 such individuals were enrolled in the study. Depressive symptoms was assessed using the WHO-5 wellbeing index. Associations between general characteristics, job-related characteristics, work environment, and depressive symptoms were tested using chi-square tests, t-tests, and multiple logistic regression analysis. The prevalence of depressive symptoms was 39% (40.7% in males and 36.5% in females). Multiple regression analysis revealed that male subjects, older subjects, subjects with higher education status, subjects with lower monthly income, current smokers, and frequent drinkers were more likely to have depressive symptoms. In addition, longer weekly work hours, occupation type (skilled, unskilled, operative, or economic sector), shift work, working to tight deadlines, exposure to stress at work, and hazard exposure were associated with depressive symptoms. This representative study will be a guide to help manage depression among Korean employees. We expect that further research will identify additional causal relationships between general or specific working conditions and depression.
Practical aspects of estimating energy components in rodents
van Klinken, Jan B.; van den Berg, Sjoerd A. A.; van Dijk, Ko Willems
2013-01-01
Recently there has been an increasing interest in exploiting computational and statistical techniques for the purpose of component analysis of indirect calorimetry data. Using these methods it becomes possible to dissect daily energy expenditure into its components and to assess the dynamic response of the resting metabolic rate (RMR) to nutritional and pharmacological manipulations. To perform robust component analysis, however, is not straightforward and typically requires the tuning of parameters and the preprocessing of data. Moreover the degree of accuracy that can be attained by these methods depends on the configuration of the system, which must be properly taken into account when setting up experimental studies. Here, we review the methods of Kalman filtering, linear, and penalized spline regression, and minimal energy expenditure estimation in the context of component analysis and discuss their results on high resolution datasets from mice and rats. In addition, we investigate the effect of the sample time, the accuracy of the activity sensor, and the washout time of the chamber on the estimation accuracy. We found that on the high resolution data there was a strong correlation between the results of Kalman filtering and penalized spline (P-spline) regression, except for the activity respiratory quotient (RQ). For low resolution data the basal metabolic rate (BMR) and resting RQ could still be estimated accurately with P-spline regression, having a strong correlation with the high resolution estimate (R2 > 0.997; sample time of 9 min). In contrast, the thermic effect of food (TEF) and activity related energy expenditure (AEE) were more sensitive to a reduction in the sample rate (R2 > 0.97). In conclusion, for component analysis on data generated by single channel systems with continuous data acquisition both Kalman filtering and P-spline regression can be used, while for low resolution data from multichannel systems P-spline regression gives more robust results. PMID:23641217
Chirombo, James; Lowe, Rachel; Kazembe, Lawrence
2014-01-01
Background After years of implementing Roll Back Malaria (RBM) interventions, the changing landscape of malaria in terms of risk factors and spatial pattern has not been fully investigated. This paper uses the 2010 malaria indicator survey data to investigate if known malaria risk factors remain relevant after many years of interventions. Methods We adopted a structured additive logistic regression model that allowed for spatial correlation, to more realistically estimate malaria risk factors. Our model included child and household level covariates, as well as climatic and environmental factors. Continuous variables were modelled by assuming second order random walk priors, while spatial correlation was specified as a Markov random field prior, with fixed effects assigned diffuse priors. Inference was fully Bayesian resulting in an under five malaria risk map for Malawi. Results Malaria risk increased with increasing age of the child. With respect to socio-economic factors, the greater the household wealth, the lower the malaria prevalence. A general decline in malaria risk was observed as altitude increased. Minimum temperatures and average total rainfall in the three months preceding the survey did not show a strong association with disease risk. Conclusions The structured additive regression model offered a flexible extension to standard regression models by enabling simultaneous modelling of possible nonlinear effects of continuous covariates, spatial correlation and heterogeneity, while estimating usual fixed effects of categorical and continuous observed variables. Our results confirmed that malaria epidemiology is a complex interaction of biotic and abiotic factors, both at the individual, household and community level and that risk factors are still relevant many years after extensive implementation of RBM activities. PMID:24991915
Chirombo, James; Lowe, Rachel; Kazembe, Lawrence
2014-01-01
After years of implementing Roll Back Malaria (RBM) interventions, the changing landscape of malaria in terms of risk factors and spatial pattern has not been fully investigated. This paper uses the 2010 malaria indicator survey data to investigate if known malaria risk factors remain relevant after many years of interventions. We adopted a structured additive logistic regression model that allowed for spatial correlation, to more realistically estimate malaria risk factors. Our model included child and household level covariates, as well as climatic and environmental factors. Continuous variables were modelled by assuming second order random walk priors, while spatial correlation was specified as a Markov random field prior, with fixed effects assigned diffuse priors. Inference was fully Bayesian resulting in an under five malaria risk map for Malawi. Malaria risk increased with increasing age of the child. With respect to socio-economic factors, the greater the household wealth, the lower the malaria prevalence. A general decline in malaria risk was observed as altitude increased. Minimum temperatures and average total rainfall in the three months preceding the survey did not show a strong association with disease risk. The structured additive regression model offered a flexible extension to standard regression models by enabling simultaneous modelling of possible nonlinear effects of continuous covariates, spatial correlation and heterogeneity, while estimating usual fixed effects of categorical and continuous observed variables. Our results confirmed that malaria epidemiology is a complex interaction of biotic and abiotic factors, both at the individual, household and community level and that risk factors are still relevant many years after extensive implementation of RBM activities.
Vidal, M; Amigo, J M; Bro, R; Ostra, M; Ubide, C; Zuriarrain, J
2011-05-23
Desktop flatbed scanners are very well-known devices that can provide digitized information of flat surfaces. They are practically present in most laboratories as a part of the computer support. Several quality levels can be found in the market, but all of them can be considered as tools with a high performance and low cost. The present paper shows how the information obtained with a scanner, from a flat surface, can be used with fine results for exploratory and quantitative purposes through image analysis. It provides cheap analytical measurements for assessment of quality parameters of coated metallic surfaces and monitoring of electrochemical coating bath lives. The samples used were steel sheets nickel-plated in an electrodeposition bath. The quality of the final deposit depends on the bath conditions and, especially, on the concentration of the additives in the bath. Some additives become degraded with the bath life and so is the quality of the plate finish. Analysis of the scanner images can be used to follow the evolution of the metal deposit and the concentration of additives in the bath. Principal component analysis (PCA) is applied to find significant differences in the coating of sheets, to find directions of maximum variability and to identify odd samples. The results found are favorably compared with those obtained by means of specular reflectance (SR), which is here used as a reference technique. Also the concentration of additives SPB and SA-1 along a nickel bath life can be followed using image data handled with algorithms such as partial least squares (PLS) regression and support vector regression (SVR). The quantitative results obtained with these and other algorithms are compared. All this opens new qualitative and quantitative possibilities to flatbed scanners. Copyright © 2011 Elsevier B.V. All rights reserved.
Analysis of Multivariate Experimental Data Using A Simplified Regression Model Search Algorithm
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert M.
2013-01-01
A new regression model search algorithm was developed that may be applied to both general multivariate experimental data sets and wind tunnel strain-gage balance calibration data. The algorithm is a simplified version of a more complex algorithm that was originally developed for the NASA Ames Balance Calibration Laboratory. The new algorithm performs regression model term reduction to prevent overfitting of data. It has the advantage that it needs only about one tenth of the original algorithm's CPU time for the completion of a regression model search. In addition, extensive testing showed that the prediction accuracy of math models obtained from the simplified algorithm is similar to the prediction accuracy of math models obtained from the original algorithm. The simplified algorithm, however, cannot guarantee that search constraints related to a set of statistical quality requirements are always satisfied in the optimized regression model. Therefore, the simplified algorithm is not intended to replace the original algorithm. Instead, it may be used to generate an alternate optimized regression model of experimental data whenever the application of the original search algorithm fails or requires too much CPU time. Data from a machine calibration of NASA's MK40 force balance is used to illustrate the application of the new search algorithm.
Chai Rui; Li Si-Man; Xu Li-Sheng; Yao Yang; Hao Li-Ling
2017-07-01
This study mainly analyzed the parameters such as ascending branch slope (A_slope), dicrotic notch height (Hn), diastolic area (Ad) and systolic area (As) diastolic blood pressure (DBP), systolic blood pressure (SBP), pulse pressure (PP), subendocardial viability ratio (SEVR), waveform parameter (k), stroke volume (SV), cardiac output (CO) and peripheral resistance (RS) of central pulse wave invasively and non-invasively measured. These parameters extracted from the central pulse wave invasively measured were compared with the parameters measured from the brachial pulse waves by a regression model and a transfer function model. The accuracy of the parameters which were estimated by the regression model and the transfer function model was compared too. Our findings showed that in addition to the k value, the above parameters of the central pulse wave and the brachial pulse wave invasively measured had positive correlation. Both the regression model parameters including A_slope, DBP, SEVR and the transfer function model parameters had good consistency with the parameters invasively measured, and they had the same effect of consistency. The regression equations of the three parameters were expressed by Y'=a+bx. The SBP, PP, SV, CO of central pulse wave could be calculated through the regression model, but their accuracies were worse than that of transfer function model.
How is the weather? Forecasting inpatient glycemic control
Saulnier, George E; Castro, Janna C; Cook, Curtiss B; Thompson, Bithika M
2017-01-01
Aim: Apply methods of damped trend analysis to forecast inpatient glycemic control. Method: Observed and calculated point-of-care blood glucose data trends were determined over 62 weeks. Mean absolute percent error was used to calculate differences between observed and forecasted values. Comparisons were drawn between model results and linear regression forecasting. Results: The forecasted mean glucose trends observed during the first 24 and 48 weeks of projections compared favorably to the results provided by linear regression forecasting. However, in some scenarios, the damped trend method changed inferences compared with linear regression. In all scenarios, mean absolute percent error values remained below the 10% accepted by demand industries. Conclusion: Results indicate that forecasting methods historically applied within demand industries can project future inpatient glycemic control. Additional study is needed to determine if forecasting is useful in the analyses of other glucometric parameters and, if so, how to apply the techniques to quality improvement. PMID:29134125
Pekala, Ronald J; Baglio, Francesca; Cabinio, Monia; Lipari, Susanna; Baglio, Gisella; Mendozzi, Laura; Cecconi, Pietro; Pugnetti, Luigi; Sciaky, Riccardo
2017-01-01
Previous research using stepwise regression analyses found self-reported hypnotic depth (srHD) to be a function of suggestibility, trance state effects, and expectancy. This study sought to replicate and expand that research using a general state measure of hypnotic responsivity, the Phenomenology of Consciousness Inventory: Hypnotic Assessment Procedure (PCI-HAP). Ninety-five participants completed an Italian translation of the PCI-HAP, with srHD scores predicted from the PCI-HAP assessment items. The regression analysis replicated the previous research results. Additionally, stepwise regression analyses were able to predict the srHD score equally well using only the PCI dimension scores. These results not only replicated prior research but suggest how this methodology to assess hypnotic responsivity, when combined with more traditional neurophysiological and cognitive-behavioral methodologies, may allow for a more comprehensive understanding of that enigma called hypnosis.
Li, Xu; Zhang, Lei; Chen, Haibing; Guo, Kaifeng; Yu, Haoyong; Zhou, Jian; Li, Ming; Li, Qing; Li, Lianxi; Yin, Jun; Liu, Fang; Bao, Yuqian; Han, Junfeng; Jia, Weiping
2017-03-31
Recent studies highlight a negative association between total bilirubin concentrations and albuminuria in patients with type 2 diabetes mellitus. Our study evaluated the relationship between bilirubin concentrations and the prevalence of diabetic nephropathy (DN) in Chinese patients with type 1 diabetes mellitus (T1DM). A total of 258 patients with T1DM were recruited and bilirubin concentrations were compared between patients with or without diabetic nephropathy. Multiple stepwise regression analysis was used to examine the relationship between bilirubin concentrations and 24 h urinary microalbumin. Binary logistic regression analysis was performed to assess independent risk factors for diabetic nephropathy. Participants were divided into four groups according to the quartile of total bilirubin concentrations (Q1, 0.20-0.60; Q2, 0.60-0.80; Q3, 0.80-1.00; Q4, 1.00-1.90 mg/dL) and the chi-square test was used to compare the prevalence of DN in patients with T1DM. The median bilirubin level was 0.56 (interquartile: 0.43-0.68 mg/dL) in the DN group, significantly lower than in the non-DN group (0.70 [interquartile: 0.58-0.89 mg/dL], P < 0.001). Spearman's correlational analysis showed bilirubin concentrations were inversely correlated with 24 h urinary microalbumin (r = -0.13, P < 0.05) and multiple stepwise regression analysis showed bilirubin concentrations were independently associated with 24 h urinary microalbumin. In logistic regression analysis, bilirubin concentrations were significantly inversely associated with nephropathy. In addition, in stratified analysis, from the first to the fourth quartile group, increased bilirubin concentrations were associated with decreased prevalence of DN from 21.90% to 2.00%. High bilirubin concentrations are independently and negatively associated with albuminuria and the prevalence of DN in patients with T1DM.
Taljaard, Monica; McKenzie, Joanne E; Ramsay, Craig R; Grimshaw, Jeremy M
2014-06-19
An interrupted time series design is a powerful quasi-experimental approach for evaluating effects of interventions introduced at a specific point in time. To utilize the strength of this design, a modification to standard regression analysis, such as segmented regression, is required. In segmented regression analysis, the change in intercept and/or slope from pre- to post-intervention is estimated and used to test causal hypotheses about the intervention. We illustrate segmented regression using data from a previously published study that evaluated the effectiveness of a collaborative intervention to improve quality in pre-hospital ambulance care for acute myocardial infarction (AMI) and stroke. In the original analysis, a standard regression model was used with time as a continuous variable. We contrast the results from this standard regression analysis with those from segmented regression analysis. We discuss the limitations of the former and advantages of the latter, as well as the challenges of using segmented regression in analysing complex quality improvement interventions. Based on the estimated change in intercept and slope from pre- to post-intervention using segmented regression, we found insufficient evidence of a statistically significant effect on quality of care for stroke, although potential clinically important effects for AMI cannot be ruled out. Segmented regression analysis is the recommended approach for analysing data from an interrupted time series study. Several modifications to the basic segmented regression analysis approach are available to deal with challenges arising in the evaluation of complex quality improvement interventions.
Standardized Regression Coefficients as Indices of Effect Sizes in Meta-Analysis
ERIC Educational Resources Information Center
Kim, Rae Seon
2011-01-01
When conducting a meta-analysis, it is common to find many collected studies that report regression analyses, because multiple regression analysis is widely used in many fields. Meta-analysis uses effect sizes drawn from individual studies as a means of synthesizing a collection of results. However, indices of effect size from regression analyses…
Resistance of nickel-chromium-aluminum alloys to cyclic oxidation at 1100 C and 1200 C
NASA Technical Reports Server (NTRS)
Barrett, C. A.; Lowell, C. E.
1976-01-01
Nickel-rich alloys in the Ni-Cr-Al system were evaluated for cyclic oxidation resistance in still air at 1,100 and 1,200 C. A first approximation oxidation attack parameter Ka was derived from specific weight change data involving both a scaling growth constant and a spalling constant. An estimating equation was derived with Ka as a function of the Cr and Al content by multiple linear regression and translated into countour ternary diagrams showing regions of minimum attack. An additional factor inferred from the regression analysis was that alloys melted in zirconia crucibles had significantly greater oxidation resistance than comparable alloys melted otherwise.
Thermophysical Property Models for Lunar Regolith
NASA Technical Reports Server (NTRS)
Schreiner, Samuel S.; Dominguez, Jesus A.; Sibille, Laurent; Hoffman, Jeffrey A.
2015-01-01
We present a set of models for a wide range of lunar regolith material properties. Data from the literature are t with regression models for the following regolith properties: composition, density, specific heat, thermal conductivity, electrical conductivity, optical absorption length, and latent heat of melting/fusion. These models contain both temperature and composition dependencies so that they can be tailored for a range of applications. These models can enable more consistent, informed analysis and design of lunar regolith processing hardware. Furthermore, these models can be utilized to further inform lunar geological simulations. In addition to regression models for each material property, the raw data is also presented to allow for further interpretation and fitting as necessary.
Allard, Alexandra; Takman, Johanna; Uddin, Gazi Salah; Ahmed, Ali
2018-02-01
We evaluate the N-shaped environmental Kuznets curve (EKC) using panel quantile regression analysis. We investigate the relationship between CO 2 emissions and GDP per capita for 74 countries over the period of 1994-2012. We include additional explanatory variables, such as renewable energy consumption, technological development, trade, and institutional quality. We find evidence for the N-shaped EKC in all income groups, except for the upper-middle-income countries. Heterogeneous characteristics are, however, observed over the N-shaped EKC. Finally, we find a negative relationship between renewable energy consumption and CO 2 emissions, which highlights the importance of promoting greener energy in order to combat global warming.
Slaughter, Andrew R; Palmer, Carolyn G; Muller, Wilhelmine J
2007-04-01
In aquatic ecotoxicology, acute to chronic ratios (ACRs) are often used to predict chronic responses from available acute data to derive water quality guidelines, despite many problems associated with this method. This paper explores the comparative protectiveness and accuracy of predicted guideline values derived from the ACR, linear regression analysis (LRA), and multifactor probit analysis (MPA) extrapolation methods applied to acute toxicity data for aquatic macroinvertebrates. Although the authors of the LRA and MPA methods advocate the use of extrapolated lethal effects in the 0.01% to 10% lethal concentration (LC0.01-LC10) range to predict safe chronic exposure levels to toxicants, the use of an extrapolated LC50 value divided by a safety factor of 5 was in addition explored here because of higher statistical confidence surrounding the LC50 value. The LRA LC50/5 method was found to compare most favorably with available experimental chronic toxicity data and was therefore most likely to be sufficiently protective, although further validation with the use of additional species is needed. Values derived by the ACR method were the least protective. It is suggested that there is an argument for the replacement of ACRs in developing water quality guidelines by the LRA LC50/5 method.
Accounting for measurement error in log regression models with applications to accelerated testing.
Richardson, Robert; Tolley, H Dennis; Evenson, William E; Lunt, Barry M
2018-01-01
In regression settings, parameter estimates will be biased when the explanatory variables are measured with error. This bias can significantly affect modeling goals. In particular, accelerated lifetime testing involves an extrapolation of the fitted model, and a small amount of bias in parameter estimates may result in a significant increase in the bias of the extrapolated predictions. Additionally, bias may arise when the stochastic component of a log regression model is assumed to be multiplicative when the actual underlying stochastic component is additive. To account for these possible sources of bias, a log regression model with measurement error and additive error is approximated by a weighted regression model which can be estimated using Iteratively Re-weighted Least Squares. Using the reduced Eyring equation in an accelerated testing setting, the model is compared to previously accepted approaches to modeling accelerated testing data with both simulations and real data.
Held, Elizabeth; Cape, Joshua; Tintle, Nathan
2016-01-01
Machine learning methods continue to show promise in the analysis of data from genetic association studies because of the high number of variables relative to the number of observations. However, few best practices exist for the application of these methods. We extend a recently proposed supervised machine learning approach for predicting disease risk by genotypes to be able to incorporate gene expression data and rare variants. We then apply 2 different versions of the approach (radial and linear support vector machines) to simulated data from Genetic Analysis Workshop 19 and compare performance to logistic regression. Method performance was not radically different across the 3 methods, although the linear support vector machine tended to show small gains in predictive ability relative to a radial support vector machine and logistic regression. Importantly, as the number of genes in the models was increased, even when those genes contained causal rare variants, model predictive ability showed a statistically significant decrease in performance for both the radial support vector machine and logistic regression. The linear support vector machine showed more robust performance to the inclusion of additional genes. Further work is needed to evaluate machine learning approaches on larger samples and to evaluate the relative improvement in model prediction from the incorporation of gene expression data.
Almalki, Mohammed J; FitzGerald, Gerry; Clark, Michele
2012-09-12
Quality of work life (QWL) has been found to influence the commitment of health professionals, including nurses. However, reliable information on QWL and turnover intention of primary health care (PHC) nurses is limited. The aim of this study was to examine the relationship between QWL and turnover intention of PHC nurses in Saudi Arabia. A cross-sectional survey was used in this study. Data were collected using Brooks' survey of Quality of Nursing Work Life, the Anticipated Turnover Scale and demographic data questions. A total of 508 PHC nurses in the Jazan Region, Saudi Arabia, completed the questionnaire (RR = 87%). Descriptive statistics, t-test, ANOVA, General Linear Model (GLM) univariate analysis, standard multiple regression, and hierarchical multiple regression were applied for analysis using SPSS v17 for Windows. Findings suggested that the respondents were dissatisfied with their work life, with almost 40% indicating a turnover intention from their current PHC centres. Turnover intention was significantly related to QWL. Using standard multiple regression, 26% of the variance in turnover intention was explained by QWL, p < 0.001, with R2 = .263. Further analysis using hierarchical multiple regression found that the total variance explained by the model as a whole (demographics and QWL) was 32.1%, p < 0.001. QWL explained an additional 19% of the variance in turnover intention, after controlling for demographic variables. Creating and maintaining a healthy work life for PHC nurses is very important to improve their work satisfaction, reduce turnover, enhance productivity and improve nursing care outcomes.
Kitagawa, Noriyuki; Okada, Hiroshi; Tanaka, Muhei; Hashimoto, Yoshitaka; Kimura, Toshihiro; Nakano, Koji; Yamazaki, Masahiro; Hasegawa, Goji; Nakamura, Naoto; Fukui, Michiaki
2016-08-01
The aim of this study was to investigate whether central systolic blood pressure (SBP) was associated with albuminuria, defined as urinary albumin excretion (UAE) ≥30 mg/g creatinine, and, if so, whether the relationship of central SBP with albuminuria was stronger than that of peripheral SBP in patients with type 2 diabetes. The authors performed a cross-sectional study in 294 outpatients with type 2 diabetes. The relationship between peripheral SBP or central SBP and UAE using regression analysis was evaluated, and the odds ratios of peripheral SBP or central SBP were calculated to identify albuminuria using logistic regression model. Moreover, the area under the receiver operating characteristic curve (AUC) of central SBP was compared with that of peripheral SBP to identify albuminuria. Multiple regression analysis demonstrated that peripheral SBP (β=0.255, P<.0001) or central SBP (r=0.227, P<.0001) was associated with UAE. Multiple logistic regression analysis demonstrated that peripheral SBP (odds ratio, 1.029; 95% confidence interval, 1.016-1.043) or central SBP (odds ratio, 1.022; 95% confidence interval, 1.011-1.034) was associated with an increased odds of albuminuria. In addition, AUC of peripheral SBP was significantly greater than that of central SBP to identify albuminuria (P=0.035). Peripheral SBP is superior to central SBP in identifying albuminuria, although both peripheral and central SBP are associated with UAE in patients with type 2 diabetes. © 2016 Wiley Periodicals, Inc.
Bae, Jong-Myon; Kim, Eun Hee
2016-03-01
Research on how the risk of gastric cancer increases with Epstein-Barr virus (EBV) infection is lacking. In a systematic review that investigated studies published until September 2014, the authors did not calculate the summary odds ratio (SOR) due to heterogeneity across studies. Therefore, we include here additional studies published until October 2015 and conduct a meta-analysis with meta-regression that controls for the heterogeneity among studies. Using the studies selected in the previously published systematic review, we formulated lists of references, cited articles, and related articles provided by PubMed. From the lists, only case-control studies that detected EBV in tissue samples were selected. In order to control for the heterogeneity among studies, subgroup analysis and meta-regression were performed. In the 33 case-control results with adjacent non-cancer tissue, the total number of test samples in the case and control groups was 5280 and 4962, respectively. In the 14 case-control results with normal tissue, the total number of test samples in case and control groups was 1393 and 945, respectively. Upon meta-regression, the type of control tissue was found to be a statistically significant variable with regard to heterogeneity. When the control tissue was normal tissue of healthy individuals, the SOR was 3.41 (95% CI, 1.78 to 6.51; I-squared, 65.5%). The results of the present study support the argument that EBV infection increases the risk of gastric cancer. In the future, age-matched and sex-matched case-control studies should be conducted.
2012-01-01
Background Quality of work life (QWL) has been found to influence the commitment of health professionals, including nurses. However, reliable information on QWL and turnover intention of primary health care (PHC) nurses is limited. The aim of this study was to examine the relationship between QWL and turnover intention of PHC nurses in Saudi Arabia. Methods A cross-sectional survey was used in this study. Data were collected using Brooks’ survey of Quality of Nursing Work Life, the Anticipated Turnover Scale and demographic data questions. A total of 508 PHC nurses in the Jazan Region, Saudi Arabia, completed the questionnaire (RR = 87%). Descriptive statistics, t-test, ANOVA, General Linear Model (GLM) univariate analysis, standard multiple regression, and hierarchical multiple regression were applied for analysis using SPSS v17 for Windows. Results Findings suggested that the respondents were dissatisfied with their work life, with almost 40% indicating a turnover intention from their current PHC centres. Turnover intention was significantly related to QWL. Using standard multiple regression, 26% of the variance in turnover intention was explained by QWL, p < 0.001, with R2 = .263. Further analysis using hierarchical multiple regression found that the total variance explained by the model as a whole (demographics and QWL) was 32.1%, p < 0.001. QWL explained an additional 19% of the variance in turnover intention, after controlling for demographic variables. Conclusions Creating and maintaining a healthy work life for PHC nurses is very important to improve their work satisfaction, reduce turnover, enhance productivity and improve nursing care outcomes. PMID:22970764
ERIC Educational Resources Information Center
Hong, Hee Kyung
2012-01-01
The purpose of this study was to simultaneously examine relationships between teacher quality and instructional time and mathematics and science achievement of 8th grade cohorts in 18 advanced and developing economies. In addition, the study examined changes in mathematics and science performance across the two groups of economies over time using…
Revisiting the southern pine growth decline: Where are we 10 years later?
Gary L. Gadbury; Michael S. Williams; Hans T. Schreuder
2004-01-01
This paper evaluates changes in growth of pine stands in the state of Georgia, U.S.A., using USDA Forest Service Forest Inventory and Analysis (FIA) data. In particular, data representing an additional 10-year growth cy-cle has been added to previously published results from two earlier growth cycles. A robust regression procedure is combined with a bootstrap technique...
Thomas C. Edwards; Gretchen G. Moisen; Tracey S. Frescino; Joshua L. Lawler
2002-01-01
We describe our collective efforts to develop and apply methods for using FIA data to model forest resources and wildlife habitat. Our work demonstrates how flexible regression techniques, such as generalized additive models, can be linked with spatially explicit environmental information for the mapping of forest type and structure. We illustrate how these maps of...
Lipiäinen, Tiina; Pessi, Jenni; Movahedi, Parisa; Koivistoinen, Juha; Kurki, Lauri; Tenhunen, Mari; Yliruusi, Jouko; Juppo, Anne M; Heikkonen, Jukka; Pahikkala, Tapio; Strachan, Clare J
2018-04-03
Raman spectroscopy is widely used for quantitative pharmaceutical analysis, but a common obstacle to its use is sample fluorescence masking the Raman signal. Time-gating provides an instrument-based method for rejecting fluorescence through temporal resolution of the spectral signal and allows Raman spectra of fluorescent materials to be obtained. An additional practical advantage is that analysis is possible in ambient lighting. This study assesses the efficacy of time-gated Raman spectroscopy for the quantitative measurement of fluorescent pharmaceuticals. Time-gated Raman spectroscopy with a 128 × (2) × 4 CMOS SPAD detector was applied for quantitative analysis of ternary mixtures of solid-state forms of the model drug, piroxicam (PRX). Partial least-squares (PLS) regression allowed quantification, with Raman-active time domain selection (based on visual inspection) improving performance. Model performance was further improved by using kernel-based regularized least-squares (RLS) regression with greedy feature selection in which the data use in both the Raman shift and time dimensions was statistically optimized. Overall, time-gated Raman spectroscopy, especially with optimized data analysis in both the spectral and time dimensions, shows potential for sensitive and relatively routine quantitative analysis of photoluminescent pharmaceuticals during drug development and manufacturing.
Characterizing mammographic images by using generic texture features
2012-01-01
Introduction Although mammographic density is an established risk factor for breast cancer, its use is limited in clinical practice because of a lack of automated and standardized measurement methods. The aims of this study were to evaluate a variety of automated texture features in mammograms as risk factors for breast cancer and to compare them with the percentage mammographic density (PMD) by using a case-control study design. Methods A case-control study including 864 cases and 418 controls was analyzed automatically. Four hundred seventy features were explored as possible risk factors for breast cancer. These included statistical features, moment-based features, spectral-energy features, and form-based features. An elaborate variable selection process using logistic regression analyses was performed to identify those features that were associated with case-control status. In addition, PMD was assessed and included in the regression model. Results Of the 470 image-analysis features explored, 46 remained in the final logistic regression model. An area under the curve of 0.79, with an odds ratio per standard deviation change of 2.88 (95% CI, 2.28 to 3.65), was obtained with validation data. Adding the PMD did not improve the final model. Conclusions Using texture features to predict the risk of breast cancer appears feasible. PMD did not show any additional value in this study. With regard to the features assessed, most of the analysis tools appeared to reflect mammographic density, although some features did not correlate with PMD. It remains to be investigated in larger case-control studies whether these features can contribute to increased prediction accuracy. PMID:22490545
Using Dominance Analysis to Determine Predictor Importance in Logistic Regression
ERIC Educational Resources Information Center
Azen, Razia; Traxel, Nicole
2009-01-01
This article proposes an extension of dominance analysis that allows researchers to determine the relative importance of predictors in logistic regression models. Criteria for choosing logistic regression R[superscript 2] analogues were determined and measures were selected that can be used to perform dominance analysis in logistic regression. A…
Logsdon, Benjamin A.; Carty, Cara L.; Reiner, Alexander P.; Dai, James Y.; Kooperberg, Charles
2012-01-01
Motivation: For many complex traits, including height, the majority of variants identified by genome-wide association studies (GWAS) have small effects, leaving a significant proportion of the heritable variation unexplained. Although many penalized multiple regression methodologies have been proposed to increase the power to detect associations for complex genetic architectures, they generally lack mechanisms for false-positive control and diagnostics for model over-fitting. Our methodology is the first penalized multiple regression approach that explicitly controls Type I error rates and provide model over-fitting diagnostics through a novel normally distributed statistic defined for every marker within the GWAS, based on results from a variational Bayes spike regression algorithm. Results: We compare the performance of our method to the lasso and single marker analysis on simulated data and demonstrate that our approach has superior performance in terms of power and Type I error control. In addition, using the Women's Health Initiative (WHI) SNP Health Association Resource (SHARe) GWAS of African-Americans, we show that our method has power to detect additional novel associations with body height. These findings replicate by reaching a stringent cutoff of marginal association in a larger cohort. Availability: An R-package, including an implementation of our variational Bayes spike regression (vBsr) algorithm, is available at http://kooperberg.fhcrc.org/soft.html. Contact: blogsdon@fhcrc.org Supplementary information: Supplementary data are available at Bioinformatics online. PMID:22563072
Association of the FGA and SLC6A4 genes with autistic spectrum disorder in a Korean population.
Ro, Myungja; Won, Seongsik; Kang, Hyunjun; Kim, Su-Yeon; Lee, Seung Ku; Nam, Min; Bang, Hee Jung; Yang, Jae Won; Choi, Kyung-Sik; Kim, Su Kang; Chung, Joo-Ho; Kwack, Kyubum
2013-01-01
Autism spectrum disorder (ASD) is a neurobiological disorder characterized by distinctive impairments in cognitive function, language, and behavior. Linkage and population studies suggest a genetic association between solute carrier family 6 member 4 (SLC6A4) variants and ASD. Logistic regression was used to identify associations between single-nucleotide polymorphisms (SNPs) and ASD with 3 alternative models (additive, dominant, and recessive). Linear regression analysis was performed to determine the influence of SNPs on Childhood Autism Rating Scale (CARS) scores as a quantitative phenotype. In the present study, we examined the associations of SNPs in the SLC6A4 gene and the fibrinogen alpha chain (FGA) gene. Logistic regression analysis showed a significant association between the risk of ASD and rs2070025 and rs2070011 in the FGA gene. The gene-gene interaction between SLC6A4 and FGA was not significantly associated with ASD susceptibility. However, polymorphisms in both SLC6A4 and the FGA gene significantly affected the symptoms of ASD. Our findings indicate that FGA and SLC6A4 gene interactions may contribute to the phenotypes of ASD rather than the incidence of ASD. © 2013 S. Karger AG, Basel.
The quantile regression approach to efficiency measurement: insights from Monte Carlo simulations.
Liu, Chunping; Laporte, Audrey; Ferguson, Brian S
2008-09-01
In the health economics literature there is an ongoing debate over approaches used to estimate the efficiency of health systems at various levels, from the level of the individual hospital - or nursing home - up to that of the health system as a whole. The two most widely used approaches to evaluating the efficiency with which various units deliver care are non-parametric data envelopment analysis (DEA) and parametric stochastic frontier analysis (SFA). Productivity researchers tend to have very strong preferences over which methodology to use for efficiency estimation. In this paper, we use Monte Carlo simulation to compare the performance of DEA and SFA in terms of their ability to accurately estimate efficiency. We also evaluate quantile regression as a potential alternative approach. A Cobb-Douglas production function, random error terms and a technical inefficiency term with different distributions are used to calculate the observed output. The results, based on these experiments, suggest that neither DEA nor SFA can be regarded as clearly dominant, and that, depending on the quantile estimated, the quantile regression approach may be a useful addition to the armamentarium of methods for estimating technical efficiency.
Forecasting Container Throughput at the Doraleh Port in Djibouti through Time Series Analysis
NASA Astrophysics Data System (ADS)
Mohamed Ismael, Hawa; Vandyck, George Kobina
The Doraleh Container Terminal (DCT) located in Djibouti has been noted as the most technologically advanced container terminal on the African continent. DCT's strategic location at the crossroads of the main shipping lanes connecting Asia, Africa and Europe put it in a unique position to provide important shipping services to vessels plying that route. This paper aims to forecast container throughput through the Doraleh Container Port in Djibouti by Time Series Analysis. A selection of univariate forecasting models has been used, namely Triple Exponential Smoothing Model, Grey Model and Linear Regression Model. By utilizing the above three models and their combination, the forecast of container throughput through the Doraleh port was realized. A comparison of the different forecasting results of the three models, in addition to the combination forecast is then undertaken, based on commonly used evaluation criteria Mean Absolute Deviation (MAD) and Mean Absolute Percentage Error (MAPE). The study found that the Linear Regression forecasting Model was the best prediction method for forecasting the container throughput, since its forecast error was the least. Based on the regression model, a ten (10) year forecast for container throughput at DCT has been made.
Test anxiety and academic performance in chiropractic students.
Zhang, Niu; Henderson, Charles N R
2014-01-01
Objective : We assessed the level of students' test anxiety, and the relationship between test anxiety and academic performance. Methods : We recruited 166 third-quarter students. The Test Anxiety Inventory (TAI) was administered to all participants. Total scores from written examinations and objective structured clinical examinations (OSCEs) were used as response variables. Results : Multiple regression analysis shows that there was a modest, but statistically significant negative correlation between TAI scores and written exam scores, but not OSCE scores. Worry and emotionality were the best predictive models for written exam scores. Mean total anxiety and emotionality scores for females were significantly higher than those for males, but not worry scores. Conclusion : Moderate-to-high test anxiety was observed in 85% of the chiropractic students examined. However, total test anxiety, as measured by the TAI score, was a very weak predictive model for written exam performance. Multiple regression analysis demonstrated that replacing total anxiety (TAI) with worry and emotionality (TAI subscales) produces a much more effective predictive model of written exam performance. Sex, age, highest current academic degree, and ethnicity contributed little additional predictive power in either regression model. Moreover, TAI scores were not found to be statistically significant predictors of physical exam skill performance, as measured by OSCEs.
Armstrong, R A
2014-01-01
Factors associated with duration of dementia in a consecutive series of 103 Alzheimer's disease (AD) cases were studied using the Kaplan-Meier estimator and Cox regression analysis (proportional hazard model). Mean disease duration was 7.1 years (range: 6 weeks-30 years, standard deviation = 5.18); 25% of cases died within four years, 50% within 6.9 years, and 75% within 10 years. Familial AD cases (FAD) had a longer duration than sporadic cases (SAD), especially cases linked to presenilin (PSEN) genes. No significant differences in duration were associated with age, sex, or apolipoprotein E (Apo E) genotype. Duration was reduced in cases with arterial hypertension. Cox regression analysis suggested longer duration was associated with an earlier disease onset and increased senile plaque (SP) and neurofibrillary tangle (NFT) pathology in the orbital gyrus (OrG), CA1 sector of the hippocampus, and nucleus basalis of Meynert (NBM). The data suggest shorter disease duration in SAD and in cases with hypertensive comorbidity. In addition, degree of neuropathology did not influence survival, but spread of SP/NFT pathology into the frontal lobe, hippocampus, and basal forebrain was associated with longer disease duration.
Ren, Y Y; Zhou, L C; Yang, L; Liu, P Y; Zhao, B W; Liu, H X
2016-09-01
The paper highlights the use of the logistic regression (LR) method in the construction of acceptable statistically significant, robust and predictive models for the classification of chemicals according to their aquatic toxic modes of action. Essentials accounting for a reliable model were all considered carefully. The model predictors were selected by stepwise forward discriminant analysis (LDA) from a combined pool of experimental data and chemical structure-based descriptors calculated by the CODESSA and DRAGON software packages. Model predictive ability was validated both internally and externally. The applicability domain was checked by the leverage approach to verify prediction reliability. The obtained models are simple and easy to interpret. In general, LR performs much better than LDA and seems to be more attractive for the prediction of the more toxic compounds, i.e. compounds that exhibit excess toxicity versus non-polar narcotic compounds and more reactive compounds versus less reactive compounds. In addition, model fit and regression diagnostics was done through the influence plot which reflects the hat-values, studentized residuals, and Cook's distance statistics of each sample. Overdispersion was also checked for the LR model. The relationships between the descriptors and the aquatic toxic behaviour of compounds are also discussed.
Effect of duration of denervation on outcomes of ansa-recurrent laryngeal nerve reinnervation.
Li, Meng; Chen, Shicai; Wang, Wei; Chen, Donghui; Zhu, Minhui; Liu, Fei; Zhang, Caiyun; Li, Yan; Zheng, Hongliang
2014-08-01
To investigate the efficacy of laryngeal reinnervation with ansa cervicalis among unilateral vocal fold paralysis (UVFP) patients with different denervation durations. We retrospectively reviewed 349 consecutive UVFP cases of delayed ansa cervicalis to the recurrent laryngeal nerve (RLN) anastomosis. Potential influencing factors were analyzed in multivariable logistic regression analysis. Stratification analysis performed was aimed at one of the identified significant variables: denervation duration. Videostroboscopy, perceptual evaluation, acoustic analysis, maximum phonation time (MPT), and laryngeal electromyography (EMG) were performed preoperatively and postoperatively. Gender, age, preoperative EMG status and denervation duration were analyzed in multivariable logistic regression analysis. Stratification analysis was performed on denervation duration, which was divided into three groups according to the interval between RLN injury and reinnervation: group A, 6 to 12 months; group B, 12 to 24 months; and group C, > 24 months. Age, preoperative EMG, and denervation duration were identified as significant variables in multivariable logistic regression analysis. Stratification analysis on denervation duration showed significant differences between group A and C and between group B and C (P < 0.05)-but showed no significant difference between group A and B (P > 0.05) with regard to parameters overall grade, jitter, shimmer, noise-to-harmonics ratio, MPT, and postoperative EMG. In addition, videostroboscopic and laryngeal EMG data, perceptual and acoustic parameters, and MPT values were significantly improved postoperatively in each denervation duration group (P < 0.01). Although delayed laryngeal reinnervation is proved valid for UVFP, surgical outcome is better if the procedure is performed within 2 years after nerve injury than that over 2 years. © 2014 The American Laryngological, Rhinological and Otological Society, Inc.
Chakraborty, Somsubhra; Weindorf, David C; Li, Bin; Ali Aldabaa, Abdalsamad Abdalsatar; Ghosh, Rakesh Kumar; Paul, Sathi; Nasim Ali, Md
2015-05-01
Using 108 petroleum contaminated soil samples, this pilot study proposed a new analytical approach of combining visible near-infrared diffuse reflectance spectroscopy (VisNIR DRS) and portable X-ray fluorescence spectrometry (PXRF) for rapid and improved quantification of soil petroleum contamination. Results indicated that an advanced fused model where VisNIR DRS spectra-based penalized spline regression (PSR) was used to predict total petroleum hydrocarbon followed by PXRF elemental data-based random forest regression was used to model the PSR residuals, it outperformed (R(2)=0.78, residual prediction deviation (RPD)=2.19) all other models tested, even producing better generalization than using VisNIR DRS alone (RPD's of 1.64, 1.86, and 1.96 for random forest, penalized spline regression, and partial least squares regression, respectively). Additionally, unsupervised principal component analysis using the PXRF+VisNIR DRS system qualitatively separated contaminated soils from control samples. Fusion of PXRF elemental data and VisNIR derivative spectra produced an optimized model for total petroleum hydrocarbon quantification in soils. Copyright © 2015 Elsevier B.V. All rights reserved.
Lien, Tonje G; Borgan, Ørnulf; Reppe, Sjur; Gautvik, Kaare; Glad, Ingrid Kristine
2018-03-07
Using high-dimensional penalized regression we studied genome-wide DNA-methylation in bone biopsies of 80 postmenopausal women in relation to their bone mineral density (BMD). The women showed BMD varying from severely osteoporotic to normal. Global gene expression data from the same individuals was available, and since DNA-methylation often affects gene expression, the overall aim of this paper was to include both of these omics data sets into an integrated analysis. The classical penalized regression uses one penalty, but we incorporated individual penalties for each of the DNA-methylation sites. These individual penalties were guided by the strength of association between DNA-methylations and gene transcript levels. DNA-methylations that were highly associated to one or more transcripts got lower penalties and were therefore favored compared to DNA-methylations showing less association to expression. Because of the complex pathways and interactions among genes, we investigated both the association between DNA-methylations and their corresponding cis gene, as well as the association between DNA-methylations and trans-located genes. Two integrating penalized methods were used: first, an adaptive group-regularized ridge regression, and secondly, variable selection was performed through a modified version of the weighted lasso. When information from gene expressions was integrated, predictive performance was considerably improved, in terms of predictive mean square error, compared to classical penalized regression without data integration. We found a 14.7% improvement in the ridge regression case and a 17% improvement for the lasso case. Our version of the weighted lasso with data integration found a list of 22 interesting methylation sites. Several corresponded to genes that are known to be important in bone formation. Using BMD as response and these 22 methylation sites as covariates, least square regression analyses resulted in R 2 =0.726, comparable to an average R 2 =0.438 for 10000 randomly selected groups of DNA-methylations with group size 22. Two recent types of penalized regression methods were adapted to integrate DNA-methylation and their association to gene expression in the analysis of bone mineral density. In both cases predictions clearly benefit from including the additional information on gene expressions.
Habyarimana, Faustin; Zewotir, Temesgen; Ramroop, Shaun
2017-06-17
Childhood anemia is among the most significant health problems faced by public health departments in developing countries. This study aims at assessing the determinants and possible spatial effects associated with childhood anemia in Rwanda. The 2014/2015 Rwanda Demographic and Health Survey (RDHS) data was used. The analysis was done using the structured spatial additive quantile regression model. The findings of this study revealed that the child's age; the duration of breastfeeding; gender of the child; the nutritional status of the child (whether underweight and/or wasting); whether the child had a fever; had a cough in the two weeks prior to the survey or not; whether the child received vitamin A supplementation in the six weeks before the survey or not; the household wealth index; literacy of the mother; mother's anemia status; mother's age at the birth are all significant factors associated with childhood anemia in Rwanda. Furthermore, significant structured spatial location effects on childhood anemia was found.
Finley, Andrew O.; Banerjee, Sudipto; Cook, Bruce D.; Bradford, John B.
2013-01-01
In this paper we detail a multivariate spatial regression model that couples LiDAR, hyperspectral and forest inventory data to predict forest outcome variables at a high spatial resolution. The proposed model is used to analyze forest inventory data collected on the US Forest Service Penobscot Experimental Forest (PEF), ME, USA. In addition to helping meet the regression model's assumptions, results from the PEF analysis suggest that the addition of multivariate spatial random effects improves model fit and predictive ability, compared with two commonly applied modeling approaches. This improvement results from explicitly modeling the covariation among forest outcome variables and spatial dependence among observations through the random effects. Direct application of such multivariate models to even moderately large datasets is often computationally infeasible because of cubic order matrix algorithms involved in estimation. We apply a spatial dimension reduction technique to help overcome this computational hurdle without sacrificing richness in modeling.
Using the PLUM procedure of SPSS to fit unequal variance and generalized signal detection models.
DeCarlo, Lawrence T
2003-02-01
The recent addition of aprocedure in SPSS for the analysis of ordinal regression models offers a simple means for researchers to fit the unequal variance normal signal detection model and other extended signal detection models. The present article shows how to implement the analysis and how to interpret the SPSS output. Examples of fitting the unequal variance normal model and other generalized signal detection models are given. The approach offers a convenient means for applying signal detection theory to a variety of research.
The Relationship Between Surface Curvature and Abdominal Aortic Aneurysm Wall Stress.
de Galarreta, Sergio Ruiz; Cazón, Aitor; Antón, Raúl; Finol, Ender A
2017-08-01
The maximum diameter (MD) criterion is the most important factor when predicting risk of rupture of abdominal aortic aneurysms (AAAs). An elevated wall stress has also been linked to a high risk of aneurysm rupture, yet is an uncommon clinical practice to compute AAA wall stress. The purpose of this study is to assess whether other characteristics of the AAA geometry are statistically correlated with wall stress. Using in-house segmentation and meshing algorithms, 30 patient-specific AAA models were generated for finite element analysis (FEA). These models were subsequently used to estimate wall stress and maximum diameter and to evaluate the spatial distributions of wall thickness, cross-sectional diameter, mean curvature, and Gaussian curvature. Data analysis consisted of statistical correlations of the aforementioned geometry metrics with wall stress for the 30 AAA inner and outer wall surfaces. In addition, a linear regression analysis was performed with all the AAA wall surfaces to quantify the relationship of the geometric indices with wall stress. These analyses indicated that while all the geometry metrics have statistically significant correlations with wall stress, the local mean curvature (LMC) exhibits the highest average Pearson's correlation coefficient for both inner and outer wall surfaces. The linear regression analysis revealed coefficients of determination for the outer and inner wall surfaces of 0.712 and 0.516, respectively, with LMC having the largest effect on the linear regression equation with wall stress. This work underscores the importance of evaluating AAA mean wall curvature as a potential surrogate for wall stress.
A model for national outcome audit in vascular surgery.
Prytherch, D R; Ridler, B M; Beard, J D; Earnshaw, J J
2001-06-01
The aim was to model vascular surgical outcome in a national study using POSSUM scoring. One hundred and twenty-one British and Irish surgeons completed data questionnaires on patients undergoing arterial surgery under their care (mean 12 patients, range 1-49) in May/June 1998. A total of 1480 completed data records were available for logistic regression analysis using P-POSSUM methodology. Information collected included all POSSUM data items plus other factors thought to have a significant bearing on patient outcome: "extra items". The main outcome measures were death and major postoperative complications. The data were checked and inconsistent records were excluded. The remaining 1313 were divided into two sets for analysis. The first "training" set was used to obtain logistic regression models that were applied prospectively to the second "test" dataset. using POSSUM data items alone, it was possible to predict both mortality and morbidity after vascular reconstruction using P-POSSUM analysis. The addition of the "extra items" found significant in regression analysis did not significantly improve the accuracy of prediction. It was possible to predict both mortality and morbidity derived from the preoperative physiology components of the POSSUM data items alone. this study has shown that P-POSSUM methodology can be used to predict outcome after arterial surgery across a range of surgeons in different hospitals and could form the basis of a national outcome audit. It was also possible to obtain accurate models for both mortality and major morbidity from the POSSUM physiology scores alone. Copyright 2001 Harcourt Publishers Limited.
Arano, Ichiro; Sugimoto, Tomoyuki; Hamasaki, Toshimitsu; Ohno, Yuko
2010-04-23
Survival analysis methods such as the Kaplan-Meier method, log-rank test, and Cox proportional hazards regression (Cox regression) are commonly used to analyze data from randomized withdrawal studies in patients with major depressive disorder. However, unfortunately, such common methods may be inappropriate when a long-term censored relapse-free time appears in data as the methods assume that if complete follow-up were possible for all individuals, each would eventually experience the event of interest. In this paper, to analyse data including such a long-term censored relapse-free time, we discuss a semi-parametric cure regression (Cox cure regression), which combines a logistic formulation for the probability of occurrence of an event with a Cox proportional hazards specification for the time of occurrence of the event. In specifying the treatment's effect on disease-free survival, we consider the fraction of long-term survivors and the risks associated with a relapse of the disease. In addition, we develop a tree-based method for the time to event data to identify groups of patients with differing prognoses (cure survival CART). Although analysis methods typically adapt the log-rank statistic for recursive partitioning procedures, the method applied here used a likelihood ratio (LR) test statistic from a fitting of cure survival regression assuming exponential and Weibull distributions for the latency time of relapse. The method is illustrated using data from a sertraline randomized withdrawal study in patients with major depressive disorder. We concluded that Cox cure regression reveals facts on who may be cured, and how the treatment and other factors effect on the cured incidence and on the relapse time of uncured patients, and that cure survival CART output provides easily understandable and interpretable information, useful both in identifying groups of patients with differing prognoses and in utilizing Cox cure regression models leading to meaningful interpretations.
Liu, Quan; Ma, Li; Fan, Shou-Zen; Abbod, Maysam F; Shieh, Jiann-Shing
2018-01-01
Estimating the depth of anaesthesia (DoA) in operations has always been a challenging issue due to the underlying complexity of the brain mechanisms. Electroencephalogram (EEG) signals are undoubtedly the most widely used signals for measuring DoA. In this paper, a novel EEG-based index is proposed to evaluate DoA for 24 patients receiving general anaesthesia with different levels of unconsciousness. Sample Entropy (SampEn) algorithm was utilised in order to acquire the chaotic features of the signals. After calculating the SampEn from the EEG signals, Random Forest was utilised for developing learning regression models with Bispectral index (BIS) as the target. Correlation coefficient, mean absolute error, and area under the curve (AUC) were used to verify the perioperative performance of the proposed method. Validation comparisons with typical nonstationary signal analysis methods (i.e., recurrence analysis and permutation entropy) and regression methods (i.e., neural network and support vector machine) were conducted. To further verify the accuracy and validity of the proposed methodology, the data is divided into four unconsciousness-level groups on the basis of BIS levels. Subsequently, analysis of variance (ANOVA) was applied to the corresponding index (i.e., regression output). Results indicate that the correlation coefficient improved to 0.72 ± 0.09 after filtering and to 0.90 ± 0.05 after regression from the initial values of 0.51 ± 0.17. Similarly, the final mean absolute error dramatically declined to 5.22 ± 2.12. In addition, the ultimate AUC increased to 0.98 ± 0.02, and the ANOVA analysis indicates that each of the four groups of different anaesthetic levels demonstrated significant difference from the nearest levels. Furthermore, the Random Forest output was extensively linear in relation to BIS, thus with better DoA prediction accuracy. In conclusion, the proposed method provides a concrete basis for monitoring patients' anaesthetic level during surgeries.
Bignardi, A B; El Faro, L; Torres Júnior, R A A; Cardoso, V L; Machado, P F; Albuquerque, L G
2011-10-31
We analyzed 152,145 test-day records from 7317 first lactations of Holstein cows recorded from 1995 to 2003. Our objective was to model variations in test-day milk yield during the first lactation of Holstein cows by random regression model (RRM), using various functions in order to obtain adequate and parsimonious models for the estimation of genetic parameters. Test-day milk yields were grouped into weekly classes of days in milk, ranging from 1 to 44 weeks. The contemporary groups were defined as herd-test-day. The analyses were performed using a single-trait RRM, including the direct additive, permanent environmental and residual random effects. In addition, contemporary group and linear and quadratic effects of the age of cow at calving were included as fixed effects. The mean trend of milk yield was modeled with a fourth-order orthogonal Legendre polynomial. The additive genetic and permanent environmental covariance functions were estimated by random regression on two parametric functions, Ali and Schaeffer and Wilmink, and on B-spline functions of days in milk. The covariance components and the genetic parameters were estimated by the restricted maximum likelihood method. Results from RRM parametric and B-spline functions were compared to RRM on Legendre polynomials and with a multi-trait analysis, using the same data set. Heritability estimates presented similar trends during mid-lactation (13 to 31 weeks) and between week 37 and the end of lactation, for all RRM. Heritabilities obtained by multi-trait analysis were of a lower magnitude than those estimated by RRM. The RRMs with a higher number of parameters were more useful to describe the genetic variation of test-day milk yield throughout the lactation. RRM using B-spline and Legendre polynomials as base functions appears to be the most adequate to describe the covariance structure of the data.
Willke, Richard J; Zheng, Zhiyuan; Subedi, Prasun; Althin, Rikard; Mullins, C Daniel
2012-12-13
Implicit in the growing interest in patient-centered outcomes research is a growing need for better evidence regarding how responses to a given intervention or treatment may vary across patients, referred to as heterogeneity of treatment effect (HTE). A variety of methods are available for exploring HTE, each associated with unique strengths and limitations. This paper reviews a selected set of methodological approaches to understanding HTE, focusing largely but not exclusively on their uses with randomized trial data. It is oriented for the "intermediate" outcomes researcher, who may already be familiar with some methods, but would value a systematic overview of both more and less familiar methods with attention to when and why they may be used. Drawing from the biomedical, statistical, epidemiological and econometrics literature, we describe the steps involved in choosing an HTE approach, focusing on whether the intent of the analysis is for exploratory, initial testing, or confirmatory testing purposes. We also map HTE methodological approaches to data considerations as well as the strengths and limitations of each approach. Methods reviewed include formal subgroup analysis, meta-analysis and meta-regression, various types of predictive risk modeling including classification and regression tree analysis, series of n-of-1 trials, latent growth and growth mixture models, quantile regression, and selected non-parametric methods. In addition to an overview of each HTE method, examples and references are provided for further reading.By guiding the selection of the methods and analysis, this review is meant to better enable outcomes researchers to understand and explore aspects of HTE in the context of patient-centered outcomes research.
NASA Astrophysics Data System (ADS)
Sivalingam, Udhayaraj; Wels, Michael; Rempfler, Markus; Grosskopf, Stefan; Suehling, Michael; Menze, Bjoern H.
2016-03-01
In this paper, we present a fully automated approach to coronary vessel segmentation, which involves calcification or soft plaque delineation in addition to accurate lumen delineation, from 3D Cardiac Computed Tomography Angiography data. Adequately virtualizing the coronary lumen plays a crucial role for simulating blood ow by means of fluid dynamics while additionally identifying the outer vessel wall in the case of arteriosclerosis is a prerequisite for further plaque compartment analysis. Our method is a hybrid approach complementing Active Contour Model-based segmentation with an external image force that relies on a Random Forest Regression model generated off-line. The regression model provides a strong estimate of the distance to the true vessel surface for every surface candidate point taking into account 3D wavelet-encoded contextual image features, which are aligned with the current surface hypothesis. The associated external image force is integrated in the objective function of the active contour model, such that the overall segmentation approach benefits from the advantages associated with snakes and from the ones associated with machine learning-based regression alike. This yields an integrated approach achieving competitive results on a publicly available benchmark data collection (Rotterdam segmentation challenge).
Carbonell, Felix; Bellec, Pierre; Shmuel, Amir
2011-01-01
The influence of the global average signal (GAS) on functional-magnetic resonance imaging (fMRI)-based resting-state functional connectivity is a matter of ongoing debate. The global average fluctuations increase the correlation between functional systems beyond the correlation that reflects their specific functional connectivity. Hence, removal of the GAS is a common practice for facilitating the observation of network-specific functional connectivity. This strategy relies on the implicit assumption of a linear-additive model according to which global fluctuations, irrespective of their origin, and network-specific fluctuations are super-positioned. However, removal of the GAS introduces spurious negative correlations between functional systems, bringing into question the validity of previous findings of negative correlations between fluctuations in the default-mode and the task-positive networks. Here we present an alternative method for estimating global fluctuations, immune to the complications associated with the GAS. Principal components analysis was applied to resting-state fMRI time-series. A global-signal effect estimator was defined as the principal component (PC) that correlated best with the GAS. The mean correlation coefficient between our proposed PC-based global effect estimator and the GAS was 0.97±0.05, demonstrating that our estimator successfully approximated the GAS. In 66 out of 68 runs, the PC that showed the highest correlation with the GAS was the first PC. Since PCs are orthogonal, our method provides an estimator of the global fluctuations, which is uncorrelated to the remaining, network-specific fluctuations. Moreover, unlike the regression of the GAS, the regression of the PC-based global effect estimator does not introduce spurious anti-correlations beyond the decrease in seed-based correlation values allowed by the assumed additive model. After regressing this PC-based estimator out of the original time-series, we observed robust anti-correlations between resting-state fluctuations in the default-mode and the task-positive networks. We conclude that resting-state global fluctuations and network-specific fluctuations are uncorrelated, supporting a Resting-State Linear-Additive Model. In addition, we conclude that the network-specific resting-state fluctuations of the default-mode and task-positive networks show artifact-free anti-correlations.
Gender-Dependent Association of FTO Polymorphisms with Body Mass Index in Mexicans
Saldaña-Alvarez, Yolanda; Salas-Martínez, María Guadalupe; García-Ortiz, Humberto; Luckie-Duque, Angélica; García-Cárdenas, Gustavo; Vicenteño-Ayala, Hermenegildo; Cordova, Emilio J.; Esparza-Aguilar, Marcelino; Contreras-Cubas, Cecilia; Carnevale, Alessandra; Chávez-Saldaña, Margarita; Orozco, Lorena
2016-01-01
To evaluate the associations between six single-nucleotide polymorphisms (SNPs) in intron 1 of FTO and body mass index (BMI), a case-control association study of 2314 unrelated Mexican-Mestizo adult subjects was performed. The association between each SNP and BMI was tested using logistic and linear regression adjusted for age, gender, and ancestry and assuming additive, recessive, and dominant effects of the minor allele. Association analysis after BMI stratification showed that all five FTO SNPs (rs1121980, rs17817449, rs3751812, rs9930506, and rs17817449), were significantly associated with obesity class II/III under an additive model (P<0.05). Interestingly, we also documented a genetic model-dependent influence of gender on the effect of FTO variants on increased BMI. Two SNPs were specifically associated in males under a dominant model, while the remainder were associated with females under additive and recessive models (P<0.05). The SNP rs9930506 showed the highest increased in obesity risk in females (odds ratio = 4.4). Linear regression using BMI as a continuous trait also revealed differential FTO SNP contributions. Homozygous individuals for the risk alleles of rs17817449, rs3751812, and rs9930506 were on average 2.18 kg/m2 heavier than homozygous for the wild-type alleles; rs1121980 and rs8044769 showed significant but less-strong effects on BMI (1.54 kg/m2 and 0.9 kg/m2, respectively). Remarkably, rs9930506 also exhibited positive interactions with age and BMI in a gender-dependent manner. Women carrying the minor allele of this variant have a significant increase in BMI by year (0.42 kg/m2, P = 1.17 x 10−10). Linear regression haplotype analysis under an additive model, confirmed that the TGTGC haplotype harboring all five minor alleles, increased the BMI of carriers by 2.36 kg/m2 (P = 1.15 x 10−5). Our data suggest that FTO SNPs make differential contributions to obesity risk and support the hypothesis that gender differences in the mechanisms involving these variants may contribute to disease development. PMID:26726774
NASA Astrophysics Data System (ADS)
Fellers, R. S.; Braly, L. B.; Saykally, R. J.; Leforestier, C.
1999-04-01
The SWPS method is improved by the addition of H.E.G. contractions for generating a more compact basis. An error in the definition of the internal fragment axis system used in our previous calculation is described and corrected. Fully coupled 6D (rigid monomers) VRT states are computed for several new water dimer potential surfaces and compared with experiment and our earlier SWPS results. This work sets the stage for refinement of such potential surfaces via regression analysis of VRT spectroscopic data.
NASA Astrophysics Data System (ADS)
Veale, Kirsty; Adali, Sarp; Pitot, Jean; Brooks, Michael
2017-12-01
Paraffin wax as a hybrid rocket fuel has not been comprehensively characterised, especially regarding the structural feasibility of the material in launch applications. Preliminary structural testing has shown paraffin wax to be a brittle, low strength material, and at risk of failure under launch loading conditions. Structural enhancing additives have been identified, but their effect on motor performance has not always been considered, nor has any standard method of testing been identified between research institutes. A review of existing regression rate measurement techniques on paraffin wax based fuels and the results obtained with various additives are collated and discussed in this paper. The review includes 2D slab motors that enable visualisation of liquefying fuel droplet entrainment and the effect of an increased viscosity on the droplet entrainment mechanism, which can occur with the addition of structural enhancing polymers. An increased viscosity has been shown to reduce the regression rate of liquefying fuels. Viscosity increasing additives that have been tested include EVA and LDPE. Both these additives increase the structural properties of paraffin wax, where the elongation and UTS are improved. Other additives, such as metal hydrides, aluminium and boron generally offer improvements on the regression rate. However, very little consideration has been given to the structural effects these additives have on the wax grain. A 40% aluminised grain, for example, offers a slight increase in the UTS but reduces the elongation of paraffin wax. Geometrically accurate lab-scale motors have also been used to determine the regression rate properties of various additives in paraffin wax. A concise review of all available regression rate testing techniques and results on paraffin wax based hybrid propellants, as well as existing structural testing data, is presented in this paper.
Liu, Gang; Mukherjee, Bhramar; Lee, Seunggeun; Lee, Alice W; Wu, Anna H; Bandera, Elisa V; Jensen, Allan; Rossing, Mary Anne; Moysich, Kirsten B; Chang-Claude, Jenny; Doherty, Jennifer A; Gentry-Maharaj, Aleksandra; Kiemeney, Lambertus; Gayther, Simon A; Modugno, Francesmary; Massuger, Leon; Goode, Ellen L; Fridley, Brooke L; Terry, Kathryn L; Cramer, Daniel W; Ramus, Susan J; Anton-Culver, Hoda; Ziogas, Argyrios; Tyrer, Jonathan P; Schildkraut, Joellen M; Kjaer, Susanne K; Webb, Penelope M; Ness, Roberta B; Menon, Usha; Berchuck, Andrew; Pharoah, Paul D; Risch, Harvey; Pearce, Celeste Leigh
2018-02-01
There have been recent proposals advocating the use of additive gene-environment interaction instead of the widely used multiplicative scale, as a more relevant public health measure. Using gene-environment independence enhances statistical power for testing multiplicative interaction in case-control studies. However, under departure from this assumption, substantial bias in the estimates and inflated type I error in the corresponding tests can occur. In this paper, we extend the empirical Bayes (EB) approach previously developed for multiplicative interaction, which trades off between bias and efficiency in a data-adaptive way, to the additive scale. An EB estimator of the relative excess risk due to interaction is derived, and the corresponding Wald test is proposed with a general regression setting under a retrospective likelihood framework. We study the impact of gene-environment association on the resultant test with case-control data. Our simulation studies suggest that the EB approach uses the gene-environment independence assumption in a data-adaptive way and provides a gain in power compared with the standard logistic regression analysis and better control of type I error when compared with the analysis assuming gene-environment independence. We illustrate the methods with data from the Ovarian Cancer Association Consortium. © The Author(s) 2017. Published by Oxford University Press on behalf of the Johns Hopkins Bloomberg School of Public Health. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
Forrest, Elizabeth Ann; Reiling, Janske; Lipka, Geraldine; Fawcett, Jonathan
2017-01-01
AIM To identify risk factors associated with the formation of biliary strictures post liver transplantation over a period of 10-year in Queensland. METHODS Data on liver donors and recipients in Queensland between 2005 and 2014 was obtained from an electronic patient data system. In addition, intra-operative and post-operative characteristics were collected and a logistical regression analysis was performed to evaluate their association with the development of biliary strictures. RESULTS Of 296 liver transplants performed, 285 (96.3%) were from brain dead donors. Biliary strictures developed in 45 (15.2%) recipients. Anastomotic stricture formation (n = 25, 48.1%) was the commonest complication, with 14 (58.3%) of these occurred within 6-mo of transplant. A percutaneous approach or endoscopic retrograde cholangiography was used to treat 17 (37.8%) patients with biliary strictures. Biliary reconstruction was initially or ultimately required in 22 (48.9%) patients. In recipients developing biliary strictures, bilirubin was significantly increased within the first post-operative week (Day 7 total bilirubin 74 μmol/L vs 49 μmol/L, P = 0.012). In both univariate and multivariate regression analysis, Day 7 total bilirubin > 55 μmol/L was associated with the development of biliary stricture formation. In addition, hepatic artery thrombosis and primary sclerosing cholangitis were identified as independent risk factors. CONCLUSION In addition to known risk factors, bilirubin levels in the early post-operative period could be used as a clinical indicator for biliary stricture formation. PMID:29312864
Regression: The Apple Does Not Fall Far From the Tree.
Vetter, Thomas R; Schober, Patrick
2018-05-15
Researchers and clinicians are frequently interested in either: (1) assessing whether there is a relationship or association between 2 or more variables and quantifying this association; or (2) determining whether 1 or more variables can predict another variable. The strength of such an association is mainly described by the correlation. However, regression analysis and regression models can be used not only to identify whether there is a significant relationship or association between variables but also to generate estimations of such a predictive relationship between variables. This basic statistical tutorial discusses the fundamental concepts and techniques related to the most common types of regression analysis and modeling, including simple linear regression, multiple regression, logistic regression, ordinal regression, and Poisson regression, as well as the common yet often underrecognized phenomenon of regression toward the mean. The various types of regression analysis are powerful statistical techniques, which when appropriately applied, can allow for the valid interpretation of complex, multifactorial data. Regression analysis and models can assess whether there is a relationship or association between 2 or more observed variables and estimate the strength of this association, as well as determine whether 1 or more variables can predict another variable. Regression is thus being applied more commonly in anesthesia, perioperative, critical care, and pain research. However, it is crucial to note that regression can identify plausible risk factors; it does not prove causation (a definitive cause and effect relationship). The results of a regression analysis instead identify independent (predictor) variable(s) associated with the dependent (outcome) variable. As with other statistical methods, applying regression requires that certain assumptions be met, which can be tested with specific diagnostics.
A meta-analysis investigating factors underlying attrition rates in infant ERP studies.
Stets, Manuela; Stahl, Daniel; Reid, Vincent M
2012-01-01
In this meta-analysis, we examined interrelationships between characteristics of infant event-related potential (ERP) studies and their attrition rates. One-hundred and forty-nine published studies provided information on 314 experimental groups of which 181 provided data on attrition. A random effects meta-analysis revealed a high average attrition rate of 49.2%. Additionally, we used meta-regression for 178 groups with attrition data to analyze which variables best explained attrition variance. Our main findings were that the nature of the stimuli-visual, auditory, or combined as well as if stimuli were animated-influenced exclusion rates from the final analysis and that infant age did not alter attrition rates.
Cost Differences in Public and Private Shipyards
1981-01-01
block number) coefficients, costs, maintenance, naval shore facilities, naval vessels, nuclear powered ships, regression analysis, repair, salaries...of overhauls of nucler submarines, we mnight exp.,_t to find both production costs and the price of labor to be higher in naval shipyardi than in...about 18 months; in addition to the type of work done during regular overhauls, they include replacement of the nuclear core which powers the submarine
Fonseca, Maria de Jesus Mendes da; Juvanhol, Leidjaira Lopes; Rotenberg, Lúcia; Nobre, Aline Araújo; Griep, Rosane Härter; Alves, Márcia Guimarães de Mello; Cardoso, Letícia de Oliveira; Giatti, Luana; Nunes, Maria Angélica; Aquino, Estela M L; Chor, Dóra
2017-11-17
This paper explores the association between job strain and adiposity, using two statistical analysis approaches and considering the role of gender. The research evaluated 11,960 active baseline participants (2008-2010) in the ELSA-Brasil study. Job strain was evaluated through a demand-control questionnaire, while body mass index (BMI) and waist circumference (WC) were evaluated in continuous form. The associations were estimated using gamma regression models with an identity link function. Quantile regression models were also estimated from the final set of co-variables established by gamma regression. The relationship that was found varied by analytical approach and gender. Among the women, no association was observed between job strain and adiposity in the fitted gamma models. In the quantile models, a pattern of increasing effects of high strain was observed at higher BMI and WC distribution quantiles. Among the men, high strain was associated with adiposity in the gamma regression models. However, when quantile regression was used, that association was found not to be homogeneous across outcome distributions. In addition, in the quantile models an association was observed between active jobs and BMI. Our results point to an association between job strain and adiposity, which follows a heterogeneous pattern. Modelling strategies can produce different results and should, accordingly, be used to complement one another.
Developing a global psychotherapeutic approach to schizophrenia: results of a five-year follow-up.
Alanen, Y. O.; Räkköläinen, V.; Rasimus, R.; Laakso, J.; Kaljonen, A.
1985-01-01
This is an account of a long-range action research project to determine indications for and effects of a comprehensive psychotherapeutic approach, including various treatment modalities, in the treatment of schizophrenics. Four diagnostic groups were established among the 100 patients. In the course of data analysis, the group of typical schizophrenics (56 percent) was contrasted to or compared with the entire series. A further diagnostic differentiation was established according to ego functioning; i.e., imminent, acute, regressive, and paranoid ego disintegrations, respectively. Patients and family members were interviewed upon admission, and again two and five years later, and the data recorded on a 163-item form from which 40 clinical and psychosocial variables were constructed after the baseline examinations. In addition to cross-tabulation, logistic regression analysis was employed. The conclusion that the follow-up study supports the effectiveness of our global psychotherapeutic approach to treating schizophrenia seems justified. Results so far indicate that five modes of therapy in addition to drug treatments are optimal for different patients. The five modes are long-term individual psychotherapy, couple or conjoint family therapy for married patients, family therapy with the family of origin, flexible short-term crisis intervention with a family focus, and extensive long-term treatment focused on social rehabilitation for the most ill-starred patients. PMID:4049919
Ding, Changfeng; Li, Xiaogang; Zhang, Taolin; Ma, Yibing; Wang, Xingxiang
2014-10-01
Soil environmental quality standards in respect of heavy metals for farmlands should be established considering both their effects on crop yield and their accumulation in the edible part. A greenhouse experiment was conducted to investigate the effects of chromium (Cr) on biomass production and Cr accumulation in carrot plants grown in a wide range of soils. The results revealed that carrot yield significantly decreased in 18 of the total 20 soils with Cr addition being the soil environmental quality standard of China. The Cr content of carrot grown in the five soils with pH>8.0 exceeded the maximum allowable level (0.5mgkg(-1)) according to the Chinese General Standard for Contaminants in Foods. The relationship between carrot Cr concentration and soil pH could be well fitted (R(2)=0.70, P<0.0001) by a linear-linear segmented regression model. The addition of Cr to soil influenced carrot yield firstly rather than the food quality. The major soil factors controlling Cr phytotoxicity and the prediction models were further identified and developed using path analysis and stepwise multiple linear regression analysis. Soil Cr thresholds for phytotoxicity meanwhile ensuring food safety were then derived on the condition of 10 percent yield reduction. Copyright © 2014 Elsevier Inc. All rights reserved.
Admixture Analysis of Spontaneous Hepatitis C Virus Clearance in Individuals of African-Descent
Wojcik, Genevieve L.; Thio, Chloe L.; Kao, WH Linda; Latanich, Rachel; Goedert, James J.; Mehta, Shruti H.; Kirk, Gregory D.; Peters, Marion G.; Cox, Andrea L.; Kim, Arthur Y.; Chung, Raymond T.; Thomas, David L.; Duggal, Priya
2015-01-01
Hepatitis C virus (HCV) infects an estimated 3% of the global population with the majority of individuals (75–85%) failing to clear the virus without treatment, leading to chronic liver disease. Individuals of African-descent have lower rates of clearance compared to individuals of European-descent and this is not fully explained by social and environmental factors. This suggests that differences in genetic background may contribute to this difference in clinical outcome following HCV infection. Using 473 individuals and 792,721 SNPs from a genome-wide association study (GWAS), we estimated local African ancestry across the genome. Using admixture mapping and logistic regression we identified two regions of interest associated with spontaneous clearance of HCV (15q24, 20p12). A genome-wide significant variant was identified on chromosome 15 at the imputed SNP, rs55817928 (P=6.18×10−8) between the genes SCAPER and RCN. Each additional copy of the African ancestral C allele is associated with 2.4 times the odds of spontaneous clearance. Conditional analysis using this SNP in the logistic regression model explained one-third of the local ancestry association. Additionally, signals of selection in this area suggest positive selection due to some ancestral pathogen or environmental pressure in African, but not in European populations. PMID:24622687
Applied Multiple Linear Regression: A General Research Strategy
ERIC Educational Resources Information Center
Smith, Brandon B.
1969-01-01
Illustrates some of the basic concepts and procedures for using regression analysis in experimental design, analysis of variance, analysis of covariance, and curvilinear regression. Applications to evaluation of instruction and vocational education programs are illustrated. (GR)
Optimal design application on the advanced aeroelastic rotor blade
NASA Technical Reports Server (NTRS)
Wei, F. S.; Jones, R.
1985-01-01
The vibration and performance optimization procedure using regression analysis was successfully applied to an advanced aeroelastic blade design study. The major advantage of this regression technique is that multiple optimizations can be performed to evaluate the effects of various objective functions and constraint functions. The data bases obtained from the rotorcraft flight simulation program C81 and Myklestad mode shape program are analytically determined as a function of each design variable. This approach has been verified for various blade radial ballast weight locations and blade planforms. This method can also be utilized to ascertain the effect of a particular cost function which is composed of several objective functions with different weighting factors for various mission requirements without any additional effort.
Threshold altitude resulting in decompression sickness
NASA Technical Reports Server (NTRS)
Kumar, K. V.; Waligora, James M.; Calkins, Dick S.
1990-01-01
A review of case reports, hypobaric chamber training data, and experimental evidence indicated that the threshold for incidence of altitude decompression sickness (DCS) was influenced by various factors such as prior denitrogenation, exercise or rest, and period of exposure, in addition to individual susceptibility. Fitting these data with appropriate statistical models makes it possible to examine the influence of various factors on the threshold for DCS. This approach was illustrated by logistic regression analysis on the incidence of DCS below 9144 m. Estimations using these regressions showed that, under a noprebreathe, 6-h exposure, simulated EVA profile, the threshold for symptoms occurred at approximately 3353 m; while under a noprebreathe, 2-h exposure profile with knee-bends exercise, the threshold occurred at 7925 m.
A Permutation Approach for Selecting the Penalty Parameter in Penalized Model Selection
Sabourin, Jeremy A; Valdar, William; Nobel, Andrew B
2015-01-01
Summary We describe a simple, computationally effcient, permutation-based procedure for selecting the penalty parameter in LASSO penalized regression. The procedure, permutation selection, is intended for applications where variable selection is the primary focus, and can be applied in a variety of structural settings, including that of generalized linear models. We briefly discuss connections between permutation selection and existing theory for the LASSO. In addition, we present a simulation study and an analysis of real biomedical data sets in which permutation selection is compared with selection based on the following: cross-validation (CV), the Bayesian information criterion (BIC), Scaled Sparse Linear Regression, and a selection method based on recently developed testing procedures for the LASSO. PMID:26243050
Recurrent shoulder dystocia: is it predictable?
Kleitman, Vered; Feldman, Roi; Walfisch, Asnat; Toledano, Ronen; Sheiner, Eyal
2016-11-01
To examine the course and outcome of deliveries occurring in women who previously experienced shoulder dystocia. In addition, recurrent shoulder dystocia risk factors were assessed. A retrospective cohort analysis comparing all singleton deliveries with and without shoulder dystocia in their preceding delivery was conducted. Independent predictors of recurrent shoulder dystocia were investigated using a multiple logistic regression model. Of the 201,422 deliveries included in the analysis, 307 occurred in women with a previous shoulder dystocia (0.015 %). Women with a history of shoulder dystocia were more likely to be older, experienced higher rates of gestational diabetes mellitus, polyhydramnios, prolonged second stage, operative delivery and macrosomia (>4000 g) in the following delivery. Previous shoulder dystocia was found to be an independent risk factor for recurrent shoulder dystocia (OR = 6.1, 95 % CI 3.2-11.8, p value <0.001) in the multivariable regression analysis. Shoulder dystocia is an independent risk factor for recurrent shoulder dystocia. Deliveries in women with a history of shoulder dystocia are characterized by higher rates of operative delivery, prolonged second stage of labor and macrosomia.
Structural Analysis of Women’s Heptathlon
Gassmann, Freya; Fröhlich, Michael; Emrich, Eike
2016-01-01
The heptathlon comprises the results of seven single disciplines, assuming an equal influence from each discipline, depending on the measured performance. Data analysis was based on the data recorded for the individual performances of the 10 winning heptathletes in the World Athletics Championships from 1987 to 2013 and the Olympic Games from 1988 to 2012. In addition to descriptive analysis methods, correlations, bivariate and multivariate linear regressions, and panel data regressions were used. The transformation of the performances from seconds, centimeters, and meters into points showed that the individual disciplines do not equally affect the overall competition result. The currently valid conversion formula for the run, jump, and throw disciplines prefers the sprint and jump disciplines but penalizes the athletes performing in the 800 m run, javelin throw, and shotput disciplines. Furthermore, 21% to 48% of the variance of the sum of points can be attributed to the performances in the disciplines of long jump, 200 m sprint, 100 m hurdles, and high jump. To balance the effects of the single disciplines in the heptathlon, the formula to calculate points should be reevaluated. PMID:29910260
Jakubovski, Ewgeni; Varigonda, Anjali L; Freemantle, Nicholas; Taylor, Matthew J; Bloch, Michael H
2016-02-01
Previous studies suggested that the treatment response to selective serotonin reuptake inhibitors (SSRIs) in major depressive disorder follows a flat response curve within the therapeutic dose range. The present study was designed to clarify the relationship between dosage and treatment response in major depressive disorder. The authors searched PubMed for randomized placebo-controlled trials examining the efficacy of SSRIs for treating adults with major depressive disorder. Trials were also required to assess improvement in depression severity at multiple time points. Additional data were collected on treatment response and all-cause and side effect-related discontinuation. All medication doses were transformed into imipramine-equivalent doses. The longitudinal data were analyzed with a mixed-regression model. Endpoint and tolerability analyses were analyzed using meta-regression and stratified subgroup analysis by predefined SSRI dose categories in order to assess the effect of SSRI dosing on the efficacy and tolerability of SSRIs for major depressive disorder. Forty studies involving 10,039 participants were included. Longitudinal modeling (dose-by-time interaction=0.0007, 95% CI=0.0001-0.0013) and endpoint analysis (meta-regression: β=0.00053, 95% CI=0.00018-0.00088, z=2.98) demonstrated a small but statistically significant positive association between SSRI dose and efficacy. Higher doses of SSRIs were associated with an increased likelihood of dropouts due to side effects (meta-regression: β=0.00207, 95% CI=0.00071-0.00342, z=2.98) and decreased likelihood of all-cause dropout (meta-regression: β=-0.00093, 95% CI=-0.00165 to -0.00021, z=-2.54). Higher doses of SSRIs appear slightly more effective in major depressive disorder. This benefit appears to plateau at around 250 mg of imipramine equivalents (50 mg of fluoxetine). The slightly increased benefits of SSRIs at higher doses are somewhat offset by decreased tolerability at high doses.
Ono, Tomohiro; Nakamura, Mitsuhiro; Hirose, Yoshinori; Kitsuda, Kenji; Ono, Yuka; Ishigaki, Takashi; Hiraoka, Masahiro
2017-09-01
To estimate the lung tumor position from multiple anatomical features on four-dimensional computed tomography (4D-CT) data sets using single regression analysis (SRA) and multiple regression analysis (MRA) approach and evaluate an impact of the approach on internal target volume (ITV) for stereotactic body radiotherapy (SBRT) of the lung. Eleven consecutive lung cancer patients (12 cases) underwent 4D-CT scanning. The three-dimensional (3D) lung tumor motion exceeded 5 mm. The 3D tumor position and anatomical features, including lung volume, diaphragm, abdominal wall, and chest wall positions, were measured on 4D-CT images. The tumor position was estimated by SRA using each anatomical feature and MRA using all anatomical features. The difference between the actual and estimated tumor positions was defined as the root-mean-square error (RMSE). A standard partial regression coefficient for the MRA was evaluated. The 3D lung tumor position showed a high correlation with the lung volume (R = 0.92 ± 0.10). Additionally, ITVs derived from SRA and MRA approaches were compared with ITV derived from contouring gross tumor volumes on all 10 phases of the 4D-CT (conventional ITV). The RMSE of the SRA was within 3.7 mm in all directions. Also, the RMSE of the MRA was within 1.6 mm in all directions. The standard partial regression coefficient for the lung volume was the largest and had the most influence on the estimated tumor position. Compared with conventional ITV, average percentage decrease of ITV were 31.9% and 38.3% using SRA and MRA approaches, respectively. The estimation accuracy of lung tumor position was improved by the MRA approach, which provided smaller ITV than conventional ITV. © 2017 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.
Chen, Hung-Yuan; Chiu, Yen-Ling; Hsu, Shih-Ping; Pai, Mei-Fen; Ju-YehYang; Lai, Chun-Fu; Lu, Hui-Min; Huang, Shu-Chen; Yang, Shao-Yu; Wen, Su-Yin; Chiu, Hsien-Ching; Hu, Fu-Chang; Peng, Yu-Sen; Jee, Shiou-Hwa
2013-01-01
Background Uremic pruritus is a common and intractable symptom in patients on chronic hemodialysis, but factors associated with the severity of pruritus remain unclear. This study aimed to explore the associations of metabolic factors and dialysis adequacy with the aggravation of pruritus. Methods We conducted a 5-year prospective cohort study on patients with maintenance hemodialysis. A visual analogue scale (VAS) was used to assess the intensity of pruritus. Patient demographic and clinical characteristics, laboratory parameters, dialysis adequacy (assessed by Kt/V), and pruritus intensity were recorded at baseline and follow-up. Change score analysis of the difference score of VAS between baseline and follow-up was performed using multiple linear regression models. The optimal threshold of Kt/V, which is associated with the aggravation of uremic pruritus, was determined by generalized additive models and receiver operating characteristic analysis. Results A total of 111 patients completed the study. Linear regression analysis showed that lower Kt/V and use of low-flux dialyzer were significantly associated with the aggravation of pruritus after adjusting for the baseline pruritus intensity and a variety of confounding factors. The optimal threshold value of Kt/V for pruritus was 1.5 suggested by both generalized additive models and receiver operating characteristic analysis. Conclusions Hemodialysis with the target of Kt/V ≥1.5 and use of high-flux dialyzer may reduce the intensity of pruritus in patients on chronic hemodialysis. Further clinical trials are required to determine the optimal dialysis dose and regimen for uremic pruritus. PMID:23940749
NASA Technical Reports Server (NTRS)
Parsons, Vickie s.
2009-01-01
The request to conduct an independent review of regression models, developed for determining the expected Launch Commit Criteria (LCC) External Tank (ET)-04 cycle count for the Space Shuttle ET tanking process, was submitted to the NASA Engineering and Safety Center NESC on September 20, 2005. The NESC team performed an independent review of regression models documented in Prepress Regression Analysis, Tom Clark and Angela Krenn, 10/27/05. This consultation consisted of a peer review by statistical experts of the proposed regression models provided in the Prepress Regression Analysis. This document is the consultation's final report.
Spatial analysis of alcohol-related motor vehicle crash injuries in southeastern Michigan.
Meliker, Jaymie R; Maio, Ronald F; Zimmerman, Marc A; Kim, Hyungjin Myra; Smith, Sarah C; Wilson, Mark L
2004-11-01
Temporal, behavioral and social risk factors that affect injuries resulting from alcohol-related motor vehicle crashes have been characterized in previous research. Much less is known about spatial patterns and environmental associations of alcohol-related motor vehicle crashes. The aim of this study was to evaluate geographic patterns of alcohol-related motor vehicle crashes and to determine if locations of alcohol outlets are associated with those crashes. In addition, we sought to demonstrate the value of integrating spatial and traditional statistical techniques in the analysis of this preventable public health risk. The study design was a cross-sectional analysis of individual-level blood alcohol content, traffic report information, census block group data, and alcohol distribution outlets. Besag and Newell's spatial analysis and traditional logistic regression both indicated that areas of low population density had more alcohol-related motor vehicle crashes than expected (P < 0.05). There was no significant association between alcohol outlets and alcohol-related motor vehicle crashes using distance analyses, logistic regression, and Chi-square. Differences in environmental or behavioral factors characteristic of areas of low population density may be responsible for the higher proportion of alcohol-related crashes occurring in these areas.
NASA Astrophysics Data System (ADS)
Candefjord, Stefan; Nyberg, Morgan; Jalkanen, Ville; Ramser, Kerstin; Lindahl, Olof A.
2010-12-01
Tissue characterization is fundamental for identification of pathological conditions. Raman spectroscopy (RS) and tactile resonance measurement (TRM) are two promising techniques that measure biochemical content and stiffness, respectively. They have potential to complement the golden standard--histological analysis. By combining RS and TRM, complementary information about tissue content can be obtained and specific drawbacks can be avoided. The aim of this study was to develop a multivariate approach to compare RS and TRM information. The approach was evaluated on measurements at the same points on porcine abdominal tissue. The measurement points were divided into five groups by multivariate analysis of the RS data. A regression analysis was performed and receiver operating characteristic (ROC) curves were used to compare the RS and TRM data. TRM identified one group efficiently (area under ROC curve 0.99). The RS data showed that the proportion of saturated fat was high in this group. The regression analysis showed that stiffness was mainly determined by the amount of fat and its composition. We concluded that RS provided additional, important information for tissue identification that was not provided by TRM alone. The results are promising for development of a method combining RS and TRM for intraoperative tissue characterization.
Baldacchino, Tara; Jacobs, William R; Anderson, Sean R; Worden, Keith; Rowson, Jennifer
2018-01-01
This contribution presents a novel methodology for myolectric-based control using surface electromyographic (sEMG) signals recorded during finger movements. A multivariate Bayesian mixture of experts (MoE) model is introduced which provides a powerful method for modeling force regression at the fingertips, while also performing finger movement classification as a by-product of the modeling algorithm. Bayesian inference of the model allows uncertainties to be naturally incorporated into the model structure. This method is tested using data from the publicly released NinaPro database which consists of sEMG recordings for 6 degree-of-freedom force activations for 40 intact subjects. The results demonstrate that the MoE model achieves similar performance compared to the benchmark set by the authors of NinaPro for finger force regression. Additionally, inherent to the Bayesian framework is the inclusion of uncertainty in the model parameters, naturally providing confidence bounds on the force regression predictions. Furthermore, the integrated clustering step allows a detailed investigation into classification of the finger movements, without incurring any extra computational effort. Subsequently, a systematic approach to assessing the importance of the number of electrodes needed for accurate control is performed via sensitivity analysis techniques. A slight degradation in regression performance is observed for a reduced number of electrodes, while classification performance is unaffected.
Analysis of Multivariate Experimental Data Using A Simplified Regression Model Search Algorithm
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred
2013-01-01
A new regression model search algorithm was developed in 2011 that may be used to analyze both general multivariate experimental data sets and wind tunnel strain-gage balance calibration data. The new algorithm is a simplified version of a more complex search algorithm that was originally developed at the NASA Ames Balance Calibration Laboratory. The new algorithm has the advantage that it needs only about one tenth of the original algorithm's CPU time for the completion of a search. In addition, extensive testing showed that the prediction accuracy of math models obtained from the simplified algorithm is similar to the prediction accuracy of math models obtained from the original algorithm. The simplified algorithm, however, cannot guarantee that search constraints related to a set of statistical quality requirements are always satisfied in the optimized regression models. Therefore, the simplified search algorithm is not intended to replace the original search algorithm. Instead, it may be used to generate an alternate optimized regression model of experimental data whenever the application of the original search algorithm either fails or requires too much CPU time. Data from a machine calibration of NASA's MK40 force balance is used to illustrate the application of the new regression model search algorithm.
Ide, Kazuki; Kawasaki, Yohei; Akutagawa, Maiko; Yamada, Hiroshi
2017-02-01
The aim of this study is to analyze the data obtained from a randomized trial on the prevention of influenza by gargling with green tea, which gave nonsignificant results based on frequentist approaches, by using Bayesian approaches. The posterior proportion, with 95% credible interval (CrI), of influenza in each group was calculated. The Bayesian index θ is the probability that a hypothesis is true. In this case, θ is the probability that the hypothesis that green tea gargling reduced influenza compared with water gargling is true. Univariate and multivariate logistic regression analyses were also performed by using the Markov chain Monte Carlo method. The full analysis set included 747 participants. During the study period, influenza occurred in 44 participants (5.9%). The difference between the two independent binominal proportions was -0.019 (95% CrI, -0.054 to 0.015; θ = 0.87). The partial regression coefficients in the univariate analysis were -0.35 (95% CrI, -1.00 to 0.24) with use of a uniform prior and -0.34 (95% CrI, -0.96 to 0.27) with use of a Jeffreys prior. In the multivariate analysis, the values were -0.37 (95% CrI, -0.96 to 0.30) and -0.36 (95% CrI, -1.03 to 0.21), respectively. The difference between the two independent binominal proportions was less than 0, and θ was greater than 0.85. Therefore, green tea gargling may slightly reduce influenza compared with water gargling. This analysis suggests that green tea gargling can be an additional preventive measure for use with other pharmaceutical and nonpharmaceutical measures and indicates the need for additional studies to confirm the effect of green tea gargling.
International consensus on preliminary definitions of improvement in adult and juvenile myositis.
Rider, Lisa G; Giannini, Edward H; Brunner, Hermine I; Ruperto, Nicola; James-Newton, Laura; Reed, Ann M; Lachenbruch, Peter A; Miller, Frederick W
2004-07-01
To use a core set of outcome measures to develop preliminary definitions of improvement for adult and juvenile myositis as composite end points for therapeutic trials. Twenty-nine experts in the assessment of myositis achieved consensus on 102 adult and 102 juvenile paper patient profiles as clinically improved or not improved. Two hundred twenty-seven candidate definitions of improvement were developed using the experts' consensus ratings as a gold standard and their judgment of clinically meaningful change in the core set of measures. Seventeen additional candidate definitions of improvement were developed from classification and regression tree analysis, a data-mining decision tree tool analysis. Six candidate definitions specifying percentage change or raw change in the core set of measures were developed using logistic regression analysis. Adult and pediatric working groups ranked the 13 top-performing candidate definitions for face validity, clinical sensibility, and ease of use, in which the sensitivity and specificity were >/=75% in adult, pediatric, and combined data sets. Nominal group technique was used to facilitate consensus formation. The definition of improvement (common to the adult and pediatric working groups) that ranked highest was 3 of any 6 of the core set measures improved by >/=20%, with no more than 2 worse by >/=25% (which could not include manual muscle testing to assess strength). Five and 4 additional preliminary definitions of improvement for adult and juvenile myositis, respectively, were also developed, with several definitions common to both groups. Participants also agreed to prospectively test 6 logistic regression definitions of improvement in clinical trials. Consensus preliminary definitions of improvement were developed for adult and juvenile myositis, and these incorporate clinically meaningful change in all myositis core set measures in a composite end point. These definitions require prospective validation, but they are now proposed for use as end points in all myositis trials.
Assessment of parametric uncertainty for groundwater reactive transport modeling,
Shi, Xiaoqing; Ye, Ming; Curtis, Gary P.; Miller, Geoffery L.; Meyer, Philip D.; Kohler, Matthias; Yabusaki, Steve; Wu, Jichun
2014-01-01
The validity of using Gaussian assumptions for model residuals in uncertainty quantification of a groundwater reactive transport model was evaluated in this study. Least squares regression methods explicitly assume Gaussian residuals, and the assumption leads to Gaussian likelihood functions, model parameters, and model predictions. While the Bayesian methods do not explicitly require the Gaussian assumption, Gaussian residuals are widely used. This paper shows that the residuals of the reactive transport model are non-Gaussian, heteroscedastic, and correlated in time; characterizing them requires using a generalized likelihood function such as the formal generalized likelihood function developed by Schoups and Vrugt (2010). For the surface complexation model considered in this study for simulating uranium reactive transport in groundwater, parametric uncertainty is quantified using the least squares regression methods and Bayesian methods with both Gaussian and formal generalized likelihood functions. While the least squares methods and Bayesian methods with Gaussian likelihood function produce similar Gaussian parameter distributions, the parameter distributions of Bayesian uncertainty quantification using the formal generalized likelihood function are non-Gaussian. In addition, predictive performance of formal generalized likelihood function is superior to that of least squares regression and Bayesian methods with Gaussian likelihood function. The Bayesian uncertainty quantification is conducted using the differential evolution adaptive metropolis (DREAM(zs)) algorithm; as a Markov chain Monte Carlo (MCMC) method, it is a robust tool for quantifying uncertainty in groundwater reactive transport models. For the surface complexation model, the regression-based local sensitivity analysis and Morris- and DREAM(ZS)-based global sensitivity analysis yield almost identical ranking of parameter importance. The uncertainty analysis may help select appropriate likelihood functions, improve model calibration, and reduce predictive uncertainty in other groundwater reactive transport and environmental modeling.
The use of generalized estimating equations in the analysis of motor vehicle crash data.
Hutchings, Caroline B; Knight, Stacey; Reading, James C
2003-01-01
The purpose of this study was to determine if it is necessary to use generalized estimating equations (GEEs) in the analysis of seat belt effectiveness in preventing injuries in motor vehicle crashes. The 1992 Utah crash dataset was used, excluding crash participants where seat belt use was not appropriate (n=93,633). The model used in the 1996 Report to Congress [Report to congress on benefits of safety belts and motorcycle helmets, based on data from the Crash Outcome Data Evaluation System (CODES). National Center for Statistics and Analysis, NHTSA, Washington, DC, February 1996] was analyzed for all occupants with logistic regression, one level of nesting (occupants within crashes), and two levels of nesting (occupants within vehicles within crashes) to compare the use of GEEs with logistic regression. When using one level of nesting compared to logistic regression, 13 of 16 variance estimates changed more than 10%, and eight of 16 parameter estimates changed more than 10%. In addition, three of the independent variables changed from significant to insignificant (alpha=0.05). With the use of two levels of nesting, two of 16 variance estimates and three of 16 parameter estimates changed more than 10% from the variance and parameter estimates in one level of nesting. One of the independent variables changed from insignificant to significant (alpha=0.05) in the two levels of nesting model; therefore, only two of the independent variables changed from significant to insignificant when the logistic regression model was compared to the two levels of nesting model. The odds ratio of seat belt effectiveness in preventing injuries was 12% lower when a one-level nested model was used. Based on these results, we stress the need to use a nested model and GEEs when analyzing motor vehicle crash data.
Hung, Shih-Chiang; Kung, Chia-Te; Hung, Chih-Wei; Liu, Ber-Ming; Liu, Jien-Wei; Chew, Ghee; Chuang, Hung-Yi; Lee, Wen-Huei; Lee, Tzu-Chi
2014-08-23
The adverse effects of delayed admission to the intensive care unit (ICU) have been recognized in previous studies. However, the definitions of delayed admission varies across studies. This study proposed a model to define "delayed admission", and explored the effect of ICU-waiting time on patients' outcome. This retrospective cohort study included non-traumatic adult patients on mechanical ventilation in the emergency department (ED), from July 2009 to June 2010. The primary outcomes measures were 21-ventilator-day mortality and prolonged hospital stays (over 30 days). Models of Cox regression and logistic regression were used for multivariate analysis. The non-delayed ICU-waiting was defined as a period in which the time effect on mortality was not statistically significant in a Cox regression model. To identify a suitable cut-off point between "delayed" and "non-delayed", subsets from the overall data were made based on ICU-waiting time and the hazard ratio of ICU-waiting hour in each subset was iteratively calculated. The cut-off time was then used to evaluate the impact of delayed ICU admission on mortality and prolonged length of hospital stay. The final analysis included 1,242 patients. The time effect on mortality emerged after 4 hours, thus we deduced ICU-waiting time in ED > 4 hours as delayed. By logistic regression analysis, delayed ICU admission affected the outcomes of 21 ventilator-days mortality and prolonged hospital stay, with odds ratio of 1.41 (95% confidence interval, 1.05 to 1.89) and 1.56 (95% confidence interval, 1.07 to 2.27) respectively. For patients on mechanical ventilation at the ED, delayed ICU admission is associated with higher probability of mortality and additional resource expenditure. A benchmark waiting time of no more than 4 hours for ICU admission is recommended.
Liu, Shujie; Kawamoto, Taisuke; Morita, Osamu; Yoshinari, Kouichi; Honda, Hiroshi
2017-03-01
Chemical exposure often results in liver hypertrophy in animal tests, characterized by increased liver weight, hepatocellular hypertrophy, and/or cell proliferation. While most of these changes are considered adaptive responses, there is concern that they may be associated with carcinogenesis. In this study, we have employed a toxicogenomic approach using a logistic ridge regression model to identify genes responsible for liver hypertrophy and hypertrophic hepatocarcinogenesis and to develop a predictive model for assessing hypertrophy-inducing compounds. Logistic regression models have previously been used in the quantification of epidemiological risk factors. DNA microarray data from the Toxicogenomics Project-Genomics Assisted Toxicity Evaluation System were used to identify hypertrophy-related genes that are expressed differently in hypertrophy induced by carcinogens and non-carcinogens. Data were collected for 134 chemicals (72 non-hypertrophy-inducing chemicals, 27 hypertrophy-inducing non-carcinogenic chemicals, and 15 hypertrophy-inducing carcinogenic compounds). After applying logistic ridge regression analysis, 35 genes for liver hypertrophy (e.g., Acot1 and Abcc3) and 13 genes for hypertrophic hepatocarcinogenesis (e.g., Asns and Gpx2) were selected. The predictive models built using these genes were 94.8% and 82.7% accurate, respectively. Pathway analysis of the genes indicates that, aside from a xenobiotic metabolism-related pathway as an adaptive response for liver hypertrophy, amino acid biosynthesis and oxidative responses appear to be involved in hypertrophic hepatocarcinogenesis. Early detection and toxicogenomic characterization of liver hypertrophy using our models may be useful for predicting carcinogenesis. In addition, the identified genes provide novel insight into discrimination between adverse hypertrophy associated with carcinogenesis and adaptive hypertrophy in risk assessment. Copyright © 2017 Elsevier Inc. All rights reserved.
Selection of higher order regression models in the analysis of multi-factorial transcription data.
Prazeres da Costa, Olivia; Hoffman, Arthur; Rey, Johannes W; Mansmann, Ulrich; Buch, Thorsten; Tresch, Achim
2014-01-01
Many studies examine gene expression data that has been obtained under the influence of multiple factors, such as genetic background, environmental conditions, or exposure to diseases. The interplay of multiple factors may lead to effect modification and confounding. Higher order linear regression models can account for these effects. We present a new methodology for linear model selection and apply it to microarray data of bone marrow-derived macrophages. This experiment investigates the influence of three variable factors: the genetic background of the mice from which the macrophages were obtained, Yersinia enterocolitica infection (two strains, and a mock control), and treatment/non-treatment with interferon-γ. We set up four different linear regression models in a hierarchical order. We introduce the eruption plot as a new practical tool for model selection complementary to global testing. It visually compares the size and significance of effect estimates between two nested models. Using this methodology we were able to select the most appropriate model by keeping only relevant factors showing additional explanatory power. Application to experimental data allowed us to qualify the interaction of factors as either neutral (no interaction), alleviating (co-occurring effects are weaker than expected from the single effects), or aggravating (stronger than expected). We find a biologically meaningful gene cluster of putative C2TA target genes that appear to be co-regulated with MHC class II genes. We introduced the eruption plot as a tool for visual model comparison to identify relevant higher order interactions in the analysis of expression data obtained under the influence of multiple factors. We conclude that model selection in higher order linear regression models should generally be performed for the analysis of multi-factorial microarray data.
Application of Temperature Sensitivities During Iterative Strain-Gage Balance Calibration Analysis
NASA Technical Reports Server (NTRS)
Ulbrich, N.
2011-01-01
A new method is discussed that may be used to correct wind tunnel strain-gage balance load predictions for the influence of residual temperature effects at the location of the strain-gages. The method was designed for the iterative analysis technique that is used in the aerospace testing community to predict balance loads from strain-gage outputs during a wind tunnel test. The new method implicitly applies temperature corrections to the gage outputs during the load iteration process. Therefore, it can use uncorrected gage outputs directly as input for the load calculations. The new method is applied in several steps. First, balance calibration data is analyzed in the usual manner assuming that the balance temperature was kept constant during the calibration. Then, the temperature difference relative to the calibration temperature is introduced as a new independent variable for each strain--gage output. Therefore, sensors must exist near the strain--gages so that the required temperature differences can be measured during the wind tunnel test. In addition, the format of the regression coefficient matrix needs to be extended so that it can support the new independent variables. In the next step, the extended regression coefficient matrix of the original calibration data is modified by using the manufacturer specified temperature sensitivity of each strain--gage as the regression coefficient of the corresponding temperature difference variable. Finally, the modified regression coefficient matrix is converted to a data reduction matrix that the iterative analysis technique needs for the calculation of balance loads. Original calibration data and modified check load data of NASA's MC60D balance are used to illustrate the new method.
Eckmann, Christian; Wasserman, Matthew; Latif, Faisal; Roberts, Graeme; Beriot-Mathiot, Axelle
2013-10-01
Hospital-onset Clostridium difficile infection (CDI) places a significant burden on health care systems throughout Europe, estimated at around €3 billion per annum. This burden is shared between national payers and hospitals that support additional bed days for patients diagnosed with CDI while in hospital or patients re-admitted from a previous hospitalisation. This study was performed to quantify additional hospital stay attributable to CDI in four countries, England, Germany, Spain, and The Netherlands, by analysing nationwide hospital-episode data. We focused upon patients at increased risk of CDI: with chronic obstructive pulmonary disease, heart failure, diabetes, or chronic kidney disease, and aged 50 years or over. Multivariate regression and propensity score matching models were developed to investigate the impact of CDI on additional length of hospital stay, controlling for confounding factors such as underlying disease severity. Patients in England had the longest additional hospital stay attributable to CDI at 16.09 days, followed by Germany at 15.47 days, Spain at 13.56 days, and The Netherlands at 12.58 days, derived using regression analysis. Propensity score matching indicated a higher attributable length of stay of 32.42 days in England, 15.31 days in Spain, and 18.64 days in The Netherlands. Outputs from this study consistently demonstrate that in European countries, for patients whose hospitalisation is complicated by CDI, the infection causes a statistically significant increase in hospital length of stay. This has implications for optimising resource allocation and budget setting at both the national and hospital level to ensure that levels of CDI-complicated hospitalisations are minimised.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Shujie; Kawamoto, Taisuke; Morita, Osamu
Chemical exposure often results in liver hypertrophy in animal tests, characterized by increased liver weight, hepatocellular hypertrophy, and/or cell proliferation. While most of these changes are considered adaptive responses, there is concern that they may be associated with carcinogenesis. In this study, we have employed a toxicogenomic approach using a logistic ridge regression model to identify genes responsible for liver hypertrophy and hypertrophic hepatocarcinogenesis and to develop a predictive model for assessing hypertrophy-inducing compounds. Logistic regression models have previously been used in the quantification of epidemiological risk factors. DNA microarray data from the Toxicogenomics Project-Genomics Assisted Toxicity Evaluation System weremore » used to identify hypertrophy-related genes that are expressed differently in hypertrophy induced by carcinogens and non-carcinogens. Data were collected for 134 chemicals (72 non-hypertrophy-inducing chemicals, 27 hypertrophy-inducing non-carcinogenic chemicals, and 15 hypertrophy-inducing carcinogenic compounds). After applying logistic ridge regression analysis, 35 genes for liver hypertrophy (e.g., Acot1 and Abcc3) and 13 genes for hypertrophic hepatocarcinogenesis (e.g., Asns and Gpx2) were selected. The predictive models built using these genes were 94.8% and 82.7% accurate, respectively. Pathway analysis of the genes indicates that, aside from a xenobiotic metabolism-related pathway as an adaptive response for liver hypertrophy, amino acid biosynthesis and oxidative responses appear to be involved in hypertrophic hepatocarcinogenesis. Early detection and toxicogenomic characterization of liver hypertrophy using our models may be useful for predicting carcinogenesis. In addition, the identified genes provide novel insight into discrimination between adverse hypertrophy associated with carcinogenesis and adaptive hypertrophy in risk assessment. - Highlights: • Hypertrophy (H) and hypertrophic carcinogenesis (C) were studied by toxicogenomics. • Important genes for H and C were selected by logistic ridge regression analysis. • Amino acid biosynthesis and oxidative responses may be involved in C. • Predictive models for H and C provided 94.8% and 82.7% accuracy, respectively. • The identified genes could be useful for assessment of liver hypertrophy.« less
Rapid increase in Japanese life expectancy after World War II.
Sugiura, Yasuo; Ju, Young-Su; Yasuoka, Junko; Jimba, Masamine
2010-02-01
Japanese life expectancy increased by about 13.7 years during the first decade after World War II, despite the country's post-war poverty. Although it is known that medical progress explains part of this increase, roles of non-medical factors have not been systematically studied. This study hypothesizes that non-medical factors, in addition to medical factors, are associated with the rapid increase in life expectancy in Japan. We analyzed the time trends of potential explanatory factors and used regression analysis with historical data from the Ministry of Internal Affairs and Communications' Historical Statistics of Japan during the period between 1946 and 1983. Time trends analysis revealed that the rapid increase in life expectancy preceded the dramatic growth of per capita Gross Domestic Product (GDP) by 10 years. In education, the nearly universal enrollment in elementary schools and increased advancement to upper secondary schools for both sexes were associated with better health. Regarding legislation, 32 health laws were passed in the first decade after the war and these laws were associated with improved health. Using regression analysis, we found that the enrollment rate in elementary schools, the number of health laws, and expansion of community-based activity staff were significantly associated with the increased life expectancy during the first decade after World War II. To conclude, in addition to medical factors, non-medical factors applied across the country, particularly education, community-based activities and legislation were associated with the rapid increase in Japanese life expectancy after World War II.
Factors Associated With Surgery Clerkship Performance and Subsequent USMLE Step Scores.
Dong, Ting; Copeland, Annesley; Gangidine, Matthew; Schreiber-Gregory, Deanna; Ritter, E Matthew; Durning, Steven J
2018-03-12
We conducted an in-depth empirical investigation to achieve a better understanding of the surgery clerkship from multiple perspectives, including the influence of clerkship sequence on performance, the relationship between self-logged work hours and performance, as well as the association between surgery clerkship performance with subsequent USMLE Step exams' scores. The study cohort consisted of medical students graduating between 2015 and 2018 (n = 687). The primary measures of interest were clerkship sequence (internal medicine clerkship before or after surgery clerkship), self-logged work hours during surgery clerkship, surgery NBME subject exam score, surgery clerkship overall grade, and Step 1, Step 2 CK, and Step 3 exam scores. We reported the descriptive statistics and conducted correlation analysis, stepwise linear regression analysis, and variable selection analysis of logistic regression to answer the research questions. Students who completed internal medicine clerkship prior to surgery clerkship had better performance on surgery subject exam. The subject exam score explained an additional 28% of the variance of the Step 2 CK score, and the clerkship overall score accounted for an additional 24% of the variance after the MCAT scores and undergraduate GPA were controlled. Our finding suggests that the clerkship sequence does matter when it comes to performance on the surgery NBME subject exam. Performance on the surgery subject exam is predictive of subsequent performance on future USMLE Step exams. Copyright © 2018 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Regression Analysis of Top of Descent Location for Idle-thrust Descents
NASA Technical Reports Server (NTRS)
Stell, Laurel; Bronsvoort, Jesper; McDonald, Greg
2013-01-01
In this paper, multiple regression analysis is used to model the top of descent (TOD) location of user-preferred descent trajectories computed by the flight management system (FMS) on over 1000 commercial flights into Melbourne, Australia. The independent variables cruise altitude, final altitude, cruise Mach, descent speed, wind, and engine type were also recorded or computed post-operations. Both first-order and second-order models are considered, where cross-validation, hypothesis testing, and additional analysis are used to compare models. This identifies the models that should give the smallest errors if used to predict TOD location for new data in the future. A model that is linear in TOD altitude, final altitude, descent speed, and wind gives an estimated standard deviation of 3.9 nmi for TOD location given the trajec- tory parameters, which means about 80% of predictions would have error less than 5 nmi in absolute value. This accuracy is better than demonstrated by other ground automation predictions using kinetic models. Furthermore, this approach would enable online learning of the model. Additional data or further knowl- edge of algorithms is necessary to conclude definitively that no second-order terms are appropriate. Possible applications of the linear model are described, including enabling arriving aircraft to fly optimized descents computed by the FMS even in congested airspace. In particular, a model for TOD location that is linear in the independent variables would enable decision support tool human-machine interfaces for which a kinetic approach would be computationally too slow.
Additivity of nonlinear biomass equations
Bernard R. Parresol
2001-01-01
Two procedures that guarantee the property of additivity among the components of tree biomass and total tree biomass utilizing nonlinear functions are developed. Procedure 1 is a simple combination approach, and procedure 2 is based on nonlinear joint-generalized regression (nonlinear seemingly unrelated regressions) with parameter restrictions. Statistical theory is...
Maintenance Operations in Mission Oriented Protective Posture Level IV (MOPPIV)
1987-10-01
Repair FADAC Printed Circuit Board ............. 6 3. Data Analysis Techniques ............................. 6 a. Multiple Linear Regression... ANALYSIS /DISCUSSION ............................... 12 1. Exa-ple of Regression Analysis ..................... 12 S2. Regression results for all tasks...6 * TABLE 9. Task Grouping for Analysis ........................ 7 "TABXLE 10. Remove/Replace H60A3 Power Pack................. 8 TABLE
Wagner, Brian J.; Gorelick, Steven M.
1986-01-01
A simulation nonlinear multiple-regression methodology for estimating parameters that characterize the transport of contaminants is developed and demonstrated. Finite difference contaminant transport simulation is combined with a nonlinear weighted least squares multiple-regression procedure. The technique provides optimal parameter estimates and gives statistics for assessing the reliability of these estimates under certain general assumptions about the distributions of the random measurement errors. Monte Carlo analysis is used to estimate parameter reliability for a hypothetical homogeneous soil column for which concentration data contain large random measurement errors. The value of data collected spatially versus data collected temporally was investigated for estimation of velocity, dispersion coefficient, effective porosity, first-order decay rate, and zero-order production. The use of spatial data gave estimates that were 2–3 times more reliable than estimates based on temporal data for all parameters except velocity. Comparison of estimated linear and nonlinear confidence intervals based upon Monte Carlo analysis showed that the linear approximation is poor for dispersion coefficient and zero-order production coefficient when data are collected over time. In addition, examples demonstrate transport parameter estimation for two real one-dimensional systems. First, the longitudinal dispersivity and effective porosity of an unsaturated soil are estimated using laboratory column data. We compare the reliability of estimates based upon data from individual laboratory experiments versus estimates based upon pooled data from several experiments. Second, the simulation nonlinear regression procedure is extended to include an additional governing equation that describes delayed storage during contaminant transport. The model is applied to analyze the trends, variability, and interrelationship of parameters in a mourtain stream in northern California.
Hołda, Mateusz K; Koziej, Mateusz; Wszołek, Karolina; Pawlik, Wiesław; Krawczyk-Ożóg, Agata; Sorysz, Danuta; Łoboda, Piotr; Kuźma, Katarzyna; Kuniewicz, Marcin; Lelakowski, Jacek; Dudek, Dariusz; Klimek-Piotrowska, Wiesława
2017-10-01
The aim of this study is to provide a morphometric description of the left-sided septal pouch (LSSP), left atrial accessory appendages, and diverticula using cardiac multi-slice computed tomography (MSCT) and to compare results between patient subgroups. Two hundred and ninety four patients (42.9% females) with a mean of 69.4±13.1years of age were investigated using MSCT. The presence of the LSSP, left atrial accessory appendages, and diverticula was evaluated. Multiple logistic regression analysis was performed to check whether the presence of additional left atrial structures is associated with increased risk of atrial fibrillation and cerebrovascular accidents. At least one additional left atrial structure was present in 51.7% of patients. A single LSSP, left atrial diverticulum, and accessory appendage were present in 35.7%, 16.0%, and 4.1% of patients, respectively. After adjusting for other risk factors via multiple logistic regression, patients with LSSP are more likely to have atrial fibrillation (OR=2.00, 95% CI=1.14-3.48, p=0.01). The presence of a LSSP was found to be associated with an increased risk of transient ischemic attack using multiple logistic regression analysis after adjustment for other risk factors (OR=3.88, 95% CI=1.10-13.69, p=0.03). In conclusion LSSPs, accessory appendages, and diverticula are highly prevalent anatomic structures within the left atrium, which could be easily identified by MSCT. The presence of LSSP is associated with increased risk for atrial fibrillation and transient ischemic attack. Copyright © 2017 Elsevier B.V. All rights reserved.
Feasibility study of palm-based fuels for hybrid rocket motor applications
NASA Astrophysics Data System (ADS)
Tarmizi Ahmad, M.; Abidin, Razali; Taha, A. Latif; Anudip, Amzaryi
2018-02-01
This paper describes the combined analysis done in pure palm-based wax that can be used as solid fuel in a hybrid rocket engine. The measurement of pure palm wax calorific value was performed using a bomb calorimeter. An experimental rocket engine and static test stand facility were established. After initial measurement and calibration, repeated procedures were performed. Instrumentation supplies carried out allow fuel regression rate measurements, oxidizer mass flow rates and stearic acid rocket motors measurements. Similar tests are also carried out with stearate acid (from palm oil by-products) dissolved with nitrocellulose and bee solution. Calculated data and experiments show that rates and regression thrust can be achieved even in pure-tested palm-based wax. Additionally, palm-based wax is mixed with beeswax characterized by higher nominal melting temperatures to increase moisturizing points to higher temperatures without affecting regression rate values. Calorie measurements and ballistic experiments were performed on this new fuel formulation. This new formulation promises driving applications in a wide range of temperatures.
The arcsine is asinine: the analysis of proportions in ecology.
Warton, David I; Hui, Francis K C
2011-01-01
The arcsine square root transformation has long been standard procedure when analyzing proportional data in ecology, with applications in data sets containing binomial and non-binomial response variables. Here, we argue that the arcsine transform should not be used in either circumstance. For binomial data, logistic regression has greater interpretability and higher power than analyses of transformed data. However, it is important to check the data for additional unexplained variation, i.e., overdispersion, and to account for it via the inclusion of random effects in the model if found. For non-binomial data, the arcsine transform is undesirable on the grounds of interpretability, and because it can produce nonsensical predictions. The logit transformation is proposed as an alternative approach to address these issues. Examples are presented in both cases to illustrate these advantages, comparing various methods of analyzing proportions including untransformed, arcsine- and logit-transformed linear models and logistic regression (with or without random effects). Simulations demonstrate that logistic regression usually provides a gain in power over other methods.
A psycholinguistic database for traditional Chinese character naming.
Chang, Ya-Ning; Hsu, Chun-Hsien; Tsai, Jie-Li; Chen, Chien-Liang; Lee, Chia-Ying
2016-03-01
In this study, we aimed to provide a large-scale set of psycholinguistic norms for 3,314 traditional Chinese characters, along with their naming reaction times (RTs), collected from 140 Chinese speakers. The lexical and semantic variables in the database include frequency, regularity, familiarity, consistency, number of strokes, homophone density, semantic ambiguity rating, phonetic combinability, semantic combinability, and the number of disyllabic compound words formed by a character. Multiple regression analyses were conducted to examine the predictive powers of these variables for the naming RTs. The results demonstrated that these variables could account for a significant portion of variance (55.8%) in the naming RTs. An additional multiple regression analysis was conducted to demonstrate the effects of consistency and character frequency. Overall, the regression results were consistent with the findings of previous studies on Chinese character naming. This database should be useful for research into Chinese language processing, Chinese education, or cross-linguistic comparisons. The database can be accessed via an online inquiry system (http://ball.ling.sinica.edu.tw/namingdatabase/index.html).
NASA Technical Reports Server (NTRS)
Rummler, D. R.
1976-01-01
The results are presented of investigations to apply regression techniques to the development of methodology for creep-rupture data analysis. Regression analysis techniques are applied to the explicit description of the creep behavior of materials for space shuttle thermal protection systems. A regression analysis technique is compared with five parametric methods for analyzing three simulated and twenty real data sets, and a computer program for the evaluation of creep-rupture data is presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jenkin, Thomas J; Larson, Andrew; Ruth, Mark F
In light of the changing electricity resource mixes across the United States, an important question in electricity modeling is how additions and retirements of generation, including additions in variable renewable energy (VRE) generation could impact markets by changing hourly wholesale energy prices. Instead of using resource-intensive production cost models (PCMs) or building and using simple generator supply curves, this analysis uses a 'top-down' approach based on regression analysis of hourly historical energy and load data to estimate the impact of supply changes on wholesale electricity prices, provided the changes are not so substantial that they fundamentally alter the market andmore » dispatch-order driven behavior of non-retiring units. The rolling supply curve (RSC) method used in this report estimates the shape of the supply curve that fits historical hourly price and load data for given time intervals, such as two-weeks, and then repeats this on a rolling basis through the year. These supply curves can then be modified on an hourly basis to reflect the impact of generation retirements or additions, including VRE and then reapplied to the same load data to estimate the change in hourly electricity price. The choice of duration over which these RSCs are estimated has a significant impact on goodness of fit. For example, in PJM in 2015, moving from fitting one curve per year to 26 rolling two-week supply curves improves the standard error of the regression from 16 dollars/MWh to 6 dollars/MWh and the R-squared of the estimate from 0.48 to 0.76. We illustrate the potential use and value of the RSC method by estimating wholesale price effects under various generator retirement and addition scenarios, and we discuss potential limits of the technique, some of which are inherent. The ability to do this type of analysis is important to a wide range of market participants and other stakeholders, and it may have a role in complementing use of or providing calibrating insights to PCMs.« less
Resting-state functional magnetic resonance imaging: the impact of regression analysis.
Yeh, Chia-Jung; Tseng, Yu-Sheng; Lin, Yi-Ru; Tsai, Shang-Yueh; Huang, Teng-Yi
2015-01-01
To investigate the impact of regression methods on resting-state functional magnetic resonance imaging (rsfMRI). During rsfMRI preprocessing, regression analysis is considered effective for reducing the interference of physiological noise on the signal time course. However, it is unclear whether the regression method benefits rsfMRI analysis. Twenty volunteers (10 men and 10 women; aged 23.4 ± 1.5 years) participated in the experiments. We used node analysis and functional connectivity mapping to assess the brain default mode network by using five combinations of regression methods. The results show that regressing the global mean plays a major role in the preprocessing steps. When a global regression method is applied, the values of functional connectivity are significantly lower (P ≤ .01) than those calculated without a global regression. This step increases inter-subject variation and produces anticorrelated brain areas. rsfMRI data processed using regression should be interpreted carefully. The significance of the anticorrelated brain areas produced by global signal removal is unclear. Copyright © 2014 by the American Society of Neuroimaging.
Semi-automatic assessment of skin capillary density: proof of principle and validation.
Gronenschild, E H B M; Muris, D M J; Schram, M T; Karaca, U; Stehouwer, C D A; Houben, A J H M
2013-11-01
Skin capillary density and recruitment have been proven to be relevant measures of microvascular function. Unfortunately, the assessment of skin capillary density from movie files is very time-consuming, since this is done manually. This impedes the use of this technique in large-scale studies. We aimed to develop a (semi-) automated assessment of skin capillary density. CapiAna (Capillary Analysis) is a newly developed semi-automatic image analysis application. The technique involves four steps: 1) movement correction, 2) selection of the frame range and positioning of the region of interest (ROI), 3) automatic detection of capillaries, and 4) manual correction of detected capillaries. To gain insight into the performance of the technique, skin capillary density was measured in twenty participants (ten women; mean age 56.2 [42-72] years). To investigate the agreement between CapiAna and the classic manual counting procedure, we used weighted Deming regression and Bland-Altman analyses. In addition, intra- and inter-observer coefficients of variation (CVs), and differences in analysis time were assessed. We found a good agreement between CapiAna and the classic manual method, with a Pearson's correlation coefficient (r) of 0.95 (P<0.001) and a Deming regression coefficient of 1.01 (95%CI: 0.91; 1.10). In addition, we found no significant differences between the two methods, with an intercept of the Deming regression of 1.75 (-6.04; 9.54), while the Bland-Altman analysis showed a mean difference (bias) of 2.0 (-13.5; 18.4) capillaries/mm(2). The intra- and inter-observer CVs of CapiAna were 2.5% and 5.6% respectively, while for the classic manual counting procedure these were 3.2% and 7.2%, respectively. Finally, the analysis time for CapiAna ranged between 25 and 35min versus 80 and 95min for the manual counting procedure. We have developed a semi-automatic image analysis application (CapiAna) for the assessment of skin capillary density, which agrees well with the classic manual counting procedure, is time-saving, and has a better reproducibility as compared to the classic manual counting procedure. As a result, the use of skin capillaroscopy is feasible in large-scale studies, which importantly extends the possibilities to perform microcirculation research in humans. © 2013.
SOCIAL STABILITY AND HIV RISK BEHAVIOR: EVALUATING THE ROLE OF ACCUMULATED VULNERABILITY
German, Danielle; Latkin, Carl A.
2011-01-01
This study evaluated a cumulative and syndromic relationship among commonly co-occurring vulnerabilites (homelessness, incarceration, low-income, residential transition) in association with HIV-related risk behaviors among 635 low-income women in Baltimore. Analysis included descriptive statistics, logistic regression, latent class analysis and latent class regression. Both methods of assessing multidimensional instability showed significant associations with risk indicators. Risk of multiple partners, sex exchange, and drug use decreased significantly with each additional domain. Higher stability class membership (77%) was associated with decreased likelihood of multiple partners, exchange partners, recent drug use, and recent STI. Multidimensional social vulnerabilities were cumulatively and synergistically linked to HIV risk behavior. Independent instability measures may miss important contextual determinants of risk. Social stability offers a useful framework to understand the synergy of social vulnerabilities that shape sexual risk behavior. Social policies and programs aiming to enhance housing and overall social stability are likely to be beneficial for HIV prevention. PMID:21259043
Tewary, Sweta; Farber, Naomi
2014-01-01
Individuals with rheumatoid arthritis (RA) struggle to maintain improved functional ability and reduced pain levels. Health education emphasizing self-efficacy helps individuals to adjust with the disease outcome and progression. As a basis to develop comprehensive evidence-based patient education programs, the aim of the study was to examine the role of marriage as a predictor of pain and functional self-efficacy among individuals with RA. Review of the regression analysis did not provide support for the relationships between marital quality and self-efficacy. Relationships were not observed between marital quality, length of marriage, and self-efficacy as predicted by the first hypothesis. Additional regression analysis examination found that marital quality, length of marriage, pain, and health assessment together reported significant variance in self-efficacy. However, only health assessment significantly predicted self-efficacy. Other nonexamined variables could have influenced the independent marital quality effects. Future longitudinal studies with larger sample sizes can further validate the current findings.
Big Data Toolsets to Pharmacometrics: Application of Machine Learning for Time‐to‐Event Analysis
Gong, Xiajing; Hu, Meng
2018-01-01
Abstract Additional value can be potentially created by applying big data tools to address pharmacometric problems. The performances of machine learning (ML) methods and the Cox regression model were evaluated based on simulated time‐to‐event data synthesized under various preset scenarios, i.e., with linear vs. nonlinear and dependent vs. independent predictors in the proportional hazard function, or with high‐dimensional data featured by a large number of predictor variables. Our results showed that ML‐based methods outperformed the Cox model in prediction performance as assessed by concordance index and in identifying the preset influential variables for high‐dimensional data. The prediction performances of ML‐based methods are also less sensitive to data size and censoring rates than the Cox regression model. In conclusion, ML‐based methods provide a powerful tool for time‐to‐event analysis, with a built‐in capacity for high‐dimensional data and better performance when the predictor variables assume nonlinear relationships in the hazard function. PMID:29536640
Kawabata, Yoshinori
2012-01-01
FOLFOX6 and FOLFIRI regimens are often selected as the first- or second-line treatment for advanced or recurrent colorectal cancer. Patients are now able to undergo at-home treatment by using a portable disposable infusion pump (SUREFUSER(®)A) for continuous intravenous infusion of 5-fluorouracil (5-FU). The duration of continuous 5-FU infusion is normally set at an average of 46 h, but large variations in the duration of infusion are observed. The relationship between the total volume of the drug solution in SUREFUSER(®)A and the duration of infusion was analyzed by regression analysis. In addition, multiple regression analysis of the total volume of the drug solution, dummy variables for temperature, and duration of infusion was carried out. The duration of infusion was affected by the coefficient of viscosity of the drug solution and the ambient temperature. The composition of the drug solutions and the ambient temperature must be considered to ensure correct duration of continuous infusion.
NASA Astrophysics Data System (ADS)
Saunders, Ian; Ottemöller, Lars; Brandt, Martin B. C.; Fourie, Christoffel J. S.
2013-04-01
A relation to determine local magnitude ( M L) based on the original Richter definition is empirically derived from synthetic Wood-Anderson seismograms recorded by the South African National Seismograph Network. In total, 263 earthquakes in the distance range 10 to 1,000 km, representing 1,681 trace amplitudes measured in nanometers from synthesized Wood-Anderson records on the vertical channel were considered to derive an attenuation relation appropriate for South Africa through multiple regression analysis. Additionally, station corrections were determined for 26 stations during the regression analysis resulting in values ranging between -0.31 and 0.50. The most appropriate M L scale for South Africa from this study satisfies the equation: {M_{{{L}}}} = {{lo}}{{{g}}_{{10}}}(A) + 1.149{{lo}}{{{g}}_{{10}}}(R) + 0.00063R + 2.04 - S The anelastic attenuation term derived from this study indicates that ground motion attenuation is significantly different from Southern California but comparable with stable continental regions.
Critical Analysis of Dual-Probe Heat-Pulse Technique Applied to Measuring Thermal Diffusivity
NASA Astrophysics Data System (ADS)
Bovesecchi, G.; Coppa, P.; Corasaniti, S.; Potenza, M.
2018-07-01
The paper presents an analysis of the experimental parameters involved in application of the dual-probe heat pulse technique, followed by a critical review of methods for processing thermal response data (e.g., maximum detection and nonlinear least square regression) and the consequent obtainable uncertainty. Glycerol was selected as testing liquid, and its thermal diffusivity was evaluated over the temperature range from - 20 °C to 60 °C. In addition, Monte Carlo simulation was used to assess the uncertainty propagation for maximum detection. It was concluded that maximum detection approach to process thermal response data gives the closest results to the reference data inasmuch nonlinear regression results are affected by major uncertainties due to partial correlation between the evaluated parameters. Besides, the interpolation of temperature data with a polynomial to find the maximum leads to a systematic difference between measured and reference data, as put into evidence by the Monte Carlo simulations; through its correction, this systematic error can be reduced to a negligible value, about 0.8 %.
Standards for Standardized Logistic Regression Coefficients
ERIC Educational Resources Information Center
Menard, Scott
2011-01-01
Standardized coefficients in logistic regression analysis have the same utility as standardized coefficients in linear regression analysis. Although there has been no consensus on the best way to construct standardized logistic regression coefficients, there is now sufficient evidence to suggest a single best approach to the construction of a…
Linear regression analysis: part 14 of a series on evaluation of scientific publications.
Schneider, Astrid; Hommel, Gerhard; Blettner, Maria
2010-11-01
Regression analysis is an important statistical method for the analysis of medical data. It enables the identification and characterization of relationships among multiple factors. It also enables the identification of prognostically relevant risk factors and the calculation of risk scores for individual prognostication. This article is based on selected textbooks of statistics, a selective review of the literature, and our own experience. After a brief introduction of the uni- and multivariable regression models, illustrative examples are given to explain what the important considerations are before a regression analysis is performed, and how the results should be interpreted. The reader should then be able to judge whether the method has been used correctly and interpret the results appropriately. The performance and interpretation of linear regression analysis are subject to a variety of pitfalls, which are discussed here in detail. The reader is made aware of common errors of interpretation through practical examples. Both the opportunities for applying linear regression analysis and its limitations are presented.
An improved multiple linear regression and data analysis computer program package
NASA Technical Reports Server (NTRS)
Sidik, S. M.
1972-01-01
NEWRAP, an improved version of a previous multiple linear regression program called RAPIER, CREDUC, and CRSPLT, allows for a complete regression analysis including cross plots of the independent and dependent variables, correlation coefficients, regression coefficients, analysis of variance tables, t-statistics and their probability levels, rejection of independent variables, plots of residuals against the independent and dependent variables, and a canonical reduction of quadratic response functions useful in optimum seeking experimentation. A major improvement over RAPIER is that all regression calculations are done in double precision arithmetic.
Jansson, Bruce S; Nyamathi, Adeline; Heidemann, Gretchen; Duan, Lei; Kaplan, Charles
2015-01-01
Although literature documents the need for hospital social workers, nurses, and medical residents to engage in patient advocacy, little information exists about what predicts the extent they do so. This study aims to identify predictors of health professionals' patient advocacy engagement with respect to a broad range of patients' problems. A cross-sectional research design was employed with a sample of 94 social workers, 97 nurses, and 104 medical residents recruited from eight hospitals in Los Angeles. Bivariate correlations explored whether seven scales (Patient Advocacy Eagerness, Ethical Commitment, Skills, Tangible Support, Organizational Receptivity, Belief Other Professionals Engage, and Belief the Hospital Empowers Patients) were associated with patient advocacy engagement, measured by the validated Patient Advocacy Engagement Scale. Regression analysis examined whether these scales, when controlling for sociodemographic and setting variables, predicted patient advocacy engagement. While all seven predictor scales were significantly associated with patient advocacy engagement in correlational analyses, only Eagerness, Skills, and Belief the Hospital Empowers Patients predicted patient advocacy engagement in regression analyses. Additionally, younger professionals engaged in higher levels of patient advocacy than older professionals, and social workers engaged in greater patient advocacy than nurses. Limitations and the utility of these findings for acute-care hospitals are discussed.
Bilgili, Mehmet; Sahin, Besir; Sangun, Levent
2013-01-01
The aim of this study is to estimate the soil temperatures of a target station using only the soil temperatures of neighboring stations without any consideration of the other variables or parameters related to soil properties. For this aim, the soil temperatures were measured at depths of 5, 10, 20, 50, and 100 cm below the earth surface at eight measuring stations in Turkey. Firstly, the multiple nonlinear regression analysis was performed with the "Enter" method to determine the relationship between the values of target station and neighboring stations. Then, the stepwise regression analysis was applied to determine the best independent variables. Finally, an artificial neural network (ANN) model was developed to estimate the soil temperature of a target station. According to the derived results for the training data set, the mean absolute percentage error and correlation coefficient ranged from 1.45% to 3.11% and from 0.9979 to 0.9986, respectively, while corresponding ranges of 1.685-3.65% and 0.9988-0.9991, respectively, were obtained based on the testing data set. The obtained results show that the developed ANN model provides a simple and accurate prediction to determine the soil temperature. In addition, the missing data at the target station could be determined within a high degree of accuracy.
Figueroa, Jennifer A; Mansoor, Jim K; Allen, Roblee P; Davis, Cristina E; Walby, William F; Aksenov, Alexander A; Zhao, Weixiang; Lewis, William R; Schelegle, Edward S
2015-04-20
With ascent to altitude, certain individuals are susceptible to high altitude pulmonary edema (HAPE), which in turn can cause disability and even death. The ability to identify individuals at risk of HAPE prior to ascent is poor. The present study examined the profile of volatile organic compounds (VOC) in exhaled breath condensate (EBC) and pulmonary artery systolic pressures (PASP) before and after exposure to normobaric hypoxia (12% O2) in healthy males with and without a history of HAPE (Hx HAPE, n = 5; Control, n = 11). In addition, hypoxic ventilatory response (HVR), and PASP response to normoxic exercise were also measured. Auto-regression/partial least square regression of whole gas chromatography/mass spectrometry (GC/MS) data and binary logistic regression (BLR) of individual GC peaks and physiologic parameters resulted in models that separate individual subjects into their groups with variable success. The result of BLR analysis highlights HVR, PASP response to hypoxia and the amount of benzyl alcohol and dimethylbenzaldehyde dimethyl in expired breath as markers of HAPE history. These findings indicate the utility of EBC VOC analysis to discriminate between individuals with and without a history of HAPE and identified potential novel biomarkers that correlated with physiological responses to hypoxia.
2012-01-01
Implicit in the growing interest in patient-centered outcomes research is a growing need for better evidence regarding how responses to a given intervention or treatment may vary across patients, referred to as heterogeneity of treatment effect (HTE). A variety of methods are available for exploring HTE, each associated with unique strengths and limitations. This paper reviews a selected set of methodological approaches to understanding HTE, focusing largely but not exclusively on their uses with randomized trial data. It is oriented for the “intermediate” outcomes researcher, who may already be familiar with some methods, but would value a systematic overview of both more and less familiar methods with attention to when and why they may be used. Drawing from the biomedical, statistical, epidemiological and econometrics literature, we describe the steps involved in choosing an HTE approach, focusing on whether the intent of the analysis is for exploratory, initial testing, or confirmatory testing purposes. We also map HTE methodological approaches to data considerations as well as the strengths and limitations of each approach. Methods reviewed include formal subgroup analysis, meta-analysis and meta-regression, various types of predictive risk modeling including classification and regression tree analysis, series of n-of-1 trials, latent growth and growth mixture models, quantile regression, and selected non-parametric methods. In addition to an overview of each HTE method, examples and references are provided for further reading. By guiding the selection of the methods and analysis, this review is meant to better enable outcomes researchers to understand and explore aspects of HTE in the context of patient-centered outcomes research. PMID:23234603
NASA Astrophysics Data System (ADS)
Krivolutsky, Alexei A.; Nazarova, Margarita; Knyazeva, Galina
Solar activity influences on atmospheric photochemical system via its changebale electromag-netic flux with eleven-year period and also by energetic particles during solar proton event (SPE). Energetic particles penetrate mostly into polar regions and induce additional produc-tion of NOx and HOx chemical compounds, which can destroy ozone in photochemical catalytic cycles. Solar irradiance variations cause in-phase variability of ozone in accordance with photo-chemical theory. However, real ozone response caused by these two factors, which has different physical nature, is not so clear on long-term time scale. In order to understand the situation multiply linear regression statistical method was used. Three data series, which covered the period 1958-2006, have been used to realize such analysis: yearly averaged total ozone at dif-ferent latitudes (World Ozone Data Centre, Canada, WMO); yearly averaged proton fluxes with E¿ 10 MeV ( IMP, GOES, METEOR satellites); yearly averaged numbers of solar spots (Solar Data). Then, before the analysis, the data sets of ozone deviations from the mean values for whole period (1958-2006) at each latitudinal belt were prepared. The results of multiply regression analysis (two factors) revealed rather complicated time-dependent behavior of ozone response with clear negative peaks for the years of strong SPEs. The magnitudes of such peaks on annual mean basis are not greater than 10 DU. The unusual effect -positive response of ozone to solar proton activity near both poles-was discovered by statistical analysis. The pos-sible photochemical nature of found effect is discussed. This work was supported by Russian Science Foundation for Basic Research (grant 09-05-009949) and by the contract 1-6-08 under Russian Sub-Program "Research and Investigation of Antarctica".
Marital status integration and suicide: A meta-analysis and meta-regression.
Kyung-Sook, Woo; SangSoo, Shin; Sangjin, Shin; Young-Jeon, Shin
2018-01-01
Marital status is an index of the phenomenon of social integration within social structures and has long been identified as an important predictor suicide. However, previous meta-analyses have focused only on a particular marital status, or not sufficiently explored moderators. A meta-analysis of observational studies was conducted to explore the relationships between marital status and suicide and to understand the important moderating factors in this association. Electronic databases were searched to identify studies conducted between January 1, 2000 and June 30, 2016. We performed a meta-analysis, subgroup analysis, and meta-regression of 170 suicide risk estimates from 36 publications. Using random effects model with adjustment for covariates, the study found that the suicide risk for non-married versus married was OR = 1.92 (95% CI: 1.75-2.12). The suicide risk was higher for non-married individuals aged <65 years than for those aged ≥65 years, and higher for men than for women. According to the results of stratified analysis by gender, non-married men exhibited a greater risk of suicide than their married counterparts in all sub-analyses, but women aged 65 years or older showed no significant association between marital status and suicide. The suicide risk in divorced individuals was higher than for non-married individuals in both men and women. The meta-regression showed that gender, age, and sample size affected between-study variation. The results of the study indicated that non-married individuals have an aggregate higher suicide risk than married ones. In addition, gender and age were confirmed as important moderating factors in the relationship between marital status and suicide. Copyright © 2017 Elsevier Ltd. All rights reserved.
A secure distributed logistic regression protocol for the detection of rare adverse drug events
El Emam, Khaled; Samet, Saeed; Arbuckle, Luk; Tamblyn, Robyn; Earle, Craig; Kantarcioglu, Murat
2013-01-01
Background There is limited capacity to assess the comparative risks of medications after they enter the market. For rare adverse events, the pooling of data from multiple sources is necessary to have the power and sufficient population heterogeneity to detect differences in safety and effectiveness in genetic, ethnic and clinically defined subpopulations. However, combining datasets from different data custodians or jurisdictions to perform an analysis on the pooled data creates significant privacy concerns that would need to be addressed. Existing protocols for addressing these concerns can result in reduced analysis accuracy and can allow sensitive information to leak. Objective To develop a secure distributed multi-party computation protocol for logistic regression that provides strong privacy guarantees. Methods We developed a secure distributed logistic regression protocol using a single analysis center with multiple sites providing data. A theoretical security analysis demonstrates that the protocol is robust to plausible collusion attacks and does not allow the parties to gain new information from the data that are exchanged among them. The computational performance and accuracy of the protocol were evaluated on simulated datasets. Results The computational performance scales linearly as the dataset sizes increase. The addition of sites results in an exponential growth in computation time. However, for up to five sites, the time is still short and would not affect practical applications. The model parameters are the same as the results on pooled raw data analyzed in SAS, demonstrating high model accuracy. Conclusion The proposed protocol and prototype system would allow the development of logistic regression models in a secure manner without requiring the sharing of personal health information. This can alleviate one of the key barriers to the establishment of large-scale post-marketing surveillance programs. We extended the secure protocol to account for correlations among patients within sites through generalized estimating equations, and to accommodate other link functions by extending it to generalized linear models. PMID:22871397
A secure distributed logistic regression protocol for the detection of rare adverse drug events.
El Emam, Khaled; Samet, Saeed; Arbuckle, Luk; Tamblyn, Robyn; Earle, Craig; Kantarcioglu, Murat
2013-05-01
There is limited capacity to assess the comparative risks of medications after they enter the market. For rare adverse events, the pooling of data from multiple sources is necessary to have the power and sufficient population heterogeneity to detect differences in safety and effectiveness in genetic, ethnic and clinically defined subpopulations. However, combining datasets from different data custodians or jurisdictions to perform an analysis on the pooled data creates significant privacy concerns that would need to be addressed. Existing protocols for addressing these concerns can result in reduced analysis accuracy and can allow sensitive information to leak. To develop a secure distributed multi-party computation protocol for logistic regression that provides strong privacy guarantees. We developed a secure distributed logistic regression protocol using a single analysis center with multiple sites providing data. A theoretical security analysis demonstrates that the protocol is robust to plausible collusion attacks and does not allow the parties to gain new information from the data that are exchanged among them. The computational performance and accuracy of the protocol were evaluated on simulated datasets. The computational performance scales linearly as the dataset sizes increase. The addition of sites results in an exponential growth in computation time. However, for up to five sites, the time is still short and would not affect practical applications. The model parameters are the same as the results on pooled raw data analyzed in SAS, demonstrating high model accuracy. The proposed protocol and prototype system would allow the development of logistic regression models in a secure manner without requiring the sharing of personal health information. This can alleviate one of the key barriers to the establishment of large-scale post-marketing surveillance programs. We extended the secure protocol to account for correlations among patients within sites through generalized estimating equations, and to accommodate other link functions by extending it to generalized linear models.
[A SAS marco program for batch processing of univariate Cox regression analysis for great database].
Yang, Rendong; Xiong, Jie; Peng, Yangqin; Peng, Xiaoning; Zeng, Xiaomin
2015-02-01
To realize batch processing of univariate Cox regression analysis for great database by SAS marco program. We wrote a SAS macro program, which can filter, integrate, and export P values to Excel by SAS9.2. The program was used for screening survival correlated RNA molecules of ovarian cancer. A SAS marco program could finish the batch processing of univariate Cox regression analysis, the selection and export of the results. The SAS macro program has potential applications in reducing the workload of statistical analysis and providing a basis for batch processing of univariate Cox regression analysis.
Multiple regression technique for Pth degree polynominals with and without linear cross products
NASA Technical Reports Server (NTRS)
Davis, J. W.
1973-01-01
A multiple regression technique was developed by which the nonlinear behavior of specified independent variables can be related to a given dependent variable. The polynomial expression can be of Pth degree and can incorporate N independent variables. Two cases are treated such that mathematical models can be studied both with and without linear cross products. The resulting surface fits can be used to summarize trends for a given phenomenon and provide a mathematical relationship for subsequent analysis. To implement this technique, separate computer programs were developed for the case without linear cross products and for the case incorporating such cross products which evaluate the various constants in the model regression equation. In addition, the significance of the estimated regression equation is considered and the standard deviation, the F statistic, the maximum absolute percent error, and the average of the absolute values of the percent of error evaluated. The computer programs and their manner of utilization are described. Sample problems are included to illustrate the use and capability of the technique which show the output formats and typical plots comparing computer results to each set of input data.
Fei, Yang; Hu, Jian; Gao, Kun; Tu, Jianfeng; Li, Wei-Qin; Wang, Wei
2017-06-01
To construct a radical basis function (RBF) artificial neural networks (ANNs) model to predict the incidence of acute pancreatitis (AP)-induced portal vein thrombosis. The analysis included 353 patients with AP who had admitted between January 2011 and December 2015. RBF ANNs model and logistic regression model were constructed based on eleven factors relevant to AP respectively. Statistical indexes were used to evaluate the value of the prediction in two models. The predict sensitivity, specificity, positive predictive value, negative predictive value and accuracy by RBF ANNs model for PVT were 73.3%, 91.4%, 68.8%, 93.0% and 87.7%, respectively. There were significant differences between the RBF ANNs and logistic regression models in these parameters (P<0.05). In addition, a comparison of the area under receiver operating characteristic curves of the two models showed a statistically significant difference (P<0.05). The RBF ANNs model is more likely to predict the occurrence of PVT induced by AP than logistic regression model. D-dimer, AMY, Hct and PT were important prediction factors of approval for AP-induced PVT. Copyright © 2017 Elsevier Inc. All rights reserved.
Exact Analysis of Squared Cross-Validity Coefficient in Predictive Regression Models
ERIC Educational Resources Information Center
Shieh, Gwowen
2009-01-01
In regression analysis, the notion of population validity is of theoretical interest for describing the usefulness of the underlying regression model, whereas the presumably more important concept of population cross-validity represents the predictive effectiveness for the regression equation in future research. It appears that the inference…
USDA-ARS?s Scientific Manuscript database
Selective principal component regression analysis (SPCR) uses a subset of the original image bands for principal component transformation and regression. For optimal band selection before the transformation, this paper used genetic algorithms (GA). In this case, the GA process used the regression co...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fellers, R.S.; Braly, L.B.; Saykally, R.J.
The SWPS method is improved by the addition of H.E.G. contractions for generating a more compact basis. An error in the definition of the internal fragment axis system used in our previous calculation is described and corrected. Fully coupled 6D (rigid monomers) VRT states are computed for several new water dimer potential surfaces and compared with experiment and our earlier SWPS results. This work sets the stage for refinement of such potential surfaces via regression analysis of VRT spectroscopic data. {copyright} {ital 1999 American Institute of Physics.}
Tolerance of ciliated protozoan Paramecium bursaria (Protozoa, Ciliophora) to ammonia and nitrites
NASA Astrophysics Data System (ADS)
Xu, Henglong; Song, Weibo; Lu, Lu; Alan, Warren
2005-09-01
The tolerance to ammonia and nitrites in freshwater ciliate Paramecium bursaria was measured in a conventional open system. The ciliate was exposed to different concentrations of ammonia and nitrites for 2h and 12h in order to determine the lethal concentrations. Linear regression analysis revealed that the 2h-LC50 value for ammonia was 95.94 mg/L and for nitrite 27.35 mg/L using probit scale method (with 95% confidence intervals). There was a linear correlation between the mortality probit scale and logarithmic concentration of ammonia which fit by a regression equation y=7.32 x 9.51 ( R 2=0.98; y, mortality probit scale; x, logarithmic concentration of ammonia), by which 2 h-LC50 value for ammonia was found to be 95.50 mg/L. A linear correlation between mortality probit scales and logarithmic concentration of nitrite is also followed the regression equation y=2.86 x+0.89 ( R 2=0.95; y, mortality probit scale; x, logarithmic concentration of nitrite). The regression analysis of toxicity curves showed that the linear correlation between exposed time of ammonia-N LC50 value and ammonia-N LC50 value followed the regression equation y=2 862.85 e -0.08 x ( R 2=0.95; y, duration of exposure to LC50 value; x, LC50 value), and that between exposed time of nitrite-N LC50 value and nitrite-N LC50 value followed the regression equation y=127.15 e -0.13 x ( R 2=0.91; y, exposed time of LC50 value; x, LC50 value). The results demonstrate that the tolerance to ammonia in P. bursaria is considerably higher than that of the larvae or juveniles of some metozoa, e.g. cultured prawns and oysters. In addition, ciliates, as bacterial predators, are likely to play a positive role in maintaining and improving water quality in aquatic environments with high-level ammonium, such as sewage treatment systems.
Ahearn, Elizabeth A.
2010-01-01
Multiple linear regression equations for determining flow-duration statistics were developed to estimate select flow exceedances ranging from 25- to 99-percent for six 'bioperiods'-Salmonid Spawning (November), Overwinter (December-February), Habitat Forming (March-April), Clupeid Spawning (May), Resident Spawning (June), and Rearing and Growth (July-October)-in Connecticut. Regression equations also were developed to estimate the 25- and 99-percent flow exceedances without reference to a bioperiod. In total, 32 equations were developed. The predictive equations were based on regression analyses relating flow statistics from streamgages to GIS-determined basin and climatic characteristics for the drainage areas of those streamgages. Thirty-nine streamgages (and an additional 6 short-term streamgages and 28 partial-record sites for the non-bioperiod 99-percent exceedance) in Connecticut and adjacent areas of neighboring States were used in the regression analysis. Weighted least squares regression analysis was used to determine the predictive equations; weights were assigned based on record length. The basin characteristics-drainage area, percentage of area with coarse-grained stratified deposits, percentage of area with wetlands, mean monthly precipitation (November), mean seasonal precipitation (December, January, and February), and mean basin elevation-are used as explanatory variables in the equations. Standard errors of estimate of the 32 equations ranged from 10.7 to 156 percent with medians of 19.2 and 55.4 percent to predict the 25- and 99-percent exceedances, respectively. Regression equations to estimate high and median flows (25- to 75-percent exceedances) are better predictors (smaller variability of the residual values around the regression line) than the equations to estimate low flows (less than 75-percent exceedance). The Habitat Forming (March-April) bioperiod had the smallest standard errors of estimate, ranging from 10.7 to 20.9 percent. In contrast, the Rearing and Growth (July-October) bioperiod had the largest standard errors, ranging from 30.9 to 156 percent. The adjusted coefficient of determination of the equations ranged from 77.5 to 99.4 percent with medians of 98.5 and 90.6 percent to predict the 25- and 99-percent exceedances, respectively. Descriptive information on the streamgages used in the regression, measured basin and climatic characteristics, and estimated flow-duration statistics are provided in this report. Flow-duration statistics and the 32 regression equations for estimating flow-duration statistics in Connecticut are stored on the U.S. Geological Survey World Wide Web application ?StreamStats? (http://water.usgs.gov/osw/streamstats/index.html). The regression equations developed in this report can be used to produce unbiased estimates of select flow exceedances statewide.
Boosting structured additive quantile regression for longitudinal childhood obesity data.
Fenske, Nora; Fahrmeir, Ludwig; Hothorn, Torsten; Rzehak, Peter; Höhle, Michael
2013-07-25
Childhood obesity and the investigation of its risk factors has become an important public health issue. Our work is based on and motivated by a German longitudinal study including 2,226 children with up to ten measurements on their body mass index (BMI) and risk factors from birth to the age of 10 years. We introduce boosting of structured additive quantile regression as a novel distribution-free approach for longitudinal quantile regression. The quantile-specific predictors of our model include conventional linear population effects, smooth nonlinear functional effects, varying-coefficient terms, and individual-specific effects, such as intercepts and slopes. Estimation is based on boosting, a computer intensive inference method for highly complex models. We propose a component-wise functional gradient descent boosting algorithm that allows for penalized estimation of the large variety of different effects, particularly leading to individual-specific effects shrunken toward zero. This concept allows us to flexibly estimate the nonlinear age curves of upper quantiles of the BMI distribution, both on population and on individual-specific level, adjusted for further risk factors and to detect age-varying effects of categorical risk factors. Our model approach can be regarded as the quantile regression analog of Gaussian additive mixed models (or structured additive mean regression models), and we compare both model classes with respect to our obesity data.
Metal hydride and pyrophoric fuel additives for dicyclopentadiene based hybrid propellants
NASA Astrophysics Data System (ADS)
Shark, Steven C.
The purpose of this study is to investigate the use of reactive energetic fuel additives that have the potential to increase the combustion performance of hybrid rocket propellants in terms of solid fuel regression rate and combustion efficiency. Additives that can augment the combustion flame zone in a hybrid rocket motor by means of increased energy feedback to the fuel grain surface are of great interest. Metal hydrides have large volumetric hydrogen densities, which gives these materials high performance potential as fuel additives in terms of specifc impulse. The excess hydrogen and corresponding base metal may also cause an increase in the hybrid rocket solid fuel regression rate. Pyrophoric additives also have potential to increase the solid fuel regression rate by reacting more readily near the burning fuel surface providing rapid energy feedback. An experimental performance evaluation of metal hydride fuel additives for hybrid rocket motor propulsion systems is examined in this study. Hypergolic ignition droplet tests and an accelerated aging study revealed the protection capabilities of Dicyclopentadiene (DCPD) as a fuel binder, and the ability for unaided ignition. Static hybrid rocket motor experiments were conducted using DCPD as the fuel. Sodium borohydride (NabH4) and aluminum hydride (AlH3) were examined as fuel additives. Ninety percent rocket grade hydrogen peroxide (RGHP) was used as the oxidizer. In this study, the sensitivity of solid fuel regression rate and characteristic velocity (C*) efficiency to total fuel grain port mass flux and particle loading is examined. These results were compared to HTPB combustion performance as a baseline. Chamber pressure histories revealed steady motor operation in most tests, with reduced ignition delays when using NabH4 as a fuel additive. The addition of NabH4 and AlH3 produced up to a 47% and 85% increase in regression rate over neat DCPD, respectively. For all test conditions examined C* efficiency ranges between 80% and 90%. The regression rate and C* efficiency mass flux dependence indicate a shift towards a more diffusion controlled system with metal hydride particle addition. Although these types of energetic particles have potential as high performing fuel additives, they can be in low supply and expensive. An opposed flow burner was investigated as a means to screen and characterize hybrid rocket fuels prior to full scale rocket motor testing. Although this type of configuration has been investigated in the past, no comparison has been made to hybrid rocket motor operation in terms of mass flux. Polymeric fuels and low melt temperature fuels with and without additives were investigated via an opposed flow burner. The effects of laminar and turbulent flow regimes on the convective heat transfer in the opposed flow system was depicted in the regression rate trends of these fuels. Regression rate trends similar to hybrid rocket motor operation were depicted, including the entrainment mechanism for paran fuel. However, there was a shift in overall magnitude of these results. A decrease in regression rate occurred for HTPB loaded with passivated nano-aluminum, due to low resonance time in the reaction zone. Previous results have shown that pyrophoric additives can cause an increase in regression rate in the opposed flow burner configuration. It is proposed that the opposed burner is useful as a screening and characterization tool for some propellant combinations. Gaseous oxygen (GOX) was investigated as an oxidizer for similar fuels evaluated with RGHP. Specifically, combustion performance sensitivity to mass flux and MH particle size was investigated. Similar results to the RGHP experiments were observed for the regression rate tends of HTPB, DPCD, and NabH 4 addition. Kinetically limited regression rate dependence on mass flux was observed at the higher mass flux levels. No major increase in C* efficiency was observed for MH addition. The C* efficiency varied with equivalence ratio by approximately 10 percentage points, which was not observed in the RGHP experiments. A 10 percentage point decrease in C* efficiency was observed with increasing mass flux in the system. This was most likely due to poorly mixed fuel and oxidizer in center of the combustion chamber at the higher mass flux levels. Detailed measurements of the hybrid rocket combustion zone is useful for understanding the mechanisms governing performance, but can be difficult to obtain. Traditional slab burner configurations have proven useful but are operationally limited in pressure and mass flux ranges. A new optical cylindrical combustor (OCC) design is presented that allows surface and flame zone imaging and tracking during hybrid rocket motor operation at appreciable mass flux and pressure levels, > 100 kg/s/m2 and > 0.69 MPa. The flame height and regression rate sensitivity to mass flux and chamber pressure was examined for the same fuels examined in the GOX hybrid rocket motor, with the addition of DCPD fuel loaded with Al and unpassivated mechanically activated Al-PTFE. The regression rate trends were on the same order of magnitude of traditional hybrid rocket motor results. A flame height decrease was observed for increased mass flux. The flame height increased with NabH 4 addition, which is most likely a function of increased blowing at the surface. There was no appreciable flame height sensitivity to NabH4 particle size. There was no relative change in flame height or regression rate between the Al and AL-PTFE addition. The OCC allowed visualization of the hybrid rocket fuel flame zone at mass flux and pressure levels that are not known to be report for traditional slab burner configurations in literature. The OCC proved to be a new useful tool for investigated hybrid rocket propellant combustion characteristics.
Rasmussen, Patrick P.; Gray, John R.; Glysson, G. Douglas; Ziegler, Andrew C.
2009-01-01
In-stream continuous turbidity and streamflow data, calibrated with measured suspended-sediment concentration data, can be used to compute a time series of suspended-sediment concentration and load at a stream site. Development of a simple linear (ordinary least squares) regression model for computing suspended-sediment concentrations from instantaneous turbidity data is the first step in the computation process. If the model standard percentage error (MSPE) of the simple linear regression model meets a minimum criterion, this model should be used to compute a time series of suspended-sediment concentrations. Otherwise, a multiple linear regression model using paired instantaneous turbidity and streamflow data is developed and compared to the simple regression model. If the inclusion of the streamflow variable proves to be statistically significant and the uncertainty associated with the multiple regression model results in an improvement over that for the simple linear model, the turbidity-streamflow multiple linear regression model should be used to compute a suspended-sediment concentration time series. The computed concentration time series is subsequently used with its paired streamflow time series to compute suspended-sediment loads by standard U.S. Geological Survey techniques. Once an acceptable regression model is developed, it can be used to compute suspended-sediment concentration beyond the period of record used in model development with proper ongoing collection and analysis of calibration samples. Regression models to compute suspended-sediment concentrations are generally site specific and should never be considered static, but they represent a set period in a continually dynamic system in which additional data will help verify any change in sediment load, type, and source.
Regression Analysis and the Sociological Imagination
ERIC Educational Resources Information Center
De Maio, Fernando
2014-01-01
Regression analysis is an important aspect of most introductory statistics courses in sociology but is often presented in contexts divorced from the central concerns that bring students into the discipline. Consequently, we present five lesson ideas that emerge from a regression analysis of income inequality and mortality in the USA and Canada.
Pieterse, Alex L; Carter, Robert T; Evans, Sarah A; Walter, Rebecca A
2010-07-01
In this study, we examined the association among perceptions of racial and/or ethnic discrimination, racial climate, and trauma-related symptoms among 289 racially diverse college undergraduates. Study measures included the Perceived Stress Scale, the Perceived Ethnic Discrimination Questionnaire, the Posttraumatic Stress Disorder Checklist-Civilian Version, and the Racial Climate Scale. Results of a multivariate analysis of variance (MANOVA) indicated that Asian and Black students reported more frequent experiences of discrimination than did White students. Additionally, the MANOVA indicated that Black students perceived the campus racial climate as being more negative than did White and Asian students. A hierarchical regression analysis showed that when controlling for generic life stress, perceptions of discrimination contributed an additional 10% of variance in trauma-related symptoms for Black students, and racial climate contributed an additional 7% of variance in trauma symptoms for Asian students. (c) 2010 APA, all rights reserved.
Spatial-temporal event detection in climate parameter imagery.
DOE Office of Scientific and Technical Information (OSTI.GOV)
McKenna, Sean Andrew; Gutierrez, Karen A.
Previously developed techniques that comprise statistical parametric mapping, with applications focused on human brain imaging, are examined and tested here for new applications in anomaly detection within remotely-sensed imagery. Two approaches to analysis are developed: online, regression-based anomaly detection and conditional differences. These approaches are applied to two example spatial-temporal data sets: data simulated with a Gaussian field deformation approach and weekly NDVI images derived from global satellite coverage. Results indicate that anomalies can be identified in spatial temporal data with the regression-based approach. Additionally, la Nina and el Nino climatic conditions are used as different stimuli applied to themore » earth and this comparison shows that el Nino conditions lead to significant decreases in NDVI in both the Amazon Basin and in Southern India.« less
Araki, Kenichiro; Shirabe, Ken; Watanabe, Akira; Kubo, Norio; Sasaki, Shigeru; Suzuki, Hideki; Asao, Takayuki; Kuwano, Hiroyuki
2017-01-01
Although single-incision laparoscopic cholecystectomy is now widely performed in patients with cholecystitis, some cases require an additional port to complete the procedure. In this study, we focused on risk factor of additional port in this surgery. We performed single-incision cholecystectomy in 75 patients with acute cholecystitis or after cholecystitis between 2010 and 2014 at Gunma University Hospital. Surgical indications followed the TG13 guidelines. Our standard procedure for single-incision cholecystectomy routinely uses two needlescopic devices. We used logistic regression analysis to identify the risk factors associated with use of an additional full-size port (5 or 10 mm). Surgical outcome was acceptable without biliary injury. Nine patients (12.0%) required an additional port, and one patient (1.3%) required conversion to open cholecystectomy because of severe adhesions around the cystic duct and common bile duct. In multivariate analysis, high C-reactive protein (CRP) values (>7.0 mg/dl) during cholecystitis attacks were significantly correlated with the need for an additional port (P = 0.009), with a sensitivity of 55.6%, specificity of 98.5%, and accuracy of 93.3%. This study indicated that the severe inflammation indicated by high CRP values during cholecystitis attacks predicts the need for an additional port. J. Med. Invest. 64: 245-249, August, 2017.
Cox regression analysis with missing covariates via nonparametric multiple imputation.
Hsu, Chiu-Hsieh; Yu, Mandi
2018-01-01
We consider the situation of estimating Cox regression in which some covariates are subject to missing, and there exists additional information (including observed event time, censoring indicator and fully observed covariates) which may be predictive of the missing covariates. We propose to use two working regression models: one for predicting the missing covariates and the other for predicting the missing probabilities. For each missing covariate observation, these two working models are used to define a nearest neighbor imputing set. This set is then used to non-parametrically impute covariate values for the missing observation. Upon the completion of imputation, Cox regression is performed on the multiply imputed datasets to estimate the regression coefficients. In a simulation study, we compare the nonparametric multiple imputation approach with the augmented inverse probability weighted (AIPW) method, which directly incorporates the two working models into estimation of Cox regression, and the predictive mean matching imputation (PMM) method. We show that all approaches can reduce bias due to non-ignorable missing mechanism. The proposed nonparametric imputation method is robust to mis-specification of either one of the two working models and robust to mis-specification of the link function of the two working models. In contrast, the PMM method is sensitive to misspecification of the covariates included in imputation. The AIPW method is sensitive to the selection probability. We apply the approaches to a breast cancer dataset from Surveillance, Epidemiology and End Results (SEER) Program.
Multivariate Regression Analysis and Slaughter Livestock,
AGRICULTURE, *ECONOMICS), (*MEAT, PRODUCTION), MULTIVARIATE ANALYSIS, REGRESSION ANALYSIS , ANIMALS, WEIGHT, COSTS, PREDICTIONS, STABILITY, MATHEMATICAL MODELS, STORAGE, BEEF, PORK, FOOD, STATISTICAL DATA, ACCURACY
Dai, Quxiu; Ma, Liping; Ren, Nanqi; Ning, Ping; Guo, Zhiying; Xie, Longgui; Gao, Haijun
2018-06-06
Modified phosphogypsum (MPG) was developed to improve dewaterability of sewage sludge, and dewatering performance, properties of treated sludge, composition and morphology distribution of EPS, dynamic analysis and multiple regression model on bound water release were investigated. The results showed that addition of MPG caused extracellular polymeric substances (EPS) disintegration through charge neutralization. Destruction of EPS promoted the formation of larger sludge flocs and the release of bound water into supernatant. Simultaneously, content of organics with molecular weight between 1000 and 7000 Da in soluble EPS (SB-EPS) increased with increasing of EPS dissolved into the liquid phase. Besides, about 8.8 kg•kg -1 DS of bound water was released after pretreatment with 40%DS MPG dosage. Additionally, a multiple linear regression model for bound water release was established, showing that lower loosely bond EPS (LB-EPS) content and specific resistance of filtration (SRF) may improve dehydration performance, and larger sludge flocs may be beneficial for sludge dewatering. Copyright © 2018 Elsevier Ltd. All rights reserved.
Regression Analysis: Legal Applications in Institutional Research
ERIC Educational Resources Information Center
Frizell, Julie A.; Shippen, Benjamin S., Jr.; Luna, Andrew L.
2008-01-01
This article reviews multiple regression analysis, describes how its results should be interpreted, and instructs institutional researchers on how to conduct such analyses using an example focused on faculty pay equity between men and women. The use of multiple regression analysis will be presented as a method with which to compare salaries of…
RAWS II: A MULTIPLE REGRESSION ANALYSIS PROGRAM,
This memorandum gives instructions for the use and operation of a revised version of RAWS, a multiple regression analysis program. The program...of preprocessed data, the directed retention of variable, listing of the matrix of the normal equations and its inverse, and the bypassing of the regression analysis to provide the input variable statistics only. (Author)
Central Plant Optimization for Waste Energy Reduction (CPOWER). ESTCP Cost and Performance Report
2016-12-01
in the regression models. The solar radiation data did not appear reliable in the weather dataset for the location, and hence it was not used. The...and additional factors (e.g., solar insolation) may be needed to obtain a better model. 2. Inputs to optimizer: During several periods of...Location: North Carolina Energy Consumption Cost Savings $ 443,698.00 Analysis Type: FEMP PV of total savings 215,698.00$ Base Date: April 1
Increased copy number of the DLX4 homeobox gene in breast axillary lymph node metastasis
Torresan, Clarissa; Oliveira, Márcia M.C.; Pereira, Silma R.F.; Ribeiro, Enilze M.S.F.; Marian, Catalin; Gusev, Yuriy; Lima, Rubens S.; Urban, Cicero A.; Berg, Patricia E.; Haddad, Bassem R.; Cavalli, Iglenir J.; Cavalli, Luciane R.
2017-01-01
DLX4 is a homeobox gene strongly implicated in breast tumor progression and invasion. Our main objective was to determine the DLX4 copy number status in sentinel lymph node (SLN) metastasis to assess its involvement in the initial stages of the axillary metastatic process. A total of 37 paired samples of SLN metastasis and primary breast tumors (PBT) were evaluated by fluorescence in situ hybridization, quantitative polymerase chain reaction and array comparative genomic hybridization assays. DLX4 increased copy number was observed in 21.6% of the PBT and 24.3% of the SLN metastasis; regression analysis demonstrated that the DLX4 alterations observed in the SLN metastasis were dependent on the ones in the PBT, indicating that they occur in the primary tumor cell populations and are maintained in the early axillary metastatic site. In addition, regression analysis demonstrated that DLX4 alterations (and other DLX and HOXB family members) occurred independently of the ones in the HER2/NEU gene, the main amplification driver on the 17q region. Additional studies evaluating DLX4 copy number in non-SLN axillary lymph nodes and/or distant breast cancer metastasis are necessary to determine if these alterations are carried on and maintained during more advanced stages of tumor progression and if could be used as a predictive marker for axillary involvement. PMID:24947980
Gracia-Marco, Luis; Tomas, Concepcion; Vicente-Rodriguez, German; Jimenez-Pavon, David; Rey-Lopez, Juan P; Ortega, Francisco B; Lanza-Saiz, Ricardo; Moreno, Luis A
2010-11-01
The aims of this study were to identify differences between the sexes in extra-curricular participation in sports and to determine its association with body fat and socio-demographic factors in Spanish adolescents. A total of 2165 adolescents (1124 males and 1041 females) aged 13.0-18.5 years from the AVENA Study participated. Participants filled in an ad hoc questionnaire for extra-curricular participation in sports, which was the dependent variable. Independent variables were: age, percent body fat, and father's and mother's educational level and occupation. Chi-square tests and logistic regression were applied. Bivariate analysis showed for male adolescents that age and father's occupation were related to extra-curricular participation in sports. In addition, body fat and mother's education and occupation (all P < 0.05) were related to extra-curricular participation of in sports for female adolescents. Logistic regression analysis showed that the likelihood of involvement in extra-curricular participation in sports was 5.3-fold (3.86-7.38) higher for males than females. Age and father's education in both males and females were independently associated with extra-curricular participation in sports. In summary, Spanish male adolescents were shown to engage in more extra-curricular sports than females. In addition, age and father's education (in both sexes) were associated with the participation of their offspring in extra-curricular sports during adolescence.
Modeling of Micro Deval abrasion loss based on some rock properties
NASA Astrophysics Data System (ADS)
Capik, Mehmet; Yilmaz, Ali Osman
2017-10-01
Aggregate is one of the most widely used construction material. The quality of the aggregate is determined using some testing methods. Among these methods, the Micro Deval Abrasion Loss (MDAL) test is commonly used for the determination of the quality and the abrasion resistance of aggregate. The main objective of this study is to develop models for the prediction of MDAL from rock properties, including uniaxial compressive strength, Brazilian tensile strength, point load index, Schmidt rebound hardness, apparent porosity, void ratio Cerchar abrasivity index and Bohme abrasion test are examined. Additionally, the MDAL is modeled using simple regression analysis and multiple linear regression analysis based on the rock properties. The study shows that the MDAL decreases with the increase of uniaxial compressive strength, Brazilian tensile strength, point load index, Schmidt rebound hardness and Cerchar abrasivity index. It is also concluded that the MDAL increases with the increase of apparent porosity, void ratio and Bohme abrasion test. The modeling results show that the models based on Bohme abrasion test and L type Schmidt rebound hardness give the better forecasting performances for the MDAL. More models, including the uniaxial compressive strength, the apparent porosity and Cerchar abrasivity index, are developed for the rapid estimation of the MDAL of the rocks. The developed models were verified by statistical tests. Additionally, it can be stated that the proposed models can be used as a forecasting for aggregate quality.
Wang, D Z; Wang, C; Shen, C F; Zhang, Y; Zhang, H; Song, G D; Xue, X D; Xu, Z L; Zhang, S; Jiang, G H
2017-05-10
We described the time trend of acute myocardial infarction (AMI) from 1999 to 2013 in Tianjin incidence rate with Cochran-Armitage trend (CAT) test and linear regression analysis, and the results were compared. Based on actual population, CAT test had much stronger statistical power than linear regression analysis for both overall incidence trend and age specific incidence trend (Cochran-Armitage trend P value
Vanderick, S; Harris, B L; Pryce, J E; Gengler, N
2009-03-01
In New Zealand, a large proportion of cows are currently crossbreds, mostly Holstein-Friesians (HF) x Jersey (JE). The genetic evaluation system for milk yields is considering the same additive genetic effects for all breeds. The objective was to model different additive effects according to parental breeds to obtain first estimates of correlations among breed-specific effects and to study the usefulness of this type of random regression test-day model. Estimates of (co)variance components for purebred HF and JE cattle in purebred herds were computed by using a single-breed model. This analysis showed differences between the 2 breeds, with a greater variability in the HF breed. (Co)variance components for purebred HF and JE and crossbred HF x JE cattle were then estimated by using a complete multibreed model in which computations of complete across-breed (co)variances were simplified by correlating only eigenvectors for HF and JE random regressions of the same order as obtained from the single-breed analysis. Parameter estimates differed more strongly than expected between the single-breed and multibreed analyses, especially for JE. This could be due to differences between animals and management in purebred and non-purebred herds. In addition, the model used only partially accounted for heterosis. The multibreed analysis showed additive genetic differences between the HF and JE breeds, expressed as genetic correlations of additive effects in both breeds, especially in linear and quadratic Legendre polynomials (respectively, 0.807 and 0.604). The differences were small for overall milk production (0.926). Results showed that permanent environmental lactation curves were highly correlated across breeds; however, intraherd lactation curves were also affected by the breed-environment interaction. This result may indicate the existence of breed-specific competition effects that vary through the different lactation stages. In conclusion, a multibreed model similar to the one presented could optimally use the environmental and genetic parameters and provide breed-dependent additive breeding values. This model could also be a useful tool to evaluate crossbred dairy cattle populations like those in New Zealand. However, a routine evaluation would still require the development of an improved methodology. It would also be computationally very challenging because of the simultaneous presence of a large number of breeds.
Safroneeva, E; Vavricka, S; Fournier, N; Seibold, F; Mottet, C; Nydegger, A; Ezri, J; Straumann, A; Rogler, G; Schoepfer, A M
2015-09-01
Studies that systematically assess change in ulcerative colitis (UC) extent over time in adult patients are scarce. To assess changes in disease extent over time and to evaluate clinical parameters associated with this change. Data from the Swiss IBD cohort study were analysed. We used logistic regression modelling to identify factors associated with a change in disease extent. A total of 918 UC patients (45.3% females) were included. At diagnosis, UC patients presented with the following disease extent: proctitis [199 patients (21.7%)], left-sided colitis [338 patients (36.8%)] and extensive colitis/pancolitis [381 (41.5%)]. During a median disease duration of 9 [4-16] years, progression and regression was documented in 145 patients (15.8%) and 149 patients (16.2%) respectively. In addition, 624 patients (68.0%) had a stable disease extent. The following factors were identified to be associated with disease progression: treatment with systemic glucocorticoids [odds ratio (OR) 1.704, P = 0.025] and calcineurin inhibitors (OR: 2.716, P = 0.005). No specific factors were found to be associated with disease regression. Over a median disease duration of 9 [4-16] years, about two-thirds of UC patients maintained the initial disease extent; the remaining one-third had experienced either progression or regression of the disease extent. © 2015 John Wiley & Sons Ltd.
Water quality parameter measurement using spectral signatures
NASA Technical Reports Server (NTRS)
White, P. E.
1973-01-01
Regression analysis is applied to the problem of measuring water quality parameters from remote sensing spectral signature data. The equations necessary to perform regression analysis are presented and methods of testing the strength and reliability of a regression are described. An efficient algorithm for selecting an optimal subset of the independent variables available for a regression is also presented.
The purpose of this report is to provide a reference manual that could be used by investigators for making informed use of logistic regression using two methods (standard logistic regression and MARS). The details for analyses of relationships between a dependent binary response ...
NASA Astrophysics Data System (ADS)
Tang, Jie; Liu, Rong; Zhang, Yue-Li; Liu, Mou-Ze; Hu, Yong-Fang; Shao, Ming-Jie; Zhu, Li-Jun; Xin, Hua-Wen; Feng, Gui-Wen; Shang, Wen-Jun; Meng, Xiang-Guang; Zhang, Li-Rong; Ming, Ying-Zi; Zhang, Wei
2017-02-01
Tacrolimus has a narrow therapeutic window and considerable variability in clinical use. Our goal was to compare the performance of multiple linear regression (MLR) and eight machine learning techniques in pharmacogenetic algorithm-based prediction of tacrolimus stable dose (TSD) in a large Chinese cohort. A total of 1,045 renal transplant patients were recruited, 80% of which were randomly selected as the “derivation cohort” to develop dose-prediction algorithm, while the remaining 20% constituted the “validation cohort” to test the final selected algorithm. MLR, artificial neural network (ANN), regression tree (RT), multivariate adaptive regression splines (MARS), boosted regression tree (BRT), support vector regression (SVR), random forest regression (RFR), lasso regression (LAR) and Bayesian additive regression trees (BART) were applied and their performances were compared in this work. Among all the machine learning models, RT performed best in both derivation [0.71 (0.67-0.76)] and validation cohorts [0.73 (0.63-0.82)]. In addition, the ideal rate of RT was 4% higher than that of MLR. To our knowledge, this is the first study to use machine learning models to predict TSD, which will further facilitate personalized medicine in tacrolimus administration in the future.
2018-01-01
Objective The objective of this study was to estimate genetic parameters of milk, fat, and protein yields within and across lactations in Tunisian Holsteins using a random regression test-day (TD) model. Methods A random regression multiple trait multiple lactation TD model was used to estimate genetic parameters in the Tunisian dairy cattle population. Data were TD yields of milk, fat, and protein from the first three lactations. Random regressions were modeled with third-order Legendre polynomials for the additive genetic, and permanent environment effects. Heritabilities, and genetic correlations were estimated by Bayesian techniques using the Gibbs sampler. Results All variance components tended to be high in the beginning and the end of lactations. Additive genetic variances for milk, fat, and protein yields were the lowest and were the least variable compared to permanent variances. Heritability values tended to increase with parity. Estimates of heritabilities for 305-d yield-traits were low to moderate, 0.14 to 0.2, 0.12 to 0.17, and 0.13 to 0.18 for milk, fat, and protein yields, respectively. Within-parity, genetic correlations among traits were up to 0.74. Genetic correlations among lactations for the yield traits were relatively high and ranged from 0.78±0.01 to 0.82±0.03, between the first and second parities, from 0.73±0.03 to 0.8±0.04 between the first and third parities, and from 0.82±0.02 to 0.84±0.04 between the second and third parities. Conclusion These results are comparable to previously reported estimates on the same population, indicating that the adoption of a random regression TD model as the official genetic evaluation for production traits in Tunisia, as developed by most Interbull countries, is possible in the Tunisian Holsteins. PMID:28823122
Genetic analysis of groups of mid-infrared predicted fatty acids in milk.
Narayana, S G; Schenkel, F S; Fleming, A; Koeck, A; Malchiodi, F; Jamrozik, J; Johnston, J; Sargolzaei, M; Miglior, F
2017-06-01
The objective of this study was to investigate genetic variability of mid-infrared predicted fatty acid groups in Canadian Holstein cattle. Genetic parameters were estimated for 5 groups of fatty acids: short-chain (4 to 10 carbons), medium-chain (11 to 16 carbons), long-chain (17 to 22 carbons), saturated, and unsaturated fatty acids. The data set included 49,127 test-day records from 10,029 first-lactation Holstein cows in 810 herds. The random regression animal test-day model included days in milk, herd-test date, and age-season of calving (polynomial regression) as fixed effects, herd-year of calving, animal additive genetic effect, and permanent environment effects as random polynomial regressions, and random residual effect. Legendre polynomials of the third degree were selected for the fixed regression for age-season of calving effect and Legendre polynomials of the fourth degree were selected for the random regression for animal additive genetic, permanent environment, and herd-year effect. The average daily heritability over the lactation for the medium-chain fatty acid group (0.32) was higher than for the short-chain (0.24) and long-chain (0.23) fatty acid groups. The average daily heritability for the saturated fatty acid group (0.33) was greater than for the unsaturated fatty acid group (0.21). Estimated average daily genetic correlations were positive among all fatty acid groups and ranged from moderate to high (0.63-0.96). The genetic correlations illustrated similarities and differences in their origin and the makeup of the groupings based on chain length and saturation. These results provide evidence for the existence of genetic variation in mid-infrared predicted fatty acid groups, and the possibility of improving milk fatty acid profile through genetic selection in Canadian dairy cattle. Copyright © 2017 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.
Ben Zaabza, Hafedh; Ben Gara, Abderrahmen; Rekik, Boulbaba
2018-05-01
The objective of this study was to estimate genetic parameters of milk, fat, and protein yields within and across lactations in Tunisian Holsteins using a random regression test-day (TD) model. A random regression multiple trait multiple lactation TD model was used to estimate genetic parameters in the Tunisian dairy cattle population. Data were TD yields of milk, fat, and protein from the first three lactations. Random regressions were modeled with third-order Legendre polynomials for the additive genetic, and permanent environment effects. Heritabilities, and genetic correlations were estimated by Bayesian techniques using the Gibbs sampler. All variance components tended to be high in the beginning and the end of lactations. Additive genetic variances for milk, fat, and protein yields were the lowest and were the least variable compared to permanent variances. Heritability values tended to increase with parity. Estimates of heritabilities for 305-d yield-traits were low to moderate, 0.14 to 0.2, 0.12 to 0.17, and 0.13 to 0.18 for milk, fat, and protein yields, respectively. Within-parity, genetic correlations among traits were up to 0.74. Genetic correlations among lactations for the yield traits were relatively high and ranged from 0.78±0.01 to 0.82±0.03, between the first and second parities, from 0.73±0.03 to 0.8±0.04 between the first and third parities, and from 0.82±0.02 to 0.84±0.04 between the second and third parities. These results are comparable to previously reported estimates on the same population, indicating that the adoption of a random regression TD model as the official genetic evaluation for production traits in Tunisia, as developed by most Interbull countries, is possible in the Tunisian Holsteins.
Fan, Shou-Zen; Abbod, Maysam F.
2018-01-01
Estimating the depth of anaesthesia (DoA) in operations has always been a challenging issue due to the underlying complexity of the brain mechanisms. Electroencephalogram (EEG) signals are undoubtedly the most widely used signals for measuring DoA. In this paper, a novel EEG-based index is proposed to evaluate DoA for 24 patients receiving general anaesthesia with different levels of unconsciousness. Sample Entropy (SampEn) algorithm was utilised in order to acquire the chaotic features of the signals. After calculating the SampEn from the EEG signals, Random Forest was utilised for developing learning regression models with Bispectral index (BIS) as the target. Correlation coefficient, mean absolute error, and area under the curve (AUC) were used to verify the perioperative performance of the proposed method. Validation comparisons with typical nonstationary signal analysis methods (i.e., recurrence analysis and permutation entropy) and regression methods (i.e., neural network and support vector machine) were conducted. To further verify the accuracy and validity of the proposed methodology, the data is divided into four unconsciousness-level groups on the basis of BIS levels. Subsequently, analysis of variance (ANOVA) was applied to the corresponding index (i.e., regression output). Results indicate that the correlation coefficient improved to 0.72 ± 0.09 after filtering and to 0.90 ± 0.05 after regression from the initial values of 0.51 ± 0.17. Similarly, the final mean absolute error dramatically declined to 5.22 ± 2.12. In addition, the ultimate AUC increased to 0.98 ± 0.02, and the ANOVA analysis indicates that each of the four groups of different anaesthetic levels demonstrated significant difference from the nearest levels. Furthermore, the Random Forest output was extensively linear in relation to BIS, thus with better DoA prediction accuracy. In conclusion, the proposed method provides a concrete basis for monitoring patients’ anaesthetic level during surgeries. PMID:29844970
Krishan, Kewal; Kanchan, Tanuj; Sharma, Abhilasha
2012-05-01
Estimation of stature is an important parameter in identification of human remains in forensic examinations. The present study is aimed to compare the reliability and accuracy of stature estimation and to demonstrate the variability in estimated stature and actual stature using multiplication factor and regression analysis methods. The study is based on a sample of 246 subjects (123 males and 123 females) from North India aged between 17 and 20 years. Four anthropometric measurements; hand length, hand breadth, foot length and foot breadth taken on the left side in each subject were included in the study. Stature was measured using standard anthropometric techniques. Multiplication factors were calculated and linear regression models were derived for estimation of stature from hand and foot dimensions. Derived multiplication factors and regression formula were applied to the hand and foot measurements in the study sample. The estimated stature from the multiplication factors and regression analysis was compared with the actual stature to find the error in estimated stature. The results indicate that the range of error in estimation of stature from regression analysis method is less than that of multiplication factor method thus, confirming that the regression analysis method is better than multiplication factor analysis in stature estimation. Copyright © 2012 Elsevier Ltd and Faculty of Forensic and Legal Medicine. All rights reserved.
Carbonell, Felix; Bellec, Pierre
2011-01-01
Abstract The influence of the global average signal (GAS) on functional-magnetic resonance imaging (fMRI)–based resting-state functional connectivity is a matter of ongoing debate. The global average fluctuations increase the correlation between functional systems beyond the correlation that reflects their specific functional connectivity. Hence, removal of the GAS is a common practice for facilitating the observation of network-specific functional connectivity. This strategy relies on the implicit assumption of a linear-additive model according to which global fluctuations, irrespective of their origin, and network-specific fluctuations are super-positioned. However, removal of the GAS introduces spurious negative correlations between functional systems, bringing into question the validity of previous findings of negative correlations between fluctuations in the default-mode and the task-positive networks. Here we present an alternative method for estimating global fluctuations, immune to the complications associated with the GAS. Principal components analysis was applied to resting-state fMRI time-series. A global-signal effect estimator was defined as the principal component (PC) that correlated best with the GAS. The mean correlation coefficient between our proposed PC-based global effect estimator and the GAS was 0.97±0.05, demonstrating that our estimator successfully approximated the GAS. In 66 out of 68 runs, the PC that showed the highest correlation with the GAS was the first PC. Since PCs are orthogonal, our method provides an estimator of the global fluctuations, which is uncorrelated to the remaining, network-specific fluctuations. Moreover, unlike the regression of the GAS, the regression of the PC-based global effect estimator does not introduce spurious anti-correlations beyond the decrease in seed-based correlation values allowed by the assumed additive model. After regressing this PC-based estimator out of the original time-series, we observed robust anti-correlations between resting-state fluctuations in the default-mode and the task-positive networks. We conclude that resting-state global fluctuations and network-specific fluctuations are uncorrelated, supporting a Resting-State Linear-Additive Model. In addition, we conclude that the network-specific resting-state fluctuations of the default-mode and task-positive networks show artifact-free anti-correlations. PMID:22444074
Handling nonnormality and variance heterogeneity for quantitative sublethal toxicity tests.
Ritz, Christian; Van der Vliet, Leana
2009-09-01
The advantages of using regression-based techniques to derive endpoints from environmental toxicity data are clear, and slowly, this superior analytical technique is gaining acceptance. As use of regression-based analysis becomes more widespread, some of the associated nuances and potential problems come into sharper focus. Looking at data sets that cover a broad spectrum of standard test species, we noticed that some model fits to data failed to meet two key assumptions-variance homogeneity and normality-that are necessary for correct statistical analysis via regression-based techniques. Failure to meet these assumptions often is caused by reduced variance at the concentrations showing severe adverse effects. Although commonly used with linear regression analysis, transformation of the response variable only is not appropriate when fitting data using nonlinear regression techniques. Through analysis of sample data sets, including Lemna minor, Eisenia andrei (terrestrial earthworm), and algae, we show that both the so-called Box-Cox transformation and use of the Poisson distribution can help to correct variance heterogeneity and nonnormality and so allow nonlinear regression analysis to be implemented. Both the Box-Cox transformation and the Poisson distribution can be readily implemented into existing protocols for statistical analysis. By correcting for nonnormality and variance heterogeneity, these two statistical tools can be used to encourage the transition to regression-based analysis and the depreciation of less-desirable and less-flexible analytical techniques, such as linear interpolation.
Informal Peer-Assisted Learning Groups Did Not Lead to Better Performance of Saudi Dental Students.
AbdelSalam, Maha; El Tantawi, Maha; Al-Ansari, Asim; AlAgl, Adel; Al-Harbi, Fahad
2017-01-01
To describe peer-assisted learning (PAL) groups formed by dental undergraduate students in a biomedical course and to investigate the association of individual and group characteristics with academic performance. In 2015, 92 fourth-year students (43 males and 49 females) in the College of Dentistry, University of Dammam, Saudi Arabia, were invited to form PAL groups to study a unit of a biomedical course. An examination was used to assess their knowledge after 2 weeks. In addition, a questionnaire and social network analysis were used to investigate (1) individual student attributes: gender, role, subject matter knowledge, grade in previous year, teaming with friends, previous communication with teammates, and content discussion, and (2) group attributes: group teacher's previous grade, number of colleagues with whom a student connected, teaming with friends, similarity of teammates' previous grades, and teacher having higher previous grades than other teammates. Regression analysis was used to assess the association of examination scores with individual and group attributes. The response rate was 80.4% (74 students: 36 males and 38 females). Students who previously scored grades A and B had higher examination scores than students with grades C/less (regression coefficient = 18.50 and 13.39) within the groups. Higher scores were not associated with working in groups including friends only (regression coefficient = 1.17) or when all students had similar previous grades (regression coefficient = 0.85). Students with previous high grades benefited to a greater extent from working in PAL groups. Similarity of teammates in PAL groups was not associated with better scores. © 2017 S. Karger AG, Basel.
Adolescent religiosity and attitudes to HIV and AIDS in Ghana.
Amoako-Agyeman, Kofi Nyame
2012-11-01
This study investigated the relationships between adolescent religiosity and attitudes to HIV/AIDS based on two major techniques of analysis, factor and regression analysis towards informing preventive school education strategies. Using cross-sectional data of 448 adolescents in junior high school, the study incorporated survey in a self-administered questionnaire and sought to identify underlying factors that affect pupils' responses, delineate the pattern of relationships between variables and select models which best explain and predict relationships among variables. A seven-factor solution described the 'attitude' construct including abstinence and protection, and six for 'religiosity'. The results showed relatively high levels of religiosity and a preference for private religiosity as opposed to organisational religiosity. The regression analysis produced significant relationships between factors of attitudes to HIV/AIDS and of religiosity. Adolescent with very high private religiosity are more likely to abstain from sex but less likely to use condoms once they initiate: protection is inversely related to religiosity. The findings suggest that religious-based adolescent interventions should focus on intrinsic religiosity. Additionally, increasing HIV prevention information and incorporating culturally relevant and socially acceptable values might lend support to improved adolescent school-based HIV/AIDS prevention programmes.
Wolf, Alexander; Leucht, Stefan; Pajonk, Frank-Gerald
2017-04-01
Behavioural and psychological symptoms in dementia (BPSD) are common and often treated with antipsychotics, which are known to have small efficacy and to cause many side effects. One potential side effect might be cognitive decline. We searched MEDLINE, Scopus, CENTRAL and www.ClincalStudyResult.org for randomized, double-blind, placebo-controlled trials using antipsychotics for treating BPSD and evaluated cognitive functioning. The studies identified were summarized in a meta-analysis with the standardized mean difference (SMD, Hedges's g) as the effect size. Meta-regression was additionally performed to identify associated factors. Ten studies provided data on the course of cognitive functioning. The random effects model of the pooled analysis showed a not significant effect (SMD = -0.065, 95 % CI -0.186 to 0.057, I 2 = 41 %). Meta-regression revealed a significant correlation between cognitive impairment and treatment duration (R 2 = 0.78, p < 0.02) as well as baseline MMSE (R 2 = 0.92, p < 0.005). These correlations depend on only two out of ten studies and should interpret cautiously.
Tani, Yuji; Ogasawara, Katsuhiko
2012-01-01
This study aimed to contribute to the management of a healthcare organization by providing management information using time-series analysis of business data accumulated in the hospital information system, which has not been utilized thus far. In this study, we examined the performance of the prediction method using the auto-regressive integrated moving-average (ARIMA) model, using the business data obtained at the Radiology Department. We made the model using the data used for analysis, which was the number of radiological examinations in the past 9 years, and we predicted the number of radiological examinations in the last 1 year. Then, we compared the actual value with the forecast value. We were able to establish that the performance prediction method was simple and cost-effective by using free software. In addition, we were able to build the simple model by pre-processing the removal of trend components using the data. The difference between predicted values and actual values was 10%; however, it was more important to understand the chronological change rather than the individual time-series values. Furthermore, our method was highly versatile and adaptable compared to the general time-series data. Therefore, different healthcare organizations can use our method for the analysis and forecasting of their business data.
Valeri, A; Briollais, L; Azzouzi, R; Fournier, G; Mangin, P; Berthon, P; Cussenot, O; Demenais, F
2003-03-01
Four segregation analyses concerning prostate cancer (CaP), three conducted in the United States and one in Northern Europe, have shown evidence for a dominant major gene but with different parameter estimates. A recent segregation analysis of Australian pedigrees has found a better fit of a two-locus model than single-locus models. This model included a dominantly inherited increased risk that was greater at younger ages and a recessively inherited or X-linked increased risk that was greater at older ages. Recent linkage analyses have led to the detection of at least 8 CaP predisposing genes, suggesting a complex inheritance and genetic heterogeneity. To assess the nature of familial aggregation of prostate cancer in France, segregation analysis was conducted in 691 families ascertained through 691 CaP patients, recruited from three French hospitals and unselected with respect to age at diagnosis, clinical stage or family history. This mode of family inclusion, without any particular selection of the probands, is unique, as probands from all previous analyses were selected according to various criteria. Segregation analysis was carried out using the logistic hazard regressive model, as incorporated in the REGRESS program, which can accommodate a major gene effect, residual familial dependences of any origin (genetic and/or environmental), and covariates, while including survival analysis concepts. Segregation analysis showed evidence for the segregation of an autosomal dominant gene (allele frequency of 0.03%) with an additional brother-brother dependence. The estimated cumulative risks of prostate cancer by age 85 years, among subjects with the at-risk genotype, were 86% in the fathers' generation and 99% in the probands' generation. This study supports the model of Mendelian transmission of a rare autosomal dominant gene with high penetrance, and demonstrates that additional genetic and/or common sibling environmental factors are involved to account for the familial clustering of CaP.
Using Robust Standard Errors to Combine Multiple Regression Estimates with Meta-Analysis
ERIC Educational Resources Information Center
Williams, Ryan T.
2012-01-01
Combining multiple regression estimates with meta-analysis has continued to be a difficult task. A variety of methods have been proposed and used to combine multiple regression slope estimates with meta-analysis, however, most of these methods have serious methodological and practical limitations. The purpose of this study was to explore the use…
A Quality Assessment Tool for Non-Specialist Users of Regression Analysis
ERIC Educational Resources Information Center
Argyrous, George
2015-01-01
This paper illustrates the use of a quality assessment tool for regression analysis. It is designed for non-specialist "consumers" of evidence, such as policy makers. The tool provides a series of questions such consumers of evidence can ask to interrogate regression analysis, and is illustrated with reference to a recent study published…
Morphology parameters for intracranial aneurysm rupture risk assessment.
Dhar, Sujan; Tremmel, Markus; Mocco, J; Kim, Minsuok; Yamamoto, Junichi; Siddiqui, Adnan H; Hopkins, L Nelson; Meng, Hui
2008-08-01
The aim of this study is to identify image-based morphological parameters that correlate with human intracranial aneurysm (IA) rupture. For 45 patients with terminal or sidewall saccular IAs (25 unruptured, 20 ruptured), three-dimensional geometries were evaluated for a range of morphological parameters. In addition to five previously studied parameters (aspect ratio, aneurysm size, ellipticity index, nonsphericity index, and undulation index), we defined three novel parameters incorporating the parent vessel geometry (vessel angle, aneurysm [inclination] angle, and [aneurysm-to-vessel] size ratio) and explored their correlation with aneurysm rupture. Parameters were analyzed with a two-tailed independent Student's t test for significance; significant parameters (P < 0.05) were further examined by multivariate logistic regression analysis. Additionally, receiver operating characteristic analyses were performed on each parameter. Statistically significant differences were found between mean values in ruptured and unruptured groups for size ratio, undulation index, nonsphericity index, ellipticity index, aneurysm angle, and aspect ratio. Logistic regression analysis further revealed that size ratio (odds ratio, 1.41; 95% confidence interval, 1.03-1.92) and undulation index (odds ratio, 1.51; 95% confidence interval, 1.08-2.11) had the strongest independent correlation with ruptured IA. From the receiver operating characteristic analysis, size ratio and aneurysm angle had the highest area under the curve values of 0.83 and 0.85, respectively. Size ratio and aneurysm angle are promising new morphological metrics for IA rupture risk assessment. Because these parameters account for vessel geometry, they may bridge the gap between morphological studies and more qualitative location-based studies.
Construction and analysis of a modular model of caspase activation in apoptosis
Harrington, Heather A; Ho, Kenneth L; Ghosh, Samik; Tung, KC
2008-01-01
Background A key physiological mechanism employed by multicellular organisms is apoptosis, or programmed cell death. Apoptosis is triggered by the activation of caspases in response to both extracellular (extrinsic) and intracellular (intrinsic) signals. The extrinsic and intrinsic pathways are characterized by the formation of the death-inducing signaling complex (DISC) and the apoptosome, respectively; both the DISC and the apoptosome are oligomers with complex formation dynamics. Additionally, the extrinsic and intrinsic pathways are coupled through the mitochondrial apoptosis-induced channel via the Bcl-2 family of proteins. Results A model of caspase activation is constructed and analyzed. The apoptosis signaling network is simplified through modularization methodologies and equilibrium abstractions for three functional modules. The mathematical model is composed of a system of ordinary differential equations which is numerically solved. Multiple linear regression analysis investigates the role of each module and reduced models are constructed to identify key contributions of the extrinsic and intrinsic pathways in triggering apoptosis for different cell lines. Conclusion Through linear regression techniques, we identified the feedbacks, dissociation of complexes, and negative regulators as the key components in apoptosis. The analysis and reduced models for our model formulation reveal that the chosen cell lines predominately exhibit strong extrinsic caspase, typical of type I cell, behavior. Furthermore, under the simplified model framework, the selected cells lines exhibit different modes by which caspase activation may occur. Finally the proposed modularized model of apoptosis may generalize behavior for additional cells and tissues, specifically identifying and predicting components responsible for the transition from type I to type II cell behavior. PMID:19077196
The Prognostic Value of Epithelial Membrane Protein 1 (EMP-1) in Patients with Laryngeal Carcinoma
Liu, Chang; Wei, Xiaojun; Li, Feng; Wang, Li; Ruan, Xinjian; Jia, Jia; Zhang, Xia
2017-01-01
Background In the present study, we aimed to investigate the prognostic value of epithelial membrane protein 1 (EMP-1) gene in patients diagnosed with laryngeal carcinoma (LC). Material/Methods Patients who were pathologically diagnosed with LC were enrolled in the present study. The expression levels of EMP-1 in tumor tissues and corresponding normal tissues collected from the LC patients were detected by semi-reverse transcriptase polymerase chain reaction (semi-RT-PCR). Chi-square analysis was used to evaluate the relationship between EMP-1 expression level and clinical characteristics. Survival analysis for the study population was analyzed by Kaplan-Meier method with log rank test. Additionally, Cox regression model was applied to evaluate the prognostic value of EMP-1 in LC patients. Results 106 LC patients, including 55 men and 51 women, were enrolled in the present study. Semi-RT-PCR demonstrated that the expression level of EMP-1 was decreased in tumor tissues, compared with adjacent normal tissues (p<0.001). Moreover, the level was significantly associated with lymph node metastasis, histological grade, and clinical stage (p<0.05 for all). In addition, low levels of EMP-1 was significantly correlated with poor survival rate (log rank test, p=0.020). Cox regression analysis indicated that EMP-1 was an independent marker for LC prognosis (HR=2.755, 95% CI=1.123–6.760, p=0.027). Conclusions The abnormal expression of EMP-1 may be associated with progression of LC and the gene may act as a prognostic marker for LC. PMID:28779068
The Prognostic Value of Epithelial Membrane Protein 1 (EMP-1) in Patients with Laryngeal Carcinoma.
Liu, Chang; Wei, Xiaojun; Li, Feng; Wang, Li; Ruan, Xinjian; Jia, Jia; Zhang, Xia
2017-08-05
BACKGROUND In the present study, we aimed to investigate the prognostic value of epithelial membrane protein 1 (EMP-1) gene in patients diagnosed with laryngeal carcinoma (LC). MATERIAL AND METHODS Patients who were pathologically diagnosed with LC were enrolled in the present study. The expression levels of EMP-1 in tumor tissues and corresponding normal tissues collected from the LC patients were detected by semi-reverse transcriptase polymerase chain reaction (semi-RT-PCR). Chi-square analysis was used to evaluate the relationship between EMP-1 expression level and clinical characteristics. Survival analysis for the study population was analyzed by Kaplan-Meier method with log rank test. Additionally, Cox regression model was applied to evaluate the prognostic value of EMP-1 in LC patients. RESULTS 106 LC patients, including 55 men and 51 women, were enrolled in the present study. Semi-RT-PCR demonstrated that the expression level of EMP-1 was decreased in tumor tissues, compared with adjacent normal tissues (p<0.001). Moreover, the level was significantly associated with lymph node metastasis, histological grade, and clinical stage (p<0.05 for all). In addition, low levels of EMP-1 was significantly correlated with poor survival rate (log rank test, p=0.020). Cox regression analysis indicated that EMP-1 was an independent marker for LC prognosis (HR=2.755, 95% CI=1.123-6.760, p=0.027). CONCLUSIONS The abnormal expression of EMP-1 may be associated with progression of LC and the gene may act as a prognostic marker for LC.
Bispectral analysis during deep sedation of pediatric oral surgery patients.
Overly, Frank L; Wright, Robert O; Connor, Francis A; Jay, Gregory D; Linakis, James G
2005-02-01
Bispectral (BIS) analysis uses electroencephalogram information from a forehead electrode to calculate an index score (0 to 100; 0 = coma; 90 to 100 = awake). This index score correlates with the level of alertness in anesthetized patients. Classically, sedation has been monitored with clinical sedation scales such as the Observers Assessment of Alertness Sedation Scale (OAA/S), Modified Ramsey Scale, or a Visual Analog Scale (VAS). Our objective was to determine the correlation between clinical sedation scales and BIS index in pediatric patients undergoing sedation in an outpatient oral surgery setting. Prospective cohort study of patients aged 2 to 17 years undergoing sedation in an outpatient oral surgery office. Sedation was performed in the customary manner with the addition of BIS monitoring. Three clinical sedation scores (OAA/S: 5 to 1; 5 = awake, 1 = unresponsive; Modified Ramsey: 1 to 6; 1-2 = awake, 6 = unresponsive; VAS: 0 to 10; 0 = awake, 10 = unresponsive) were assigned every 5 minutes by an investigator blinded to the BIS index. Data were analyzed using a repeated measures linear regression model. Sixteen subjects undergoing oral surgery, ages 4.5 years to 17 years, were enrolled, mean age 12.6 years +/- 4.3 years (standard deviation). Patients received methohexital in addition to 1 or more of the following: nitrous oxide, fentanyl, or midazolam. The results of the longitudinal regression analysis showed a highly significant association between the sedation scales and the BIS index. The BIS monitor may be a useful adjunct in monitoring pediatric patients receiving sedation in the outpatient setting.
Park, Ji Hyun; Kim, Hyeon-Young; Lee, Hanna; Yun, Eun Kyoung
2015-12-01
This study compares the performance of the logistic regression and decision tree analysis methods for assessing the risk factors for infection in cancer patients undergoing chemotherapy. The subjects were 732 cancer patients who were receiving chemotherapy at K university hospital in Seoul, Korea. The data were collected between March 2011 and February 2013 and were processed for descriptive analysis, logistic regression and decision tree analysis using the IBM SPSS Statistics 19 and Modeler 15.1 programs. The most common risk factors for infection in cancer patients receiving chemotherapy were identified as alkylating agents, vinca alkaloid and underlying diabetes mellitus. The logistic regression explained 66.7% of the variation in the data in terms of sensitivity and 88.9% in terms of specificity. The decision tree analysis accounted for 55.0% of the variation in the data in terms of sensitivity and 89.0% in terms of specificity. As for the overall classification accuracy, the logistic regression explained 88.0% and the decision tree analysis explained 87.2%. The logistic regression analysis showed a higher degree of sensitivity and classification accuracy. Therefore, logistic regression analysis is concluded to be the more effective and useful method for establishing an infection prediction model for patients undergoing chemotherapy. Copyright © 2015 Elsevier Ltd. All rights reserved.
Zarb, Francis; McEntee, Mark F; Rainford, Louise
2015-06-01
To evaluate visual grading characteristics (VGC) and ordinal regression analysis during head CT optimisation as a potential alternative to visual grading assessment (VGA), traditionally employed to score anatomical visualisation. Patient images (n = 66) were obtained using current and optimised imaging protocols from two CT suites: a 16-slice scanner at the national Maltese centre for trauma and a 64-slice scanner in a private centre. Local resident radiologists (n = 6) performed VGA followed by VGC and ordinal regression analysis. VGC alone indicated that optimised protocols had similar image quality as current protocols. Ordinal logistic regression analysis provided an in-depth evaluation, criterion by criterion allowing the selective implementation of the protocols. The local radiology review panel supported the implementation of optimised protocols for brain CT examinations (including trauma) in one centre, achieving radiation dose reductions ranging from 24 % to 36 %. In the second centre a 29 % reduction in radiation dose was achieved for follow-up cases. The combined use of VGC and ordinal logistic regression analysis led to clinical decisions being taken on the implementation of the optimised protocols. This improved method of image quality analysis provided the evidence to support imaging protocol optimisation, resulting in significant radiation dose savings. • There is need for scientifically based image quality evaluation during CT optimisation. • VGC and ordinal regression analysis in combination led to better informed clinical decisions. • VGC and ordinal regression analysis led to dose reductions without compromising diagnostic efficacy.
NASA Astrophysics Data System (ADS)
Esquinca, Alberto
This is a study of language use in the context of an inquiry-based science curriculum in which conceptual understanding ratings are used split texts into groups of "successful" and "unsuccessful" texts. "Successful" texts could include known features of science language. 420 texts generated by students in 14 classrooms from three school districts, culled from a prior study on the effectiveness of science notebooks to assess understanding, in addition to the aforementioned ratings are the data sources. In science notebooks, students write in the process of learning (here, a unit on electricity). The analytical framework is systemic functional linguistics (Halliday and Matthiessen, 2004; Eggins, 2004), specifically the concepts of genre, register and nominalization. Genre classification involves an analysis of the purpose and register features in the text (Schleppegrell, 2004). The use of features of the scientific academic register, namely the use relational processes and nominalization (Halliday and Martin, 1993), requires transitivity analysis and noun analysis. Transitivity analysis, consisting of the identification of the process type, is conducted on 4737 ranking clauses. A manual count of each noun used in the corpus allows for a typology of nouns. Four school science genres, procedures, procedural recounts reports and explanations, are found. Most texts (85.4%) are factual, and 14.1% are classified as explanations, the analytical genre. Logistic regression analysis indicates that there is no significant probability that the texts classified as explanation are placed in the group of "successful" texts. In addition, material process clauses predominate in the corpus, followed by relational process clauses. Results of a logistic regression analysis indicate that there is a significant probability (Chi square = 15.23, p < .0001) that texts with a high rate of relational processes are placed in the group of "successful" texts. In addition, 59.5% of 6511 nouns are references to physical materials, followed by references to abstract concepts (35.54%). Only two of the concept nouns were found to be nominalized referents in definition model sentences. In sum, the corpus has recognizable genres and features science language, and relational processes are more prevalent in "successful" texts. However, the pervasive feature of science language, nominalization, is scarce.
Park, In Ja; Kim, Dae Yong; Kim, Hee Cheol; Kim, Nam Kyu; Kim, Hyeong-Rok; Kang, Sung-Bum; Choi, Gyu-Seog; Lee, Kang Young; Kim, Seon-Hahn; Oh, Seung Taek; Lim, Seok-Byung; Kim, Jin Cheon; Oh, Jae Hwan; Kim, Sun Young; Lee, Woo Yong; Lee, Jung Bok; Yu, Chang Sik
2015-07-01
To explore the role of adjuvant chemotherapy for patients with ypT0-2N0 rectal cancer treated by preoperative chemoradiation therapy (PCRT) and radical resection. A national consortium of 10 institutions was formed, and patients with ypT0-2N0 mid- and low-rectal cancer after PCRT and radical resection from 2004 to 2009 were included. Patients were categorized into 2 groups according to receipt of additional adjuvant chemotherapy: Adj CTx (+) versus Adj CTx (-). Propensity scores were calculated and used to perform matched and adjusted analyses comparing relapse-free survival (RFS) between treatment groups while controlling for potential confounding. A total of 1016 patients, who met the selection criteria, were evaluated. Of these, 106 (10.4%) did not receive adjuvant chemotherapy. There was no overall improvement in 5-year RFS as a result of adjuvant chemotherapy [91.6% for Adj CTx (+) vs 87.5% for Adj CTx (-), P=.18]. There were no differences in 5-year local recurrence and distant metastasis rate between the 2 groups. In patients who show moderate, minimal, or no regression in tumor regression grade, however, possible association of adjuvant chemotherapy with RFS would be considered (hazard ratio 0.35; 95% confidence interval 0.14-0.88; P=.03). Cox regression analysis after propensity score matching failed to show that addition of adjuvant chemotherapy was associated with improved RFS (hazard ratio 0.81; 95% confidence interval 0.39-1.70; P=.58). Adjuvant chemotherapy seemed to not influence the RFS of patients with ypT0-2N0 rectal cancer after PCRT followed by radical resection. Thus, the addition of adjuvant chemotherapy needs to be weighed against its oncologic benefits. Copyright © 2015 Elsevier Inc. All rights reserved.
REGRESSION ANALYSIS OF SEA-SURFACE-TEMPERATURE PATTERNS FOR THE NORTH PACIFIC OCEAN.
SEA WATER, *SURFACE TEMPERATURE, *OCEANOGRAPHIC DATA, PACIFIC OCEAN, REGRESSION ANALYSIS , STATISTICAL ANALYSIS, UNDERWATER EQUIPMENT, DETECTION, UNDERWATER COMMUNICATIONS, DISTRIBUTION, THERMAL PROPERTIES, COMPUTERS.
USE OF LETHALITY DATA DURING CATEGORICAL REGRESSION MODELING OF ACUTE REFERENCE EXPOSURES
Categorical regression is being considered by the U.S. EPA as an additional tool for derivation of acute reference exposures (AREs) to be used for human health risk assessment for exposure to inhaled chemicals. Categorical regression is used to calculate probability-response fun...
The process and utility of classification and regression tree methodology in nursing research
Kuhn, Lisa; Page, Karen; Ward, John; Worrall-Carter, Linda
2014-01-01
Aim This paper presents a discussion of classification and regression tree analysis and its utility in nursing research. Background Classification and regression tree analysis is an exploratory research method used to illustrate associations between variables not suited to traditional regression analysis. Complex interactions are demonstrated between covariates and variables of interest in inverted tree diagrams. Design Discussion paper. Data sources English language literature was sourced from eBooks, Medline Complete and CINAHL Plus databases, Google and Google Scholar, hard copy research texts and retrieved reference lists for terms including classification and regression tree* and derivatives and recursive partitioning from 1984–2013. Discussion Classification and regression tree analysis is an important method used to identify previously unknown patterns amongst data. Whilst there are several reasons to embrace this method as a means of exploratory quantitative research, issues regarding quality of data as well as the usefulness and validity of the findings should be considered. Implications for Nursing Research Classification and regression tree analysis is a valuable tool to guide nurses to reduce gaps in the application of evidence to practice. With the ever-expanding availability of data, it is important that nurses understand the utility and limitations of the research method. Conclusion Classification and regression tree analysis is an easily interpreted method for modelling interactions between health-related variables that would otherwise remain obscured. Knowledge is presented graphically, providing insightful understanding of complex and hierarchical relationships in an accessible and useful way to nursing and other health professions. PMID:24237048
The process and utility of classification and regression tree methodology in nursing research.
Kuhn, Lisa; Page, Karen; Ward, John; Worrall-Carter, Linda
2014-06-01
This paper presents a discussion of classification and regression tree analysis and its utility in nursing research. Classification and regression tree analysis is an exploratory research method used to illustrate associations between variables not suited to traditional regression analysis. Complex interactions are demonstrated between covariates and variables of interest in inverted tree diagrams. Discussion paper. English language literature was sourced from eBooks, Medline Complete and CINAHL Plus databases, Google and Google Scholar, hard copy research texts and retrieved reference lists for terms including classification and regression tree* and derivatives and recursive partitioning from 1984-2013. Classification and regression tree analysis is an important method used to identify previously unknown patterns amongst data. Whilst there are several reasons to embrace this method as a means of exploratory quantitative research, issues regarding quality of data as well as the usefulness and validity of the findings should be considered. Classification and regression tree analysis is a valuable tool to guide nurses to reduce gaps in the application of evidence to practice. With the ever-expanding availability of data, it is important that nurses understand the utility and limitations of the research method. Classification and regression tree analysis is an easily interpreted method for modelling interactions between health-related variables that would otherwise remain obscured. Knowledge is presented graphically, providing insightful understanding of complex and hierarchical relationships in an accessible and useful way to nursing and other health professions. © 2013 The Authors. Journal of Advanced Nursing Published by John Wiley & Sons Ltd.
Hoch, Jeffrey S; Dewa, Carolyn S
2014-04-01
Economic evaluations commonly accompany trials of new treatments or interventions; however, regression methods and their corresponding advantages for the analysis of cost-effectiveness data are not well known. To illustrate regression-based economic evaluation, we present a case study investigating the cost-effectiveness of a collaborative mental health care program for people receiving short-term disability benefits for psychiatric disorders. We implement net benefit regression to illustrate its strengths and limitations. Net benefit regression offers a simple option for cost-effectiveness analyses of person-level data. By placing economic evaluation in a regression framework, regression-based techniques can facilitate the analysis and provide simple solutions to commonly encountered challenges. Economic evaluations of person-level data (eg, from a clinical trial) should use net benefit regression to facilitate analysis and enhance results.
Gotvald, Anthony J.; Barth, Nancy A.; Veilleux, Andrea G.; Parrett, Charles
2012-01-01
Methods for estimating the magnitude and frequency of floods in California that are not substantially affected by regulation or diversions have been updated. Annual peak-flow data through water year 2006 were analyzed for 771 streamflow-gaging stations (streamgages) in California having 10 or more years of data. Flood-frequency estimates were computed for the streamgages by using the expected moments algorithm to fit a Pearson Type III distribution to logarithms of annual peak flows for each streamgage. Low-outlier and historic information were incorporated into the flood-frequency analysis, and a generalized Grubbs-Beck test was used to detect multiple potentially influential low outliers. Special methods for fitting the distribution were developed for streamgages in the desert region in southeastern California. Additionally, basin characteristics for the streamgages were computed by using a geographical information system. Regional regression analysis, using generalized least squares regression, was used to develop a set of equations for estimating flows with 50-, 20-, 10-, 4-, 2-, 1-, 0.5-, and 0.2-percent annual exceedance probabilities for ungaged basins in California that are outside of the southeastern desert region. Flood-frequency estimates and basin characteristics for 630 streamgages were combined to form the final database used in the regional regression analysis. Five hydrologic regions were developed for the area of California outside of the desert region. The final regional regression equations are functions of drainage area and mean annual precipitation for four of the five regions. In one region, the Sierra Nevada region, the final equations are functions of drainage area, mean basin elevation, and mean annual precipitation. Average standard errors of prediction for the regression equations in all five regions range from 42.7 to 161.9 percent. For the desert region of California, an analysis of 33 streamgages was used to develop regional estimates of all three parameters (mean, standard deviation, and skew) of the log-Pearson Type III distribution. The regional estimates were then used to develop a set of equations for estimating flows with 50-, 20-, 10-, 4-, 2-, 1-, 0.5-, and 0.2-percent annual exceedance probabilities for ungaged basins. The final regional regression equations are functions of drainage area. Average standard errors of prediction for these regression equations range from 214.2 to 856.2 percent. Annual peak-flow data through water year 2006 were analyzed for eight streamgages in California having 10 or more years of data considered to be affected by urbanization. Flood-frequency estimates were computed for the urban streamgages by fitting a Pearson Type III distribution to logarithms of annual peak flows for each streamgage. Regression analysis could not be used to develop flood-frequency estimation equations for urban streams because of the limited number of sites. Flood-frequency estimates for the eight urban sites were graphically compared to flood-frequency estimates for 630 non-urban sites. The regression equations developed from this study will be incorporated into the U.S. Geological Survey (USGS) StreamStats program. The StreamStats program is a Web-based application that provides streamflow statistics and basin characteristics for USGS streamgages and ungaged sites of interest. StreamStats can also compute basin characteristics and provide estimates of streamflow statistics for ungaged sites when users select the location of a site along any stream in California.
Bowden, Jack; Del Greco M, Fabiola; Minelli, Cosetta; Davey Smith, George; Sheehan, Nuala A; Thompson, John R
2016-12-01
: MR-Egger regression has recently been proposed as a method for Mendelian randomization (MR) analyses incorporating summary data estimates of causal effect from multiple individual variants, which is robust to invalid instruments. It can be used to test for directional pleiotropy and provides an estimate of the causal effect adjusted for its presence. MR-Egger regression provides a useful additional sensitivity analysis to the standard inverse variance weighted (IVW) approach that assumes all variants are valid instruments. Both methods use weights that consider the single nucleotide polymorphism (SNP)-exposure associations to be known, rather than estimated. We call this the `NO Measurement Error' (NOME) assumption. Causal effect estimates from the IVW approach exhibit weak instrument bias whenever the genetic variants utilized violate the NOME assumption, which can be reliably measured using the F-statistic. The effect of NOME violation on MR-Egger regression has yet to be studied. An adaptation of the I2 statistic from the field of meta-analysis is proposed to quantify the strength of NOME violation for MR-Egger. It lies between 0 and 1, and indicates the expected relative bias (or dilution) of the MR-Egger causal estimate in the two-sample MR context. We call it IGX2 . The method of simulation extrapolation is also explored to counteract the dilution. Their joint utility is evaluated using simulated data and applied to a real MR example. In simulated two-sample MR analyses we show that, when a causal effect exists, the MR-Egger estimate of causal effect is biased towards the null when NOME is violated, and the stronger the violation (as indicated by lower values of IGX2 ), the stronger the dilution. When additionally all genetic variants are valid instruments, the type I error rate of the MR-Egger test for pleiotropy is inflated and the causal effect underestimated. Simulation extrapolation is shown to substantially mitigate these adverse effects. We demonstrate our proposed approach for a two-sample summary data MR analysis to estimate the causal effect of low-density lipoprotein on heart disease risk. A high value of IGX2 close to 1 indicates that dilution does not materially affect the standard MR-Egger analyses for these data. : Care must be taken to assess the NOME assumption via the IGX2 statistic before implementing standard MR-Egger regression in the two-sample summary data context. If IGX2 is sufficiently low (less than 90%), inferences from the method should be interpreted with caution and adjustment methods considered. © The Author 2016. Published by Oxford University Press on behalf of the International Epidemiological Association.
CADDIS Volume 4. Data Analysis: Basic Analyses
Use of statistical tests to determine if an observation is outside the normal range of expected values. Details of CART, regression analysis, use of quantile regression analysis, CART in causal analysis, simplifying or pruning resulting trees.
GIS and statistical analysis for landslide susceptibility mapping in the Daunia area, Italy
NASA Astrophysics Data System (ADS)
Mancini, F.; Ceppi, C.; Ritrovato, G.
2010-09-01
This study focuses on landslide susceptibility mapping in the Daunia area (Apulian Apennines, Italy) and achieves this by using a multivariate statistical method and data processing in a Geographical Information System (GIS). The Logistic Regression (hereafter LR) method was chosen to produce a susceptibility map over an area of 130 000 ha where small settlements are historically threatened by landslide phenomena. By means of LR analysis, the tendency to landslide occurrences was, therefore, assessed by relating a landslide inventory (dependent variable) to a series of causal factors (independent variables) which were managed in the GIS, while the statistical analyses were performed by means of the SPSS (Statistical Package for the Social Sciences) software. The LR analysis produced a reliable susceptibility map of the investigated area and the probability level of landslide occurrence was ranked in four classes. The overall performance achieved by the LR analysis was assessed by local comparison between the expected susceptibility and an independent dataset extrapolated from the landslide inventory. Of the samples classified as susceptible to landslide occurrences, 85% correspond to areas where landslide phenomena have actually occurred. In addition, the consideration of the regression coefficients provided by the analysis demonstrated that a major role is played by the "land cover" and "lithology" causal factors in determining the occurrence and distribution of landslide phenomena in the Apulian Apennines.
Yamazaki, Takeshi; Takeda, Hisato; Hagiya, Koichi; Yamaguchi, Satoshi; Sasaki, Osamu
2018-03-13
Because lactation periods in dairy cows lengthen with increasing total milk production, it is important to predict individual productivities after 305 days in milk (DIM) to determine the optimal lactation period. We therefore examined whether the random regression (RR) coefficient from 306 to 450 DIM (M2) can be predicted from those during the first 305 DIM (M1) by using a random regression model. We analyzed test-day milk records from 85690 Holstein cows in their first lactations and 131727 cows in their later (second to fifth) lactations. Data in M1 and M2 were analyzed separately by using different single-trait RR animal models. We then performed a multiple regression analysis of the RR coefficients of M2 on those of M1 during the first and later lactations. The first-order Legendre polynomials were practical covariates of random regression for the milk yields of M2. All RR coefficients for the additive genetic (AG) effect and the intercept for the permanent environmental (PE) effect of M2 had moderate to strong correlations with the intercept for the AG effect of M1. The coefficients of determination for multiple regression of the combined intercepts for the AG and PE effects of M2 on the coefficients for the AG effect of M1 were moderate to high. The daily milk yields of M2 predicted by using the RR coefficients for the AG effect of M1 were highly correlated with those obtained by using the coefficients of M2. Milk production after 305 DIM can be predicted by using the RR coefficient estimates of the AG effect during the first 305 DIM.
Access disparities to Magnet hospitals for patients undergoing neurosurgical operations
Missios, Symeon; Bekelis, Kimon
2017-01-01
Background Centers of excellence focusing on quality improvement have demonstrated superior outcomes for a variety of surgical interventions. We investigated the presence of access disparities to hospitals recognized by the Magnet Recognition Program of the American Nurses Credentialing Center (ANCC) for patients undergoing neurosurgical operations. Methods We performed a cohort study of all neurosurgery patients who were registered in the New York Statewide Planning and Research Cooperative System (SPARCS) database from 2009–2013. We examined the association of African-American race and lack of insurance with Magnet status hospitalization for neurosurgical procedures. A mixed effects propensity adjusted multivariable regression analysis was used to control for confounding. Results During the study period, 190,535 neurosurgical patients met the inclusion criteria. Using a multivariable logistic regression, we demonstrate that African-Americans had lower admission rates to Magnet institutions (OR 0.62; 95% CI, 0.58–0.67). This persisted in a mixed effects logistic regression model (OR 0.77; 95% CI, 0.70–0.83) to adjust for clustering at the patient county level, and a propensity score adjusted logistic regression model (OR 0.75; 95% CI, 0.69–0.82). Additionally, lack of insurance was associated with lower admission rates to Magnet institutions (OR 0.71; 95% CI, 0.68–0.73), in a multivariable logistic regression model. This persisted in a mixed effects logistic regression model (OR 0.72; 95% CI, 0.69–0.74), and a propensity score adjusted logistic regression model (OR 0.72; 95% CI, 0.69–0.75). Conclusions Using a comprehensive all-payer cohort of neurosurgery patients in New York State we identified an association of African-American race and lack of insurance with lower rates of admission to Magnet hospitals. PMID:28684152
Population heterogeneity in the salience of multiple risk factors for adolescent delinquency.
Lanza, Stephanie T; Cooper, Brittany R; Bray, Bethany C
2014-03-01
To present mixture regression analysis as an alternative to more standard regression analysis for predicting adolescent delinquency. We demonstrate how mixture regression analysis allows for the identification of population subgroups defined by the salience of multiple risk factors. We identified population subgroups (i.e., latent classes) of individuals based on their coefficients in a regression model predicting adolescent delinquency from eight previously established risk indices drawn from the community, school, family, peer, and individual levels. The study included N = 37,763 10th-grade adolescents who participated in the Communities That Care Youth Survey. Standard, zero-inflated, and mixture Poisson and negative binomial regression models were considered. Standard and mixture negative binomial regression models were selected as optimal. The five-class regression model was interpreted based on the class-specific regression coefficients, indicating that risk factors had varying salience across classes of adolescents. Standard regression showed that all risk factors were significantly associated with delinquency. Mixture regression provided more nuanced information, suggesting a unique set of risk factors that were salient for different subgroups of adolescents. Implications for the design of subgroup-specific interventions are discussed. Copyright © 2014 Society for Adolescent Health and Medicine. Published by Elsevier Inc. All rights reserved.
Paramedic-Initiated Home Care Referrals and Use of Home Care and Emergency Medical Services.
Verma, Amol A; Klich, John; Thurston, Adam; Scantlebury, Jordan; Kiss, Alex; Seddon, Gayle; Sinha, Samir K
2018-01-01
We examined the association between paramedic-initiated home care referrals and utilization of home care, 9-1-1, and Emergency Department (ED) services. This was a retrospective cohort study of individuals who received a paramedic-initiated home care referral after a 9-1-1 call between January 1, 2011 and December 31, 2012 in Toronto, Ontario, Canada. Home care, 9-1-1, and ED utilization were compared in the 6 months before and after home care referral. Nonparametric longitudinal regression was performed to assess changes in hours of home care service use and zero-inflated Poisson regression was performed to assess changes in the number of 9-1-1 calls and ambulance transports to ED. During the 24-month study period, 2,382 individuals received a paramedic-initiated home care referral. After excluding individuals who died, were hospitalized, or were admitted to a nursing home, the final study cohort was 1,851. The proportion of the study population receiving home care services increased from 18.2% to 42.5% after referral, representing 450 additional people receiving services. In longitudinal regression analysis, there was an increase of 17.4 hours in total services per person in the six months after referral (95% CI: 1.7-33.1, p = 0.03). The mean number of 9-1-1 calls per person was 1.44 (SD 9.58) before home care referral and 1.20 (SD 7.04) after home care referral in the overall study cohort. This represented a 10% reduction in 9-1-1 calls (95% CI: 7-13%, p < 0.001) in Poisson regression analysis. The mean number of ambulance transports to ED per person was 0.91 (SD 8.90) before home care referral and 0.79 (SD 6.27) after home care referral, representing a 7% reduction (95% CI: 3-11%, p < 0.001) in Poisson regression analysis. When only the participants with complete paramedic and home care records were included in the analysis, the reductions in 9-1-1 calls and ambulance transports to ED were attenuated but remained statistically significant. Paramedic-initiated home care referrals in Toronto were associated with improved access to and use of home care services and may have been associated with reduced 9-1-1 calls and ambulance transports to ED.
Okechukwu, Cassandra A; Kelly, Erin L; Bacic, Janine; DePasquale, Nicole; Hurtado, David; Kossek, Ellen; Sembajwe, Grace
2016-05-01
We analyzed qualitative and quantitative data from U.S.-based employees in 30 long-term care facilities. Analysis of semi-structured interviews from 154 managers informed quantitative analyses. Quantitative data include 1214 employees' scoring of their supervisors and their organizations on family supportiveness (individual scores and aggregated to facility level), and three outcomes: (1), care quality indicators assessed at facility level (n = 30) and collected monthly for six months after employees' data collection; (2), employees' dichotomous survey response on having additional off-site jobs; and (3), proportion of employees with additional jobs at each facility. Thematic analyses revealed that managers operate within the constraints of an industry that simultaneously: (a) employs low-wage employees with multiple work-family challenges, and (b) has firmly institutionalized goals of prioritizing quality of care and minimizing labor costs. Managers universally described providing work-family support and prioritizing care quality as antithetical to each other. Concerns surfaced that family-supportiveness encouraged employees to work additional jobs off-site, compromising care quality. Multivariable linear regression analysis of facility-level data revealed that higher family-supportive supervision was associated with significant decreases in residents' incidence of all pressure ulcers (-2.62%) and other injuries (-9.79%). Higher family-supportive organizational climate was associated with significant decreases in all falls (-17.94%) and falls with injuries (-7.57%). Managers' concerns about additional jobs were not entirely unwarranted: multivariable logistic regression of employee-level data revealed that among employees with children, having family-supportive supervision was associated with significantly higher likelihood of additional off-site jobs (RR 1.46, 95%CI 1.08-1.99), but family-supportive organizational climate was associated with lower likelihood (RR 0.76, 95%CI 0.59-0.99). However, proportion of workers with additional off-site jobs did not significantly predict care quality at facility levels. Although managers perceived providing work-family support and ensuring high care quality as conflicting goals, results suggest that family-supportiveness is associated with better care quality. Copyright © 2016 Elsevier Ltd. All rights reserved.
Okechukwu, Cassandra A.; Kelly, Erin L.; Bacic, Janine; DePasquale, Nicole; Hurtado, David; Kossek, Ellen; Sembajwe, Grace
2016-01-01
We analyzed qualitative and quantitative data from U.S.-based employees in 30 long-term care facilities. Analysis of semi-structured interviews from 154 managers informed quantitative analyses. Quantitative data include 1,214 employees’ scoring of their supervisors and their organizations on family supportiveness (individual scores and aggregated to facility level), and three outcomes: (1), care quality indicators assessed at facility level (n=30) and collected monthly for six months after employees’ data collection; (2), employees’ dichotomous survey response on having additional off-site jobs; and (3), proportion of employees with additional jobs at each facility. Thematic analyses revealed that managers operate within the constraints of an industry that simultaneously: (a) employs low-wage employees with multiple work-family challenges, and (b) has firmly institutionalized goals of prioritizing quality of care and minimizing labor costs. Managers universally described providing work-family support and prioritizing care quality as antithetical to each other. Concerns surfaced that family-supportiveness encouraged employees to work additional jobs off-site, compromising care quality. Multivariable linear regression analysis of facility-level data revealed that higher family-supportive supervision was associated with significant decreases in residents’ incidence of all pressure ulcers (−2.62%) and other injuries (−9.79%). Higher family-supportive organizational climate was associated with significant decreases in all falls (−17.94%) and falls with injuries (−7.57%). Managers’ concerns about additional jobs were not entirely unwarranted: multivariable logistic regression of employee-level data revealed that among employees with children, having family-supportive supervision was associated with significantly higher likelihood of additional off-site jobs (RR 1.46, 95%CI 1.08-1.99), but family-supportive organizational climate was associated with lower likelihood (RR 0.76, 95%CI 0.59-0.99). However, proportion of workers with additional off-site jobs did not significantly predict care quality at facility levels. Although managers perceived providing work-family support and ensuring high care quality as conflicting goals, results suggest that family-supportiveness is associated with better care quality. PMID:27082022
ERIC Educational Resources Information Center
Dolan, Conor V.; Wicherts, Jelte M.; Molenaar, Peter C. M.
2004-01-01
We consider the question of how variation in the number and reliability of indicators affects the power to reject the hypothesis that the regression coefficients are zero in latent linear regression analysis. We show that power remains constant as long as the coefficient of determination remains unchanged. Any increase in the number of indicators…
Methods for estimating streamflow at mountain fronts in southern New Mexico
Waltemeyer, S.D.
1994-01-01
The infiltration of streamflow is potential recharge to alluvial-basin aquifers at or near mountain fronts in southern New Mexico. Data for 13 streamflow-gaging stations were used to determine a relation between mean annual stream- flow and basin and climatic conditions. Regression analysis was used to develop an equation that can be used to estimate mean annual streamflow on the basis of drainage areas and mean annual precipi- tation. The average standard error of estimate for this equation is 46 percent. Regression analysis also was used to develop an equation to estimate mean annual streamflow on the basis of active- channel width. Measurements of the width of active channels were determined for 6 of the 13 gaging stations. The average standard error of estimate for this relation is 29 percent. Stream- flow estimates made using a regression equation based on channel geometry are considered more reliable than estimates made from an equation based on regional relations of basin and climatic conditions. The sample size used to develop these relations was small, however, and the reported standard error of estimate may not represent that of the entire population. Active-channel-width measurements were made at 23 ungaged sites along the Rio Grande upstream from Elephant Butte Reservoir. Data for additional sites would be needed for a more comprehensive assessment of mean annual streamflow in southern New Mexico.
Discrete mixture modeling to address genetic heterogeneity in time-to-event regression
Eng, Kevin H.; Hanlon, Bret M.
2014-01-01
Motivation: Time-to-event regression models are a critical tool for associating survival time outcomes with molecular data. Despite mounting evidence that genetic subgroups of the same clinical disease exist, little attention has been given to exploring how this heterogeneity affects time-to-event model building and how to accommodate it. Methods able to diagnose and model heterogeneity should be valuable additions to the biomarker discovery toolset. Results: We propose a mixture of survival functions that classifies subjects with similar relationships to a time-to-event response. This model incorporates multivariate regression and model selection and can be fit with an expectation maximization algorithm, we call Cox-assisted clustering. We illustrate a likely manifestation of genetic heterogeneity and demonstrate how it may affect survival models with little warning. An application to gene expression in ovarian cancer DNA repair pathways illustrates how the model may be used to learn new genetic subsets for risk stratification. We explore the implications of this model for censored observations and the effect on genomic predictors and diagnostic analysis. Availability and implementation: R implementation of CAC using standard packages is available at https://gist.github.com/programeng/8620b85146b14b6edf8f Data used in the analysis are publicly available. Contact: kevin.eng@roswellpark.org Supplementary information: Supplementary data are available at Bioinformatics online. PMID:24532723
Cao, Zhiyong; Zhang, Ping; He, Zhiqing; Yang, Jing; Liang, Chun; Ren, Yusheng; Wu, Zonggui
2016-05-26
Current study was designed to investigate the effects of obstructive sleep apnea (OSA) combined dyslipidemia on the prevalence of atherosclerotic cardiovascular diseases (ASCVD). This was a cross-sectional study and subjects with documented dyslipidemia and without previous diagnosis of OSA were enrolled. Polysomnography was applied to evaluate apnea-hypopnea index (AHI). Based on AHI value, subjects were classified into four groups: without OSA, mild, moderate and severe OSA groups. Clinical characteristics and laboratory examination data were recorded. Relationship between AHI event and lipid profiles was analyzed, and logistic regression analysis was used to evaluate the effects of OSA combined dyslipidemia on ASCVD prevalence. Totally 248 subjects with dyslipidemia were enrolled. Compared to the other 3 groups, subjects with severe OSA were older, male predominant and had higher smoking rate. In addition, subjects with severe OSA had higher body mass index, waist-hip ratio, blood pressure, and higher rates of overweight and obesity. Serum levels of fasting plasma glucose, glycated hemoglobin, LDL-C and CRP were all significantly higher. ASCVD prevalence was considerably higher in subjects with severe OSA. AHI event in the severe OSA group was up to 35.4 ± 5.1 events per hour which was significantly higher than the other groups (P < 0.05 for trend). Pearson correlation analysis showed that only LDL-C was positively correlated with AHI events (r = 0.685, P < 0.05). Logistic regression analysis revealed that in unadjusted model, compared to dyslipidemia plus no-OSA group (reference group), OSA enhanced ASCVD risk in subjects with dyslipidemia, regardless of OSA severity. After extensively adjusted for confounding variables, the odds of dyslipidemia plus mild-OSA was reduced to insignificance. While the effects of moderate- and severe-OSA on promoting ASCVD risk in subjects with dyslipidemia remained significant, with severe-OSA most prominent (odds ratio: 1.52, 95% confidence interval: 1.13-2.02). OSA combined dyslipidemia conferred additive adverse effects on cardiovascular system, with severe-OSA most prominent.
Bustamante, Alejandro; Vilar-Bergua, Andrea; Guettier, Sophie; Sánchez-Poblet, Josep; García-Berrocoso, Teresa; Giralt, Dolors; Fluri, Felix; Topakian, Raffi; Worthmann, Hans; Hug, Andreas; Molnar, Tihamer; Waje-Andreassen, Ulrike; Katan, Mira; Smith, Craig J; Montaner, Joan
2017-04-01
We conducted a systematic review and individual participant data meta-analysis to explore the role of C-reactive protein (CRP) in early detection or prediction of post-stroke infections. CRP, an acute-phase reactant binds to the phosphocholine expressed on the surface of dead or dying cells and some bacteria, thereby activating complement and promoting phagocytosis by macrophages. We searched PubMed up to May-2015 for studies measuring CRP in stroke and evaluating post-stroke infections. Individual participants' data were merged into a single database. CRP levels were standardized and divided into quartiles. Factors independently associated with post-stroke infections were determined by logistic regression analysis and the additional predictive value of CRP was assessed by comparing areas under receiver operating characteristic curves and integrated discrimination improvement index. Data from seven studies including 699 patients were obtained. Standardized CRP levels were higher in patients with post-stroke infections beyond 24 h. Standardized CRP levels in the fourth quartile were independently associated with infection in two different logistic regression models, model 1 [stroke severity and dysphagia, odds ratio = 9.70 (3.10-30.41)] and model 2 [age, sex, and stroke severity, odds ratio = 3.21 (1.93-5.32)]. Addition of CRP improved discrimination in both models [integrated discrimination improvement = 9.83% (0.89-18.77) and 5.31% (2.83-7.79), respectively], but accuracy was only improved for model 1 (area under the curve 0.806-0.874, p = 0.036). In this study, CRP was independently associated with development of post-stroke infections, with the optimal time-window for measurement at 24-48 h. However, its additional predictive value is moderate over clinical information. Combination with other biomarkers in a panel seems a promising strategy for future studies. © 2017 International Society for Neurochemistry.
NASA Astrophysics Data System (ADS)
Tustison, Nicholas J.; Contrella, Benjamin; Altes, Talissa A.; Avants, Brian B.; de Lange, Eduard E.; Mugler, John P.
2013-03-01
The utitlity of pulmonary functional imaging techniques, such as hyperpolarized 3He MRI, has encouraged their inclusion in research studies for longitudinal assessment of disease progression and the study of treatment effects. We present methodology for performing voxelwise statistical analysis of ventilation maps derived from hyper polarized 3He MRI which incorporates multivariate template construction using simultaneous acquisition of IH and 3He images. Additional processing steps include intensity normalization, bias correction, 4-D longitudinal segmentation, and generation of expected ventilation maps prior to voxelwise regression analysis. Analysis is demonstrated on a cohort of eight individuals with diagnosed cystic fibrosis (CF) undergoing treatment imaged five times every two weeks with a prescribed treatment schedule.
Association between sociability and diffusion tensor imaging in BALB/cJ mice.
Kim, Sungheon; Pickup, Stephen; Fairless, Andrew H; Ittyerah, Ranjit; Dow, Holly C; Abel, Ted; Brodkin, Edward S; Poptani, Harish
2012-01-01
The purpose of this study was to use high-resolution diffusion tensor imaging (DTI) to investigate the association between DTI metrics and sociability in BALB/c inbred mice. The sociability of prepubescent (30-day-old) BALB/cJ mice was operationally defined as the time that the mice spent sniffing a stimulus mouse in a social choice test. High-resolution ex vivo DTI data on 12 BALB/cJ mouse brains were acquired using a 9.4-T vertical-bore magnet. Regression analysis was conducted to investigate the association between DTI metrics and sociability. Significant positive regression (p < 0.001) between social sniffing time and fractional anisotropy was found in 10 regions located in the thalamic nuclei, zona incerta/substantia nigra, visual/orbital/somatosensory cortices and entorhinal cortex. In addition, significant negative regression (p < 0.001) between social sniffing time and mean diffusivity was found in five areas located in the sensory cortex, motor cortex, external capsule and amygdaloid region. In all regions showing significant regression with either the mean diffusivity or fractional anisotropy, the tertiary eigenvalue correlated negatively with the social sniffing time. This study demonstrates the feasibility of using DTI to detect brain regions associated with sociability in a mouse model system. Copyright © 2011 John Wiley & Sons, Ltd.
Curran, Christopher A.; Eng, Ken; Konrad, Christopher P.
2012-01-01
Regional low-flow regression models for estimating Q7,10 at ungaged stream sites are developed from the records of daily discharge at 65 continuous gaging stations (including 22 discontinued gaging stations) for the purpose of evaluating explanatory variables. By incorporating the base-flow recession time constant τ as an explanatory variable in the regression model, the root-mean square error for estimating Q7,10 at ungaged sites can be lowered to 72 percent (for known values of τ), which is 42 percent less than if only basin area and mean annual precipitation are used as explanatory variables. If partial-record sites are included in the regression data set, τ must be estimated from pairs of discharge measurements made during continuous periods of declining low flows. Eight measurement pairs are optimal for estimating τ at partial-record sites, and result in a lowering of the root-mean square error by 25 percent. A low-flow survey strategy that includes paired measurements at partial-record sites requires additional effort and planning beyond a standard strategy, but could be used to enhance regional estimates of τ and potentially reduce the error of regional regression models for estimating low-flow characteristics at ungaged sites.
NASA Astrophysics Data System (ADS)
Masselot, Pierre; Chebana, Fateh; Bélanger, Diane; St-Hilaire, André; Abdous, Belkacem; Gosselin, Pierre; Ouarda, Taha B. M. J.
2018-07-01
In environmental epidemiology studies, health response data (e.g. hospitalization or mortality) are often noisy because of hospital organization and other social factors. The noise in the data can hide the true signal related to the exposure. The signal can be unveiled by performing a temporal aggregation on health data and then using it as the response in regression analysis. From aggregated series, a general methodology is introduced to account for the particularities of an aggregated response in a regression setting. This methodology can be used with usually applied regression models in weather-related health studies, such as generalized additive models (GAM) and distributed lag nonlinear models (DLNM). In particular, the residuals are modelled using an autoregressive-moving average (ARMA) model to account for the temporal dependence. The proposed methodology is illustrated by modelling the influence of temperature on cardiovascular mortality in Canada. A comparison with classical DLNMs is provided and several aggregation methods are compared. Results show that there is an increase in the fit quality when the response is aggregated, and that the estimated relationship focuses more on the outcome over several days than the classical DLNM. More precisely, among various investigated aggregation schemes, it was found that an aggregation with an asymmetric Epanechnikov kernel is more suited for studying the temperature-mortality relationship.
Decreased levels of sRAGE in follicular fluid from patients with PCOS.
Wang, BiJun; Li, Jing; Yang, QingLing; Zhang, FuLi; Hao, MengMeng; Guo, YiHong
2017-03-01
This study aimed to explore the association between soluble receptor for advanced glycation end products (sRAGE) levels in follicular fluid and the number of oocytes retrieved and to evaluate the effect of sRAGE on vascular endothelial growth factor (VEGF) in granulosa cells in patients with polycystic ovarian syndrome (PCOS). Two sets of experiments were performed in this study. In part one, sRAGE and VEGF protein levels in follicular fluid samples from 39 patients with PCOS and 35 non-PCOS patients were measured by ELISA. In part two, ovarian granulosa cells were isolated from an additional 10 patients with PCOS and cultured. VEGF and SP1 mRNA and protein levels, as well as pAKT levels, were detected by real-time PCR and Western blotting after cultured cells were treated with different concentrations of sRAGE. Compared with the non-PCOS patients, patients with PCOS had lower sRAGE levels in follicular fluid. Multi-adjusted regression analysis showed that high sRAGE levels in follicular fluid predicted a lower Gn dose, more oocytes retrieved, and a better IVF outcome in the non-PCOS group. Logistic regression analysis showed that higher sRAGE levels predicted favorably IVF outcomes in the non-PCOS group. Multi-adjusted regression analysis also showed that high sRAGE levels in follicular fluid predicted a lower Gn dose in the PCOS group. Treating granulosa cells isolated from patients with PCOS with recombinant sRAGE decreased VEGF and SP1 mRNA and protein expression and pAKT levels in a dose-dependent manner. © 2017 Society for Reproduction and Fertility.
Linden, Ariel
2018-04-01
Interrupted time series analysis (ITSA) is an evaluation methodology in which a single treatment unit's outcome is studied over time and the intervention is expected to "interrupt" the level and/or trend of the outcome. The internal validity is strengthened considerably when the treated unit is contrasted with a comparable control group. In this paper, we introduce a robust evaluation framework that combines the synthetic controls method (SYNTH) to generate a comparable control group and ITSA regression to assess covariate balance and estimate treatment effects. We evaluate the effect of California's Proposition 99 for reducing cigarette sales, by comparing California to other states not exposed to smoking reduction initiatives. SYNTH is used to reweight nontreated units to make them comparable to the treated unit. These weights are then used in ITSA regression models to assess covariate balance and estimate treatment effects. Covariate balance was achieved for all but one covariate. While California experienced a significant decrease in the annual trend of cigarette sales after Proposition 99, there was no statistically significant treatment effect when compared to synthetic controls. The advantage of using this framework over regression alone is that it ensures that a comparable control group is generated. Additionally, it offers a common set of statistical measures familiar to investigators, the capability for assessing covariate balance, and enhancement of the evaluation with a comprehensive set of postestimation measures. Therefore, this robust framework should be considered as a primary approach for evaluating treatment effects in multiple group time series analysis. © 2018 John Wiley & Sons, Ltd.
Ghasemi, Jahan B; Safavi-Sohi, Reihaneh; Barbosa, Euzébio G
2012-02-01
A quasi 4D-QSAR has been carried out on a series of potent Gram-negative LpxC inhibitors. This approach makes use of the molecular dynamics (MD) trajectories and topology information retrieved from the GROMACS package. This new methodology is based on the generation of a conformational ensemble profile, CEP, for each compound instead of only one conformation, followed by the calculation intermolecular interaction energies at each grid point considering probes and all aligned conformations resulting from MD simulations. These interaction energies are independent variables employed in a QSAR analysis. The comparison of the proposed methodology to comparative molecular field analysis (CoMFA) formalism was performed. This methodology explores jointly the main features of CoMFA and 4D-QSAR models. Step-wise multiple linear regression was used for the selection of the most informative variables. After variable selection, multiple linear regression (MLR) and partial least squares (PLS) methods used for building the regression models. Leave-N-out cross-validation (LNO), and Y-randomization were performed in order to confirm the robustness of the model in addition to analysis of the independent test set. Best models provided the following statistics: [Formula in text] (PLS) and [Formula in text] (MLR). Docking study was applied to investigate the major interactions in protein-ligand complex with CDOCKER algorithm. Visualization of the descriptors of the best model helps us to interpret the model from the chemical point of view, supporting the applicability of this new approach in rational drug design.
NASA Astrophysics Data System (ADS)
Buermeyer, Jonas; Gundlach, Matthias; Grund, Anna-Lisa; Grimm, Volker; Spizyn, Alexander; Breckow, Joachim
2016-09-01
This work is part of the analysis of the effects of constructional energy-saving measures to radon concentration levels in dwellings performed on behalf of the German Federal Office for Radiation Protection. In parallel to radon measurements for five buildings, both meteorological data outside the buildings and the indoor climate factors were recorded. In order to access effects of inhabited buildings, the amount of carbon dioxide (CO2) was measured. For a statistical linear regression model, the data of one object was chosen as an example. Three dummy variables were extracted from the process of the CO2 concentration to provide information on the usage and ventilation of the room. The analysis revealed a highly autoregressive model for the radon concentration with additional influence by the natural environmental factors. The autoregression implies a strong dependency on a radon source since it reflects a backward dependency in time. At this point of the investigation, it cannot be determined whether the influence by outside factors affects the source of radon or the habitant’s ventilation behavior resulting in variation of the occurring concentration levels. In any case, the regression analysis might provide further information that would help to distinguish these effects. In the next step, the influence factors will be weighted according to their impact on the concentration levels. This might lead to a model that enables the prediction of radon concentration levels based on the measurement of CO2 in combination with environmental parameters, as well as the development of advices for ventilation.
Hu, Meng; Clark, Kelsey L.; Gong, Xiajing; Noudoost, Behrad; Li, Mingyao; Moore, Tirin
2015-01-01
Inferotemporal (IT) neurons are known to exhibit persistent, stimulus-selective activity during the delay period of object-based working memory tasks. Frontal eye field (FEF) neurons show robust, spatially selective delay period activity during memory-guided saccade tasks. We present a copula regression paradigm to examine neural interaction of these two types of signals between areas IT and FEF of the monkey during a working memory task. This paradigm is based on copula models that can account for both marginal distribution over spiking activity of individual neurons within each area and joint distribution over ensemble activity of neurons between areas. Considering the popular GLMs as marginal models, we developed a general and flexible likelihood framework that uses the copula to integrate separate GLMs into a joint regression analysis. Such joint analysis essentially leads to a multivariate analog of the marginal GLM theory and hence efficient model estimation. In addition, we show that Granger causality between spike trains can be readily assessed via the likelihood ratio statistic. The performance of this method is validated by extensive simulations, and compared favorably to the widely used GLMs. When applied to spiking activity of simultaneously recorded FEF and IT neurons during working memory task, we observed significant Granger causality influence from FEF to IT, but not in the opposite direction, suggesting the role of the FEF in the selection and retention of visual information during working memory. The copula model has the potential to provide unique neurophysiological insights about network properties of the brain. PMID:26063909
Secular trends in Cherokee cranial morphology: Eastern vs Western bands.
Sutphin, Rebecca; Ross, Ann H; Jantz, Richard L
2014-01-01
The research objective was to examine if secular trends can be identified for cranial data commissioned by Boas in 1892, specifically for cranial breadth and cranial length of the Eastern and Western band Cherokee who experienced environmental hardships. Multiple regression analysis was used to test the degree of relationship between each of the cranial measures: cranial length, cranial breadth and cephalic index, along with predictor variables (year-of-birth, location, sex, admixture); the model revealed a significant difference for all craniometric variables. Additional regression analysis was performed with smoothing Loess plots to observe cranial length and cranial breadth change over time (year-of-birth) separately for Eastern and Western Cherokee band females and males born between 1783-1874. This revealed the Western and Eastern bands show a decrease in cranial length over time. Eastern band individuals maintain a relatively constant head breadth, while Western Band individuals show a sharp decline beginning around 1860. These findings support negative secular trend occurring for both Cherokee bands where the environment made a detrimental impact; this is especially marked with the Western Cherokee band.
Big Data Toolsets to Pharmacometrics: Application of Machine Learning for Time-to-Event Analysis.
Gong, Xiajing; Hu, Meng; Zhao, Liang
2018-05-01
Additional value can be potentially created by applying big data tools to address pharmacometric problems. The performances of machine learning (ML) methods and the Cox regression model were evaluated based on simulated time-to-event data synthesized under various preset scenarios, i.e., with linear vs. nonlinear and dependent vs. independent predictors in the proportional hazard function, or with high-dimensional data featured by a large number of predictor variables. Our results showed that ML-based methods outperformed the Cox model in prediction performance as assessed by concordance index and in identifying the preset influential variables for high-dimensional data. The prediction performances of ML-based methods are also less sensitive to data size and censoring rates than the Cox regression model. In conclusion, ML-based methods provide a powerful tool for time-to-event analysis, with a built-in capacity for high-dimensional data and better performance when the predictor variables assume nonlinear relationships in the hazard function. © 2018 The Authors. Clinical and Translational Science published by Wiley Periodicals, Inc. on behalf of American Society for Clinical Pharmacology and Therapeutics.
Hejri-Zarifi, Sudiyeh; Ahmadian-Kouchaksaraei, Zahra; Pourfarzad, Amir; Khodaparast, Mohammad Hossein Haddad
2014-12-01
Germinated palm date seeds were milled into two fractions: germ and residue. Dough rheological characteristics, baking (specific volume and sensory evaluation), and textural properties (at first day and during storage for 5 days) were determined in Barbari flat bread. Germ and residue fractions were incorporated at various levels ranged in 0.5-3 g/100 g of wheat flour. Water absorption, arrival time and gelatination temperature were decreased by germ fraction but accompanied by an increasing effect on the mixing tolerance index and degree of softening in most levels. Although improvement in dough stability was monitored but specific volume of bread was not affected by both fractions. Texture analysis of bread samples during 5 days of storage indicated that both fractions of germinated date seeds were able to diminish bread staling. Avrami non-linear regression equation was chosen as useful mathematical model to properly study bread hardening kinetics. In addition, principal component analysis (PCA) allowed discriminating among dough and bread specialties. Partial least squares regression (PLSR) models were applied to determine the relationships between sensory and instrumental data.
Psychosocial variables and time to injury onset: a hurdle regression analysis model.
Sibold, Jeremy; Zizzi, Samuel
2012-01-01
Psychological variables have been shown to be related to athletic injury and time missed from participation in sport. We are unaware of any empirical examination of the influence of psychological variables on time to onset of injury. To examine the influence of orthopaedic and psychosocial variables on time to injury in college athletes. One hundred seventy-seven (men 5 116, women 5 61; age 5 19.45 6 1.39 years) National Collegiate Athletic Association Division II athletes. Hurdle regression analysis (HRA) was used to determine the influence of predictor variables on days to first injury. Worry (z = 2.98, P = .003), concentration disruption (z = -3.95, P < .001), and negative life-event stress (z = 5.02, P < .001) were robust predictors of days to injury. Orthopaedic risk score was not a predictor (z = 1.28, P = .20). These findings support previous research on the stress-injury relationship, and our group is the first to use HRA in athletic injury data. These data support the addition of psychological screening as part of preseason health examinations for collegiate athletes.
Cross-cultural relationships between self-concept and body image in high school-age boys.
Austin, J K; Champion, V L; Tzeng, O C
1989-08-01
The relationship between self-concept and body image was investigated through a secondary analysis of data from a sample of 1,200 high school male students from 30 language/culture communities (Osgood, May, & Myron, 1975). Subjects rated adjectives pertaining to self-concept and body image using 7-step semantic differential bipolar scales. Adjectives were related to the dimensions of Evaluation, Potency, and Activity. Correlation, factor analysis, and multiple regression were utilized to examine multivariate relationships among self-concept dimensions and body-image dimensions. Significant positive correlations were found between self-concept and body image. In addition, significant positive relationships were found when self-concept factors were regressed on the body-image factor (R2 = .49 to .57, p less than or equal to .001) for Activity and Potency. Results support the existence of a strong positive relationship between self-concept and body image across the 30 cultures involved. Findings have important implications for nursing in assessment and interventions with clients who have deficits in either self-concept or body image.
Applications of modern statistical methods to analysis of data in physical science
NASA Astrophysics Data System (ADS)
Wicker, James Eric
Modern methods of statistical and computational analysis offer solutions to dilemmas confronting researchers in physical science. Although the ideas behind modern statistical and computational analysis methods were originally introduced in the 1970's, most scientists still rely on methods written during the early era of computing. These researchers, who analyze increasingly voluminous and multivariate data sets, need modern analysis methods to extract the best results from their studies. The first section of this work showcases applications of modern linear regression. Since the 1960's, many researchers in spectroscopy have used classical stepwise regression techniques to derive molecular constants. However, problems with thresholds of entry and exit for model variables plagues this analysis method. Other criticisms of this kind of stepwise procedure include its inefficient searching method, the order in which variables enter or leave the model and problems with overfitting data. We implement an information scoring technique that overcomes the assumptions inherent in the stepwise regression process to calculate molecular model parameters. We believe that this kind of information based model evaluation can be applied to more general analysis situations in physical science. The second section proposes new methods of multivariate cluster analysis. The K-means algorithm and the EM algorithm, introduced in the 1960's and 1970's respectively, formed the basis of multivariate cluster analysis methodology for many years. However, several shortcomings of these methods include strong dependence on initial seed values and inaccurate results when the data seriously depart from hypersphericity. We propose new cluster analysis methods based on genetic algorithms that overcomes the strong dependence on initial seed values. In addition, we propose a generalization of the Genetic K-means algorithm which can accurately identify clusters with complex hyperellipsoidal covariance structures. We then use this new algorithm in a genetic algorithm based Expectation-Maximization process that can accurately calculate parameters describing complex clusters in a mixture model routine. Using the accuracy of this GEM algorithm, we assign information scores to cluster calculations in order to best identify the number of mixture components in a multivariate data set. We will showcase how these algorithms can be used to process multivariate data from astronomical observations.
Berardi, Cecilia; Larson, Nicholas B.; Decker, Paul A.; Wassel, Christina L.; Kirsch, Phillip S.; Pankow, James S.; Sale, Michele M.; de Andrade, Mariza; Sicotte, Hugues; Tang, Weihong; Hanson, Naomi Q.; Tsai, Michael Y.; da Chen, Yii-Der I; Bielinski, Suzette J.
2015-01-01
L-selectin is constitutively expressed on leukocytes and mediates their interaction with endothelial cells during inflammation. Previous studies on the association of soluble L-selectin (sL-selectin) with cardiovascular disease (CVD) are inconsistent. Genetic variants associated with sL-selectin levels may be a better surrogate of levels over a lifetime. We explored the association of genetic variants and sL-selectin levels in a race/ethnicity stratified random sample of 2,403 participants in the Multi-Ethnic Study of Atherosclerosis (MESA). Through a genome-wide analysis with additive linear regression models, we found that rs12938 on the SELL gene accounted for a significant portion of the protein level variance across all four races/ethnicities. To evaluate potential additional associations, elastic net models were used for variants located in the SELL/SELP/SELE genetic region and an additional two SNPs, rs3917768 and rs4987361, were associated with sL-selectin levels in African Americans. These variants accounted for a portion of protein variance that ranged from 4% in Hispanic to 14% in African Americans. To investigate the relationship of these variants with CVD, 6,317 subjects were used. No significant association was found between any of the identified SNPs and carotid intima-media thickness or presence of carotid plaque using linear and logistic regression, respectively. Similarly no significant results were found for coronary artery calcium or coronary heart disease events. In conclusion, we found that variants within the SELL gene are associated with sL-selectin levels. Despite accounting for a significant portion of the protein level variance, none of the variants was associated with clinical or subclinical CVD. PMID:25576479
Sörös, Peter; Bachmann, Katharina; Lam, Alexandra P; Kanat, Manuela; Hoxhaj, Eliza; Matthies, Swantje; Feige, Bernd; Müller, Helge H O; Thiel, Christiane; Philipsen, Alexandra
2017-01-01
Attention-deficit/hyperactivity disorder (ADHD) in adulthood is a serious and frequent psychiatric disorder with the core symptoms inattention, impulsivity, and hyperactivity. The principal aim of this study was to investigate associations between brain morphology, i.e., cortical thickness and volumes of subcortical gray matter, and individual symptom severity in adult ADHD. Surface-based brain morphometry was performed in 35 women and 29 men with ADHD using FreeSurfer. Linear regressions were calculated between cortical thickness and the volumes of subcortical gray matter and the inattention, hyperactivity, and impulsivity subscales of the Conners Adult ADHD Rating Scales (CAARS). Two separate analyses were performed. For the first analysis, age was included as additional regressor. For the second analysis, both age and severity of depression were included as additional regressors. Study participants were recruited between June 2012 and January 2014. Linear regression identified an area in the left occipital cortex of men, covering parts of the middle occipital sulcus and gyrus, in which the score on the CAARS inattention subscale predicted increased mean cortical thickness [ F (1,27) = 26.27, p < 0.001, adjusted R 2 = 0.4744]. No significant associations were found between cortical thickness and the scores on CAARS subscales in women. No significant associations were found between the volumes of subcortical gray matter and the scores on CAARS subscales, neither in men nor in women. These results remained stable when severity of depression was included as additional regressor, together with age. Increased cortical thickness in the left occipital cortex may represent a mechanism to compensate for dysfunctional attentional networks in male adult ADHD patients.
Moderation analysis using a two-level regression model.
Yuan, Ke-Hai; Cheng, Ying; Maxwell, Scott
2014-10-01
Moderation analysis is widely used in social and behavioral research. The most commonly used model for moderation analysis is moderated multiple regression (MMR) in which the explanatory variables of the regression model include product terms, and the model is typically estimated by least squares (LS). This paper argues for a two-level regression model in which the regression coefficients of a criterion variable on predictors are further regressed on moderator variables. An algorithm for estimating the parameters of the two-level model by normal-distribution-based maximum likelihood (NML) is developed. Formulas for the standard errors (SEs) of the parameter estimates are provided and studied. Results indicate that, when heteroscedasticity exists, NML with the two-level model gives more efficient and more accurate parameter estimates than the LS analysis of the MMR model. When error variances are homoscedastic, NML with the two-level model leads to essentially the same results as LS with the MMR model. Most importantly, the two-level regression model permits estimating the percentage of variance of each regression coefficient that is due to moderator variables. When applied to data from General Social Surveys 1991, NML with the two-level model identified a significant moderation effect of race on the regression of job prestige on years of education while LS with the MMR model did not. An R package is also developed and documented to facilitate the application of the two-level model.
Park, Hyunjin; Yang, Jin-ju; Seo, Jongbum; Choi, Yu-yong; Lee, Kun-ho; Lee, Jong-min
2014-04-01
Cortical features derived from magnetic resonance imaging (MRI) provide important information to account for human intelligence. Cortical thickness, surface area, sulcal depth, and mean curvature were considered to explain human intelligence. One region of interest (ROI) of a cortical structure consisting of thousands of vertices contained thousands of measurements, and typically, one mean value (first order moment), was used to represent a chosen ROI, which led to a potentially significant loss of information. We proposed a technological improvement to account for human intelligence in which a second moment (variance) in addition to the mean value was adopted to represent a chosen ROI, so that the loss of information would be less severe. Two computed moments for the chosen ROIs were analyzed with partial least squares regression (PLSR). Cortical features for 78 adults were measured and analyzed in conjunction with the full-scale intelligence quotient (FSIQ). Our results showed that 45% of the variance of the FSIQ could be explained using the combination of four cortical features using two moments per chosen ROI. Our results showed improvement over using a mean value for each ROI, which explained 37% of the variance of FSIQ using the same set of cortical measurements. Our results suggest that using additional second order moments is potentially better than using mean values of chosen ROIs for regression analysis to account for human intelligence. Copyright © 2014 Elsevier Ltd. All rights reserved.
Kinoshita, Kodzue; Kuze, Noko; Kobayashi, Toshio; Miyakawa, Etsuko; Narita, Hiromitsu; Inoue-Murayama, Miho; Idani, Gen'ichi; Tsenkova, Roumiana
2016-01-01
For promoting in situ conservation, it is important to estimate the density distribution of fertile individuals, and there is a need for developing an easy monitoring method to discriminate between physiological states. To date, physiological state has generally been determined by measuring hormone concentration using radioimmunoassay or enzyme immunoassay (EIA) methods. However, these methods have rarely been applied in situ because of the requirements for a large amount of reagent, instruments, and a radioactive isotope. In addition, the proper storage of the sample (including urine and feces) on site until analysis is difficult. On the other hand, near infrared (NIR) spectroscopy requires no reagent and enables rapid measurement. In the present study, we attempted urinary NIR spectroscopy to determine the estrogen levels of orangutans in Japanese zoos and in the Danum Valley Conservation Area, Sabah, Malaysia. Reflectance NIR spectra were obtained from urine stored using a filter paper. Filter paper is easy to use to store dried urine, even in the wild. Urinary estrogen and creatinine concentrations measured by EIA were used as the reference data of partial least square (PLS) regression of urinary NIR spectra. High accuracies (R(2) > 0.68) were obtained in both estrogen and creatinine regression models. In addition, the PLS regressions in both standards showed higher accuracies (R(2) > 0.70). Therefore, the present study demonstrates that urinary NIR spectra have the potential to estimate the estrogen and creatinine concentrations.
Bramness, Jørgen G; Walby, Fredrik A; Morken, Gunnar; Røislien, Jo
2015-08-01
Seasonal variation in the number of suicides has long been acknowledged. It has been suggested that this seasonality has declined in recent years, but studies have generally used statistical methods incapable of confirming this. We examined all suicides occurring in Norway during 1969-2007 (more than 20,000 suicides in total) to establish whether seasonality decreased over time. Fitting of additive Fourier Poisson time-series regression models allowed for formal testing of a possible linear decrease in seasonality, or a reduction at a specific point in time, while adjusting for a possible smooth nonlinear long-term change without having to categorize time into discrete yearly units. The models were compared using Akaike's Information Criterion and analysis of variance. A model with a seasonal pattern was significantly superior to a model without one. There was a reduction in seasonality during the period. Both the model assuming a linear decrease in seasonality and the model assuming a change at a specific point in time were both superior to a model assuming constant seasonality, thus confirming by formal statistical testing that the magnitude of the seasonality in suicides has diminished. The additive Fourier Poisson time-series regression model would also be useful for studying other temporal phenomena with seasonal components. © The Author 2015. Published by Oxford University Press on behalf of the Johns Hopkins Bloomberg School of Public Health. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
Multiple Correlation versus Multiple Regression.
ERIC Educational Resources Information Center
Huberty, Carl J.
2003-01-01
Describes differences between multiple correlation analysis (MCA) and multiple regression analysis (MRA), showing how these approaches involve different research questions and study designs, different inferential approaches, different analysis strategies, and different reported information. (SLD)
Vargas, Maria; Chiumello, Davide; Sutherasan, Yuda; Ball, Lorenzo; Esquinas, Antonio M; Pelosi, Paolo; Servillo, Giuseppe
2017-05-29
The aims of this systematic review and meta-analysis of randomized controlled trials are to evaluate the effects of active heated humidifiers (HHs) and moisture exchangers (HMEs) in preventing artificial airway occlusion and pneumonia, and on mortality in adult critically ill patients. In addition, we planned to perform a meta-regression analysis to evaluate the relationship between the incidence of artificial airway occlusion, pneumonia and mortality and clinical features of adult critically ill patients. Computerized databases were searched for randomized controlled trials (RCTs) comparing HHs and HMEs and reporting artificial airway occlusion, pneumonia and mortality as predefined outcomes. Relative risk (RR), 95% confidence interval for each outcome and I 2 were estimated for each outcome. Furthermore, weighted random-effect meta-regression analysis was performed to test the relationship between the effect size on each considered outcome and covariates. Eighteen RCTs and 2442 adult critically ill patients were included in the analysis. The incidence of artificial airway occlusion (RR = 1.853; 95% CI 0.792-4.338), pneumonia (RR = 932; 95% CI 0.730-1.190) and mortality (RR = 1.023; 95% CI 0.878-1.192) were not different in patients treated with HMEs and HHs. However, in the subgroup analyses the incidence of airway occlusion was higher in HMEs compared with HHs with non-heated wire (RR = 3.776; 95% CI 1.560-9.143). According to the meta-regression, the effect size in the treatment group on artificial airway occlusion was influenced by the percentage of patients with pneumonia (β = -0.058; p = 0.027; favors HMEs in studies with high prevalence of pneumonia), and a trend was observed for an effect of the duration of mechanical ventilation (MV) (β = -0.108; p = 0.054; favors HMEs in studies with longer MV time). In this meta-analysis we found no superiority of HMEs and HHs, in terms of artificial airway occlusion, pneumonia and mortality. A trend favoring HMEs was observed in studies including a high percentage of patients with pneumonia diagnosis at admission and those with prolonged MV. However, the choice of humidifiers should be made according to the clinical context, trying to avoid possible complications and reaching the appropriate performance at lower costs.
Functional Relationships and Regression Analysis.
ERIC Educational Resources Information Center
Preece, Peter F. W.
1978-01-01
Using a degenerate multivariate normal model for the distribution of organismic variables, the form of least-squares regression analysis required to estimate a linear functional relationship between variables is derived. It is suggested that the two conventional regression lines may be considered to describe functional, not merely statistical,…
Isolating and Examining Sources of Suppression and Multicollinearity in Multiple Linear Regression
ERIC Educational Resources Information Center
Beckstead, Jason W.
2012-01-01
The presence of suppression (and multicollinearity) in multiple regression analysis complicates interpretation of predictor-criterion relationships. The mathematical conditions that produce suppression in regression analysis have received considerable attention in the methodological literature but until now nothing in the way of an analytic…
General Nature of Multicollinearity in Multiple Regression Analysis.
ERIC Educational Resources Information Center
Liu, Richard
1981-01-01
Discusses multiple regression, a very popular statistical technique in the field of education. One of the basic assumptions in regression analysis requires that independent variables in the equation should not be highly correlated. The problem of multicollinearity and some of the solutions to it are discussed. (Author)
Logistic Regression: Concept and Application
ERIC Educational Resources Information Center
Cokluk, Omay
2010-01-01
The main focus of logistic regression analysis is classification of individuals in different groups. The aim of the present study is to explain basic concepts and processes of binary logistic regression analysis intended to determine the combination of independent variables which best explain the membership in certain groups called dichotomous…
Predictors of High Profit and High Deficit Outliers under SwissDRG of a Tertiary Care Center
Mehra, Tarun; Müller, Christian Thomas Benedikt; Volbracht, Jörk; Seifert, Burkhardt; Moos, Rudolf
2015-01-01
Principles Case weights of Diagnosis Related Groups (DRGs) are determined by the average cost of cases from a previous billing period. However, a significant amount of cases are largely over- or underfunded. We therefore decided to analyze earning outliers of our hospital as to search for predictors enabling a better grouping under SwissDRG. Methods 28,893 inpatient cases without additional private insurance discharged from our hospital in 2012 were included in our analysis. Outliers were defined by the interquartile range method. Predictors for deficit and profit outliers were determined with logistic regressions. Predictors were shortlisted with the LASSO regularized logistic regression method and compared to results of Random forest analysis. 10 of these parameters were selected for quantile regression analysis as to quantify their impact on earnings. Results Psychiatric diagnosis and admission as an emergency case were significant predictors for higher deficit with negative regression coefficients for all analyzed quantiles (p<0.001). Admission from an external health care provider was a significant predictor for a higher deficit in all but the 90% quantile (p<0.001 for Q10, Q20, Q50, Q80 and p = 0.0017 for Q90). Burns predicted higher earnings for cases which were favorably remunerated (p<0.001 for the 90% quantile). Osteoporosis predicted a higher deficit in the most underfunded cases, but did not predict differences in earnings for balanced or profitable cases (Q10 and Q20: p<0.00, Q50: p = 0.10, Q80: p = 0.88 and Q90: p = 0.52). ICU stay, mechanical and patient clinical complexity level score (PCCL) predicted higher losses at the 10% quantile but also higher profits at the 90% quantile (p<0.001). Conclusion We suggest considering psychiatric diagnosis, admission as an emergencay case and admission from an external health care provider as DRG split criteria as they predict large, consistent and significant losses. PMID:26517545
Ramsthaler, Frank; Kettner, Mattias; Verhoff, Marcel A
2014-01-01
In forensic anthropological casework, estimating age-at-death is key to profiling unknown skeletal remains. The aim of this study was to examine the reliability of a new, simple, fast, and inexpensive digital odontological method for age-at-death estimation. The method is based on the original Lamendin method, which is a widely used technique in the repertoire of odontological aging methods in forensic anthropology. We examined 129 single root teeth employing a digital camera and imaging software for the measurement of the luminance of the teeth's translucent root zone. Variability in luminance detection was evaluated using statistical technical error of measurement analysis. The method revealed stable values largely unrelated to observer experience, whereas requisite formulas proved to be camera-specific and should therefore be generated for an individual recording setting based on samples of known chronological age. Multiple regression analysis showed a highly significant influence of the coefficients of the variables "arithmetic mean" and "standard deviation" of luminance for the regression formula. For the use of this primer multivariate equation for age-at-death estimation in casework, a standard error of the estimate of 6.51 years was calculated. Step-by-step reduction of the number of embedded variables to linear regression analysis employing the best contributor "arithmetic mean" of luminance yielded a regression equation with a standard error of 6.72 years (p < 0.001). The results of this study not only support the premise of root translucency as an age-related phenomenon, but also demonstrate that translucency reflects a number of other influencing factors in addition to age. This new digital measuring technique of the zone of dental root luminance can broaden the array of methods available for estimating chronological age, and furthermore facilitate measurement and age classification due to its low dependence on observer experience.
Sophocleous, M.
2000-01-01
A practical methodology for recharge characterization was developed based on several years of field-oriented research at 10 sites in the Great Bend Prairie of south-central Kansas. This methodology combines the soil-water budget on a storm-by-storm year-round basis with the resulting watertable rises. The estimated 1985-1992 average annual recharge was less than 50mm/year with a range from 15 mm/year (during the 1998 drought) to 178 mm/year (during the 1993 flood year). Most of this recharge occurs during the spring months. To regionalize these site-specific estimates, an additional methodology based on multiple (forward) regression analysis combined with classification and GIS overlay analyses was developed and implemented. The multiple regression analysis showed that the most influential variables were, in order of decreasing importance, total annual precipitation, average maximum springtime soil-profile water storage, average shallowest springtime depth to watertable, and average springtime precipitation rate. Therefore, four GIS (ARC/INFO) data "layers" or coverages were constructed for the study region based on these four variables, and each such coverage was classified into the same number of data classes to avoid biasing the results. The normalized regression coefficients were employed to weigh the class rankings of each recharge-affecting variable. This approach resulted in recharge zonations that agreed well with the site recharge estimates. During the "Great Flood of 1993," when rainfall totals exceeded normal levels by -200% in the northern portion of the study region, the developed regionalization methodology was tested against such extreme conditions, and proved to be both practical, based on readily available or easily measurable data, and robust. It was concluded that the combination of multiple regression and GIS overlay analyses is a powerful and practical approach to regionalizing small samples of recharge estimates.
Predictors of High Profit and High Deficit Outliers under SwissDRG of a Tertiary Care Center.
Mehra, Tarun; Müller, Christian Thomas Benedikt; Volbracht, Jörk; Seifert, Burkhardt; Moos, Rudolf
2015-01-01
Case weights of Diagnosis Related Groups (DRGs) are determined by the average cost of cases from a previous billing period. However, a significant amount of cases are largely over- or underfunded. We therefore decided to analyze earning outliers of our hospital as to search for predictors enabling a better grouping under SwissDRG. 28,893 inpatient cases without additional private insurance discharged from our hospital in 2012 were included in our analysis. Outliers were defined by the interquartile range method. Predictors for deficit and profit outliers were determined with logistic regressions. Predictors were shortlisted with the LASSO regularized logistic regression method and compared to results of Random forest analysis. 10 of these parameters were selected for quantile regression analysis as to quantify their impact on earnings. Psychiatric diagnosis and admission as an emergency case were significant predictors for higher deficit with negative regression coefficients for all analyzed quantiles (p<0.001). Admission from an external health care provider was a significant predictor for a higher deficit in all but the 90% quantile (p<0.001 for Q10, Q20, Q50, Q80 and p = 0.0017 for Q90). Burns predicted higher earnings for cases which were favorably remunerated (p<0.001 for the 90% quantile). Osteoporosis predicted a higher deficit in the most underfunded cases, but did not predict differences in earnings for balanced or profitable cases (Q10 and Q20: p<0.00, Q50: p = 0.10, Q80: p = 0.88 and Q90: p = 0.52). ICU stay, mechanical and patient clinical complexity level score (PCCL) predicted higher losses at the 10% quantile but also higher profits at the 90% quantile (p<0.001). We suggest considering psychiatric diagnosis, admission as an emergency case and admission from an external health care provider as DRG split criteria as they predict large, consistent and significant losses.
Spatial Assessment of Model Errors from Four Regression Techniques
Lianjun Zhang; Jeffrey H. Gove; Jeffrey H. Gove
2005-01-01
Fomst modelers have attempted to account for the spatial autocorrelations among trees in growth and yield models by applying alternative regression techniques such as linear mixed models (LMM), generalized additive models (GAM), and geographicalIy weighted regression (GWR). However, the model errors are commonly assessed using average errors across the entire study...
Applying Regression Analysis to Problems in Institutional Research.
ERIC Educational Resources Information Center
Bohannon, Tom R.
1988-01-01
Regression analysis is one of the most frequently used statistical techniques in institutional research. Principles of least squares, model building, residual analysis, influence statistics, and multi-collinearity are described and illustrated. (Author/MSE)
Factor weighting in DRASTIC modeling.
Pacheco, F A L; Pires, L M G R; Santos, R M B; Sanches Fernandes, L F
2015-02-01
Evaluation of aquifer vulnerability comprehends the integration of very diverse data, including soil characteristics (texture), hydrologic settings (recharge), aquifer properties (hydraulic conductivity), environmental parameters (relief), and ground water quality (nitrate contamination). It is therefore a multi-geosphere problem to be handled by a multidisciplinary team. The DRASTIC model remains the most popular technique in use for aquifer vulnerability assessments. The algorithm calculates an intrinsic vulnerability index based on a weighted addition of seven factors. In many studies, the method is subject to adjustments, especially in the factor weights, to meet the particularities of the studied regions. However, adjustments made by different techniques may lead to markedly different vulnerabilities and hence to insecurity in the selection of an appropriate technique. This paper reports the comparison of 5 weighting techniques, an enterprise not attempted before. The studied area comprises 26 aquifer systems located in Portugal. The tested approaches include: the Delphi consensus (original DRASTIC, used as reference), Sensitivity Analysis, Spearman correlations, Logistic Regression and Correspondence Analysis (used as adjustment techniques). In all cases but Sensitivity Analysis, adjustment techniques have privileged the factors representing soil characteristics, hydrologic settings, aquifer properties and environmental parameters, by leveling their weights to ≈4.4, and have subordinated the factors describing the aquifer media by downgrading their weights to ≈1.5. Logistic Regression predicts the highest and Sensitivity Analysis the lowest vulnerabilities. Overall, the vulnerability indices may be separated by a maximum value of 51 points. This represents an uncertainty of 2.5 vulnerability classes, because they are 20 points wide. Given this ambiguity, the selection of a weighting technique to integrate a vulnerability index may require additional expertise to be set up satisfactorily. Following a general criterion that weights must be proportional to the range of the ratings, Correspondence Analysis may be recommended as the best adjustment technique. Copyright © 2014 Elsevier B.V. All rights reserved.
Prognostic value of tumor suppressors in osteosarcoma before and after neoadjuvant chemotherapy.
Robl, Bernhard; Pauli, Chantal; Botter, Sander Martijn; Bode-Lesniewska, Beata; Fuchs, Bruno
2015-05-09
Primary bone cancers are among the deadliest cancer types in adolescents, with osteosarcomas being the most prevalent form. Osteosarcomas are commonly treated with multi-drug neoadjuvant chemotherapy and therapy success as well as patient survival is affected by the presence of tumor suppressors. In order to assess the prognostic value of tumor-suppressive biomarkers, primary osteosarcoma tissues were analyzed prior to and after neoadjuvant chemotherapy. We constructed a tissue microarray from high grade osteosarcoma samples, consisting of 48 chemotherapy naïve biopsies (BXs) and 47 tumor resections (RXs) after neoadjuvant chemotherapy. We performed immunohistochemical stainings of P53, P16, maspin, PTEN, BMI1 and Ki67, characterized the subcellular localization and related staining outcome with chemotherapy response and overall survival. Binary logistic regression analysis was used to analyze chemotherapy response and Kaplan-Meier-analysis as well as the Cox proportional hazards model was applied for analysis of patient survival. No significant associations between biomarker expression in BXs and patient survival or chemotherapy response were detected. In univariate analysis, positive immunohistochemistry of P53 (P = 0.008) and P16 (P16; P = 0.033) in RXs was significantly associated with poor survival prognosis. In addition, presence of P16 in RXs was associated with poor survival in multivariate regression analysis (P = 0.003; HR = 0.067) while absence of P16 was associated with good chemotherapy response (P = 0.004; OR = 74.076). Presence of PTEN on tumor RXs was significantly associated with an improved survival prognosis (P = 0.022). Positive immunohistochemistry (IHC) of P16 and P53 in RXs was indicative for poor overall patient survival whereas positive IHC of PTEN was prognostic for good overall patient survival. In addition, we found that P16 might be a marker of osteosarcoma chemotherapy resistance. Therefore, our study supports the use of tumor RXs to assess the prognostic value of biomarkers.
Multicollinearity in Regression Analyses Conducted in Epidemiologic Studies
Vatcheva, Kristina P.; Lee, MinJae; McCormick, Joseph B.; Rahbar, Mohammad H.
2016-01-01
The adverse impact of ignoring multicollinearity on findings and data interpretation in regression analysis is very well documented in the statistical literature. The failure to identify and report multicollinearity could result in misleading interpretations of the results. A review of epidemiological literature in PubMed from January 2004 to December 2013, illustrated the need for a greater attention to identifying and minimizing the effect of multicollinearity in analysis of data from epidemiologic studies. We used simulated datasets and real life data from the Cameron County Hispanic Cohort to demonstrate the adverse effects of multicollinearity in the regression analysis and encourage researchers to consider the diagnostic for multicollinearity as one of the steps in regression analysis. PMID:27274911
Multicollinearity in Regression Analyses Conducted in Epidemiologic Studies.
Vatcheva, Kristina P; Lee, MinJae; McCormick, Joseph B; Rahbar, Mohammad H
2016-04-01
The adverse impact of ignoring multicollinearity on findings and data interpretation in regression analysis is very well documented in the statistical literature. The failure to identify and report multicollinearity could result in misleading interpretations of the results. A review of epidemiological literature in PubMed from January 2004 to December 2013, illustrated the need for a greater attention to identifying and minimizing the effect of multicollinearity in analysis of data from epidemiologic studies. We used simulated datasets and real life data from the Cameron County Hispanic Cohort to demonstrate the adverse effects of multicollinearity in the regression analysis and encourage researchers to consider the diagnostic for multicollinearity as one of the steps in regression analysis.
NASA Astrophysics Data System (ADS)
Kim, Jin-Young; Kwon, Hyun-Han; Kim, Hung-Soo
2015-04-01
The existing regional frequency analysis has disadvantages in that it is difficult to consider geographical characteristics in estimating areal rainfall. In this regard, this study aims to develop a hierarchical Bayesian model based nonstationary regional frequency analysis in that spatial patterns of the design rainfall with geographical information (e.g. latitude, longitude and altitude) are explicitly incorporated. This study assumes that the parameters of Gumbel (or GEV distribution) are a function of geographical characteristics within a general linear regression framework. Posterior distribution of the regression parameters are estimated by Bayesian Markov Chain Monte Carlo (MCMC) method, and the identified functional relationship is used to spatially interpolate the parameters of the distributions by using digital elevation models (DEM) as inputs. The proposed model is applied to derive design rainfalls over the entire Han-river watershed. It was found that the proposed Bayesian regional frequency analysis model showed similar results compared to L-moment based regional frequency analysis. In addition, the model showed an advantage in terms of quantifying uncertainty of the design rainfall and estimating the area rainfall considering geographical information. Finally, comprehensive discussion on design rainfall in the context of nonstationary will be presented. KEYWORDS: Regional frequency analysis, Nonstationary, Spatial information, Bayesian Acknowledgement This research was supported by a grant (14AWMP-B082564-01) from Advanced Water Management Research Program funded by Ministry of Land, Infrastructure and Transport of Korean government.
Stepwise versus Hierarchical Regression: Pros and Cons
ERIC Educational Resources Information Center
Lewis, Mitzi
2007-01-01
Multiple regression is commonly used in social and behavioral data analysis. In multiple regression contexts, researchers are very often interested in determining the "best" predictors in the analysis. This focus may stem from a need to identify those predictors that are supportive of theory. Alternatively, the researcher may simply be interested…
Interpreting Bivariate Regression Coefficients: Going beyond the Average
ERIC Educational Resources Information Center
Halcoussis, Dennis; Phillips, G. Michael
2010-01-01
Statistics, econometrics, investment analysis, and data analysis classes often review the calculation of several types of averages, including the arithmetic mean, geometric mean, harmonic mean, and various weighted averages. This note shows how each of these can be computed using a basic regression framework. By recognizing when a regression model…
Regression Commonality Analysis: A Technique for Quantitative Theory Building
ERIC Educational Resources Information Center
Nimon, Kim; Reio, Thomas G., Jr.
2011-01-01
When it comes to multiple linear regression analysis (MLR), it is common for social and behavioral science researchers to rely predominately on beta weights when evaluating how predictors contribute to a regression model. Presenting an underutilized statistical technique, this article describes how organizational researchers can use commonality…
Precision Efficacy Analysis for Regression.
ERIC Educational Resources Information Center
Brooks, Gordon P.
When multiple linear regression is used to develop a prediction model, sample size must be large enough to ensure stable coefficients. If the derivation sample size is inadequate, the model may not predict well for future subjects. The precision efficacy analysis for regression (PEAR) method uses a cross- validity approach to select sample sizes…
Regression Analysis of Stage Variability for West-Central Florida Lakes
Sacks, Laura A.; Ellison, Donald L.; Swancar, Amy
2008-01-01
The variability in a lake's stage depends upon many factors, including surface-water flows, meteorological conditions, and hydrogeologic characteristics near the lake. An understanding of the factors controlling lake-stage variability for a population of lakes may be helpful to water managers who set regulatory levels for lakes. The goal of this study is to determine whether lake-stage variability can be predicted using multiple linear regression and readily available lake and basin characteristics defined for each lake. Regressions were evaluated for a recent 10-year period (1996-2005) and for a historical 10-year period (1954-63). Ground-water pumping is considered to have affected stage at many of the 98 lakes included in the recent period analysis, and not to have affected stage at the 20 lakes included in the historical period analysis. For the recent period, regression models had coefficients of determination (R2) values ranging from 0.60 to 0.74, and up to five explanatory variables. Standard errors ranged from 21 to 37 percent of the average stage variability. Net leakage was the most important explanatory variable in regressions describing the full range and low range in stage variability for the recent period. The most important explanatory variable in the model predicting the high range in stage variability was the height over median lake stage at which surface-water outflow would occur. Other explanatory variables in final regression models for the recent period included the range in annual rainfall for the period and several variables related to local and regional hydrogeology: (1) ground-water pumping within 1 mile of each lake, (2) the amount of ground-water inflow (by category), (3) the head gradient between the lake and the Upper Floridan aquifer, and (4) the thickness of the intermediate confining unit. Many of the variables in final regression models are related to hydrogeologic characteristics, underscoring the importance of ground-water exchange in controlling the stage of karst lakes in Florida. Regression equations were used to predict lake-stage variability for the recent period for 12 additional lakes, and the median difference between predicted and observed values ranged from 11 to 23 percent. Coefficients of determination for the historical period were considerably lower (maximum R2 of 0.28) than for the recent period. Reasons for these low R2 values are probably related to the small number of lakes (20) with stage data for an equivalent time period that were unaffected by ground-water pumping, the similarity of many of the lake types (large surface-water drainage lakes), and the greater uncertainty in defining historical basin characteristics. The lack of lake-stage data unaffected by ground-water pumping and the poor regression results obtained for that group of lakes limit the ability to predict natural lake-stage variability using this method in west-central Florida.
Athanasopoulos, Leonidas V; Dritsas, Athanasios; Doll, Helen A; Cokkinos, Dennis V
2010-08-01
This study was conducted to explain the variance in quality of life (QoL) and activity capacity of patients with congestive heart failure from pathophysiological changes as estimated by laboratory data. Peak oxygen consumption (peak VO2) and ventilation (VE)/carbon dioxide output (VCO2) slope derived from cardiopulmonary exercise testing, plasma N-terminal prohormone of B-type natriuretic peptide (NT-proBNP), and echocardiographic markers [left atrium (LA), left ventricular ejection fraction (LVEF)] were measured in 62 patients with congestive heart failure, who also completed the Minnesota Living with Heart Failure Questionnaire and the Specific Activity Questionnaire. All regression models were adjusted for age and sex. On linear regression analysis, peak VO2 with P value less than 0.001, VE/VCO2 slope with P value less than 0.01, LVEF with P value less than 0.001, LA with P=0.001, and logNT-proBNP with P value less than 0.01 were found to be associated with QoL. On stepwise multiple linear regression, peak VO2 and LVEF continued to be predictive, accounting for 40% of the variability in Minnesota Living with Heart Failure Questionnaire score. On linear regression analysis, peak VO2 with P value less than 0.001, VE/VCO2 slope with P value less than 0.001, LVEF with P value less than 0.05, LA with P value less than 0.001, and logNT-proBNP with P value less than 0.001 were found to be associated with activity capacity. On stepwise multiple linear regression, peak VO2 and LA continued to be predictive, accounting for 53% of the variability in Specific Activity Questionnaire score. Peak VO2 is independently associated both with QoL and activity capacity. In addition to peak VO2, LVEF is independently associated with QoL, and LA with activity capacity.
Epidemiologic Evaluation of Measurement Data in the Presence of Detection Limits
Lubin, Jay H.; Colt, Joanne S.; Camann, David; Davis, Scott; Cerhan, James R.; Severson, Richard K.; Bernstein, Leslie; Hartge, Patricia
2004-01-01
Quantitative measurements of environmental factors greatly improve the quality of epidemiologic studies but can pose challenges because of the presence of upper or lower detection limits or interfering compounds, which do not allow for precise measured values. We consider the regression of an environmental measurement (dependent variable) on several covariates (independent variables). Various strategies are commonly employed to impute values for interval-measured data, including assignment of one-half the detection limit to nondetected values or of “fill-in” values randomly selected from an appropriate distribution. On the basis of a limited simulation study, we found that the former approach can be biased unless the percentage of measurements below detection limits is small (5–10%). The fill-in approach generally produces unbiased parameter estimates but may produce biased variance estimates and thereby distort inference when 30% or more of the data are below detection limits. Truncated data methods (e.g., Tobit regression) and multiple imputation offer two unbiased approaches for analyzing measurement data with detection limits. If interest resides solely on regression parameters, then Tobit regression can be used. If individualized values for measurements below detection limits are needed for additional analysis, such as relative risk regression or graphical display, then multiple imputation produces unbiased estimates and nominal confidence intervals unless the proportion of missing data is extreme. We illustrate various approaches using measurements of pesticide residues in carpet dust in control subjects from a case–control study of non-Hodgkin lymphoma. PMID:15579415
Crane, Paul K; Gibbons, Laura E; Jolley, Lance; van Belle, Gerald
2006-11-01
We present an ordinal logistic regression model for identification of items with differential item functioning (DIF) and apply this model to a Mini-Mental State Examination (MMSE) dataset. We employ item response theory ability estimation in our models. Three nested ordinal logistic regression models are applied to each item. Model testing begins with examination of the statistical significance of the interaction term between ability and the group indicator, consistent with nonuniform DIF. Then we turn our attention to the coefficient of the ability term in models with and without the group term. If including the group term has a marked effect on that coefficient, we declare that it has uniform DIF. We examined DIF related to language of test administration in addition to self-reported race, Hispanic ethnicity, age, years of education, and sex. We used PARSCALE for IRT analyses and STATA for ordinal logistic regression approaches. We used an iterative technique for adjusting IRT ability estimates on the basis of DIF findings. Five items were found to have DIF related to language. These same items also had DIF related to other covariates. The ordinal logistic regression approach to DIF detection, when combined with IRT ability estimates, provides a reasonable alternative for DIF detection. There appear to be several items with significant DIF related to language of test administration in the MMSE. More attention needs to be paid to the specific criteria used to determine whether an item has DIF, not just the technique used to identify DIF.
Norrie gene product is necessary for regression of hyaloid vessels.
Ohlmann, Anne V; Adamek, Edith; Ohlmann, Andreas; Lütjen-Drecoll, Elke
2004-07-01
To investigate the nature and origin of the vitreous membranes in mice with knock-out of the Norrie gene product (ND mice). Eighty-two eyes of ND mice of different age groups (postnatal day [P]0-13 months) and 95 age-matched wild-type control mice were investigated. In vitreoretinal wholemounts and in sagittal sections, vessels and free cells were visualized by labeling for lectin. In addition, staining with a marker for macrophages (F4/80) and collagen XVIII/endostatin known to be involved in regression of hyaloid vessels was performed for light and electron microscopic investigations. Endostatin expression was confirmed by Western blot analysis. Wild-type controls showed the typical pattern of hyaloid vessels, their regression and concomitantly retinal vasculogenesis and angiogenesis. Hyaloid vessels all stained for endostatin, whereas retinal vessels remained unstained. In ND mice, 1 to 5 days after birth, the hyaloid and retinal vasculatures were comparable to that in control mice. The hyaloid vessels also stained for endostatin. Numerous F4/80-positive cells were present adjacent to the vessels. With increasing age, only a few connecting branches of the hyaloid vessels regressed. Even in old mice most of the hyaloid vessels persisted. The vessels still stained for endostatin. Retinal angiogenesis was impaired. Retrolental membranes in ND mice consist of persistent hyaloid vessels, indicating that the ND gene product is important for the process of regression of these vessels. The ND gene product neither influences endostatin expression nor the presence of macrophages.
A Decomposition of Hospital Profitability: An Application of DuPont Analysis to the US Market.
Turner, Jason; Broom, Kevin; Elliott, Michael; Lee, Jen-Fu
2015-01-01
This paper evaluates the drivers of profitability for a large sample of U.S. hospitals. Following a methodology frequently used by financial analysts, we use a DuPont analysis as a framework to evaluate the quality of earnings. By decomposing returns on equity (ROE) into profit margin, total asset turnover, and capital structure, the DuPont analysis reveals what drives overall profitability. Profit margin, the efficiency with which services are rendered (total asset turnover), and capital structure is calculated for 3,255 U.S. hospitals between 2007 and 2012 using data from the Centers for Medicare & Medicaid Services' Healthcare Cost Report Information System (CMS Form 2552). The sample is then stratified by ownership, size, system affiliation, teaching status, critical access designation, and urban or non-urban location. Those hospital characteristics and interaction terms are then regressed (OLS) against the ROE and the respective DuPont components. Sensitivity to regression methodology is also investigated using a seemingly unrelated regression. When the sample is stratified by hospital characteristics, the results indicate investor-owned hospitals have higher profit margins, higher efficiency, and are substantially more leveraged. Hospitals in systems are found to have higher ROE, margins, and efficiency but are associated with less leverage. In addition, a number of important and significant interactions between teaching status, ownership, location, critical access designation, and inclusion in a system are documented. Many of the significant relationships, most notably not-for-profit ownership, lose significance or are predominately associated with one interaction effect when interaction terms are introduced as explanatory variables. Results are not sensitive to the alternative methodology. The results of the DuPont analysis suggest that although there appears to be convergence in the behavior of NFP and IO hospitals, significant financial differences remain depending on their respective hospital characteristics. Those differences are tempered or exacerbated by location, size, teaching status, system affiliation, and critical access designation. With the exception of cost-based reimbursement for critical access hospitals, emerging payment systems are placing additional financial pressures on hospitals. The financial pressures being applied treat hospitals as a monolithic category and, given the delicate and often negative ROE for many hospitals, the long-term stability of the healthcare facility infrastructure may be negatively impacted.
Regression Model Optimization for the Analysis of Experimental Data
NASA Technical Reports Server (NTRS)
Ulbrich, N.
2009-01-01
A candidate math model search algorithm was developed at Ames Research Center that determines a recommended math model for the multivariate regression analysis of experimental data. The search algorithm is applicable to classical regression analysis problems as well as wind tunnel strain gage balance calibration analysis applications. The algorithm compares the predictive capability of different regression models using the standard deviation of the PRESS residuals of the responses as a search metric. This search metric is minimized during the search. Singular value decomposition is used during the search to reject math models that lead to a singular solution of the regression analysis problem. Two threshold dependent constraints are also applied. The first constraint rejects math models with insignificant terms. The second constraint rejects math models with near-linear dependencies between terms. The math term hierarchy rule may also be applied as an optional constraint during or after the candidate math model search. The final term selection of the recommended math model depends on the regressor and response values of the data set, the user s function class combination choice, the user s constraint selections, and the result of the search metric minimization. A frequently used regression analysis example from the literature is used to illustrate the application of the search algorithm to experimental data.
Lou, Zhengcai; Yang, Jian; Tang, Yongmei; Xiao, Jian
2015-01-01
The use of growth factors to achieve closure of human traumatic tympanic membrane perforations (TMPs) has recently been demonstrated. However, pretreatment factors affecting healing outcomes have seldom been discussed. The objective of this study was to evaluate pretreatment factors contributing to the success or failure of healing of TMPs using fibroblast growth factor-2 (FGF-2). A retrospective cohort study of 99 patients (43 males, 56 females) with traumatic TMPs who were observed for at least 6 months after FGF-2 treatment between March 2011 and December 2012. Eleven factors considered likely to affect the outcome of perforation closure were evaluated statistically using univariate and multivariate logistic regression analysis. Each traumatic TMP was treated by direct application of FGF-2. Complete closure versus failure to close. In total, 99 patients were analyzed. The total closure rate was 92/99 (92.9%) at 6 months; the mean closure time was 10.59 ± 6.81 days. The closure rate did not significantly differ between perforations with or without inverted edges (100.0% vs. 91.4%, p = 0.087), among different size groups (p = 0.768), or among different periods of exposure to injury (p = 0.051). However, the closure rate was significantly different between the high- and low-dose FGF-2 groups (85.0% vs. 98.3%, p = 0.010) and between perforations where the umbo or malleus was or was not involved in perforation (85.4% vs. 98.3%, p = 0.012). Additionally, univariate logistic regression analysis tests showed that it was difficult to achieve healing of these perforations with a history of chronic otitis media or residual TM calcification (p = 0.006), the umbo or malleus was involved in perforation (p = 0.038), and with a high dose of FGF-2 (p = 0.035) compared with control groups. Multivariate logistic regression analysis showed that only a history of chronic otitis media and residual TM calcification and perforation close to the umbo or malleus were associated with non-healing of the TM perforation (p = 0.03 and p = 0.017, respectively) with relative risk factors. Direct application of FGF-2 can be used in all traumatic TMPs, the size of the perforation and inverted edges did not affect the closure rate, and the most beneficial dose was sufficient to keep the residual eardrum environment moist, but without adding liquid. Additionally, multivariate logistic regression analysis revealed that a large perforation was not a major risk factor for nonhealing of TM perforations. However, a history of chronic otitis media, residual TM calcification and involvement of the umbo or malleus in perforation were significant risk factors.
Cao, Qingqing; Wu, Zhenqiang; Sun, Ying; Wang, Tiezhu; Han, Tengwei; Gu, Chaomei; Sun, Yehuan
2011-11-01
To Eexplore the application of negative binomial regression and modified Poisson regression analysis in analyzing the influential factors for injury frequency and the risk factors leading to the increase of injury frequency. 2917 primary and secondary school students were selected from Hefei by cluster random sampling method and surveyed by questionnaire. The data on the count event-based injuries used to fitted modified Poisson regression and negative binomial regression model. The risk factors incurring the increase of unintentional injury frequency for juvenile students was explored, so as to probe the efficiency of these two models in studying the influential factors for injury frequency. The Poisson model existed over-dispersion (P < 0.0001) based on testing by the Lagrangemultiplier. Therefore, the over-dispersion dispersed data using a modified Poisson regression and negative binomial regression model, was fitted better. respectively. Both showed that male gender, younger age, father working outside of the hometown, the level of the guardian being above junior high school and smoking might be the results of higher injury frequencies. On a tendency of clustered frequency data on injury event, both the modified Poisson regression analysis and negative binomial regression analysis can be used. However, based on our data, the modified Poisson regression fitted better and this model could give a more accurate interpretation of relevant factors affecting the frequency of injury.
Deng, Yingyuan; Wang, Tianfu; Chen, Siping; Liu, Weixiang
2017-01-01
The aim of the study is to screen the significant sonographic features by logistic regression analysis and fit a model to diagnose thyroid nodules. A total of 525 pathological thyroid nodules were retrospectively analyzed. All the nodules underwent conventional ultrasonography (US), strain elastosonography (SE), and contrast -enhanced ultrasound (CEUS). Those nodules’ 12 suspicious sonographic features were used to assess thyroid nodules. The significant features of diagnosing thyroid nodules were picked out by logistic regression analysis. All variables that were statistically related to diagnosis of thyroid nodules, at a level of p < 0.05 were embodied in a logistic regression analysis model. The significant features in the logistic regression model of diagnosing thyroid nodules were calcification, suspected cervical lymph node metastasis, hypoenhancement pattern, margin, shape, vascularity, posterior acoustic, echogenicity, and elastography score. According to the results of logistic regression analysis, the formula that could predict whether or not thyroid nodules are malignant was established. The area under the receiver operating curve (ROC) was 0.930 and the sensitivity, specificity, accuracy, positive predictive value, and negative predictive value were 83.77%, 89.56%, 87.05%, 86.04%, and 87.79% respectively. PMID:29228030
Pang, Tiantian; Huang, Leidan; Deng, Yingyuan; Wang, Tianfu; Chen, Siping; Gong, Xuehao; Liu, Weixiang
2017-01-01
The aim of the study is to screen the significant sonographic features by logistic regression analysis and fit a model to diagnose thyroid nodules. A total of 525 pathological thyroid nodules were retrospectively analyzed. All the nodules underwent conventional ultrasonography (US), strain elastosonography (SE), and contrast -enhanced ultrasound (CEUS). Those nodules' 12 suspicious sonographic features were used to assess thyroid nodules. The significant features of diagnosing thyroid nodules were picked out by logistic regression analysis. All variables that were statistically related to diagnosis of thyroid nodules, at a level of p < 0.05 were embodied in a logistic regression analysis model. The significant features in the logistic regression model of diagnosing thyroid nodules were calcification, suspected cervical lymph node metastasis, hypoenhancement pattern, margin, shape, vascularity, posterior acoustic, echogenicity, and elastography score. According to the results of logistic regression analysis, the formula that could predict whether or not thyroid nodules are malignant was established. The area under the receiver operating curve (ROC) was 0.930 and the sensitivity, specificity, accuracy, positive predictive value, and negative predictive value were 83.77%, 89.56%, 87.05%, 86.04%, and 87.79% respectively.
Bonander, Carl; Gustavsson, Johanna; Nilson, Finn
2016-12-01
Fall-related injuries are a global public health problem, especially in elderly populations. The effect of an intervention aimed at reducing the risk of falls in the homes of community-dwelling elderly persons was evaluated. The intervention mainly involves the performance of complicated tasks and hazards assessment by a trained assessor, and has been adopted gradually over the last decade by 191 of 290 Swedish municipalities. A quasi-experimental design was used where intention-to-treat effect estimates were derived using panel regression analysis and a regression discontinuity (RD) design. The outcome measure was the incidence of fall-related hospitalisations in the treatment population, the age of which varied by municipality (≥65 years, ≥67 years, ≥70 years or ≥75 years). We found no statistically significant reductions in injury incidence in the panel regression (IRR 1.01 (95% CI 0.98 to 1.05)) or RD (IRR 1.00 (95% CI 0.97 to 1.03)) analyses. The results are robust to several different model specifications, including segmented panel regression analysis with linear trend change and community fixed effects parameters. It is unclear whether the absence of an effect is due to a low efficacy of the services provided, or a result of low adherence. Additional studies of the effects on other quality-of-life measures are recommended before conclusions are drawn regarding the cost-effectiveness of the provision of home help service programmes. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
Torija, Antonio J; Ruiz, Diego P
2015-02-01
The prediction of environmental noise in urban environments requires the solution of a complex and non-linear problem, since there are complex relationships among the multitude of variables involved in the characterization and modelling of environmental noise and environmental-noise magnitudes. Moreover, the inclusion of the great spatial heterogeneity characteristic of urban environments seems to be essential in order to achieve an accurate environmental-noise prediction in cities. This problem is addressed in this paper, where a procedure based on feature-selection techniques and machine-learning regression methods is proposed and applied to this environmental problem. Three machine-learning regression methods, which are considered very robust in solving non-linear problems, are used to estimate the energy-equivalent sound-pressure level descriptor (LAeq). These three methods are: (i) multilayer perceptron (MLP), (ii) sequential minimal optimisation (SMO), and (iii) Gaussian processes for regression (GPR). In addition, because of the high number of input variables involved in environmental-noise modelling and estimation in urban environments, which make LAeq prediction models quite complex and costly in terms of time and resources for application to real situations, three different techniques are used to approach feature selection or data reduction. The feature-selection techniques used are: (i) correlation-based feature-subset selection (CFS), (ii) wrapper for feature-subset selection (WFS), and the data reduction technique is principal-component analysis (PCA). The subsequent analysis leads to a proposal of different schemes, depending on the needs regarding data collection and accuracy. The use of WFS as the feature-selection technique with the implementation of SMO or GPR as regression algorithm provides the best LAeq estimation (R(2)=0.94 and mean absolute error (MAE)=1.14-1.16 dB(A)). Copyright © 2014 Elsevier B.V. All rights reserved.
Ma, Teng; Lu, Deyi; Zhu, Yin-Sheng; Chu, Xue-Feng; Wang, Yong; Shi, Guo-Ping; Wang, Zheng-Dong; Yu, Li; Jiang, Xiao-Yan; Wang, Xiao-Feng
2018-05-01
To examine the associations of the actinin alpha 3 gene (ACTN3) R577X polymorphism with physical performance and frailty in an older Chinese population. Data from 1,463 individuals (57.8% female) aged 70-87 years from the Rugao Longevity and Ageing Study were used. The associations between R577X and timed 5-m walk, grip strength, timed Up and Go test, and frailty index (FI) based on deficits of 23 laboratory tests (FI-Lab) were examined. Analysis of variance and linear regression models were used to evaluate the genetic effects of ACTN3 R577X on physical performance and FI-Lab. The XX and RX genotypes of the ACTN3 R557X polymorphism accounted for 17.1 and 46.9%, respectively. Multivariate regression analysis revealed that in men aged 70-79 years, the ACTN3 577X allele was significantly associated with physical performance (5-m walk time, regression coefficient (β) = 0.258, P = 0.006; grip strength, β = -1.062, P = 0.012; Up and Go test time β = 0.368, P = 0.019). In women aged 70-79 years, a significant association between the ACTN3 577X allele and the FI-Lab score was observed, with a regression coefficient of β = 0.019 (P = 0.003). These findings suggest an age- and gender-specific X-additive model of R577X for 5-m walk time, grip strength, Up and Go Test time, and FI-Lab score. The ACTN3 577X allele is associated with an age- and sex-specific decrease in physical performance and an increase in frailty in an older population.
Rosswog, Carolina; Schmidt, Rene; Oberthuer, André; Juraeva, Dilafruz; Brors, Benedikt; Engesser, Anne; Kahlert, Yvonne; Volland, Ruth; Bartenhagen, Christoph; Simon, Thorsten; Berthold, Frank; Hero, Barbara; Faldum, Andreas; Fischer, Matthias
2017-12-01
Current risk stratification systems for neuroblastoma patients consider clinical, histopathological, and genetic variables, and additional prognostic markers have been proposed in recent years. We here sought to select highly informative covariates in a multistep strategy based on consecutive Cox regression models, resulting in a risk score that integrates hazard ratios of prognostic variables. A cohort of 695 neuroblastoma patients was divided into a discovery set (n=75) for multigene predictor generation, a training set (n=411) for risk score development, and a validation set (n=209). Relevant prognostic variables were identified by stepwise multivariable L1-penalized least absolute shrinkage and selection operator (LASSO) Cox regression, followed by backward selection in multivariable Cox regression, and then integrated into a novel risk score. The variables stage, age, MYCN status, and two multigene predictors, NB-th24 and NB-th44, were selected as independent prognostic markers by LASSO Cox regression analysis. Following backward selection, only the multigene predictors were retained in the final model. Integration of these classifiers in a risk scoring system distinguished three patient subgroups that differed substantially in their outcome. The scoring system discriminated patients with diverging outcome in the validation cohort (5-year event-free survival, 84.9±3.4 vs 63.6±14.5 vs 31.0±5.4; P<.001), and its prognostic value was validated by multivariable analysis. We here propose a translational strategy for developing risk assessment systems based on hazard ratios of relevant prognostic variables. Our final neuroblastoma risk score comprised two multigene predictors only, supporting the notion that molecular properties of the tumor cells strongly impact clinical courses of neuroblastoma patients. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.
The association between intelligence and lifespan is mostly genetic.
Arden, Rosalind; Luciano, Michelle; Deary, Ian J; Reynolds, Chandra A; Pedersen, Nancy L; Plassman, Brenda L; McGue, Matt; Christensen, Kaare; Visscher, Peter M
2016-02-01
Several studies in the new field of cognitive epidemiology have shown that higher intelligence predicts longer lifespan. This positive correlation might arise from socioeconomic status influencing both intelligence and health; intelligence leading to better health behaviours; and/or some shared genetic factors influencing both intelligence and health. Distinguishing among these hypotheses is crucial for medicine and public health, but can only be accomplished by studying a genetically informative sample. We analysed data from three genetically informative samples containing information on intelligence and mortality: Sample 1, 377 pairs of male veterans from the NAS-NRC US World War II Twin Registry; Sample 2, 246 pairs of twins from the Swedish Twin Registry; and Sample 3, 784 pairs of twins from the Danish Twin Registry. The age at which intelligence was measured differed between the samples. We used three methods of genetic analysis to examine the relationship between intelligence and lifespan: we calculated the proportion of the more intelligent twins who outlived their co-twin; we regressed within-twin-pair lifespan differences on within-twin-pair intelligence differences; and we used the resulting regression coefficients to model the additive genetic covariance. We conducted a meta-analysis of the regression coefficients across the three samples. The combined (and all three individual samples) showed a small positive phenotypic correlation between intelligence and lifespan. In the combined sample observed r = .12 (95% confidence interval .06 to .18). The additive genetic covariance model supported a genetic relationship between intelligence and lifespan. In the combined sample the genetic contribution to the covariance was 95%; in the US study, 84%; in the Swedish study, 86%, and in the Danish study, 85%. The finding of common genetic effects between lifespan and intelligence has important implications for public health, and for those interested in the genetics of intelligence, lifespan or inequalities in health outcomes including lifespan. © The Author 2015; Published by Oxford University Press on behalf of the International Epidemiological Association.
The association between intelligence and lifespan is mostly genetic
Arden, Rosalind; Deary, Ian J; Reynolds, Chandra A; Pedersen, Nancy L; Plassman, Brenda L; McGue, Matt; Christensen, Kaare; Visscher, Peter M
2016-01-01
Abstract Background: Several studies in the new field of cognitive epidemiology have shown that higher intelligence predicts longer lifespan. This positive correlation might arise from socioeconomic status influencing both intelligence and health; intelligence leading to better health behaviours; and/or some shared genetic factors influencing both intelligence and health. Distinguishing among these hypotheses is crucial for medicine and public health, but can only be accomplished by studying a genetically informative sample. Methods: We analysed data from three genetically informative samples containing information on intelligence and mortality: Sample 1, 377 pairs of male veterans from the NAS-NRC US World War II Twin Registry; Sample 2, 246 pairs of twins from the Swedish Twin Registry; and Sample 3, 784 pairs of twins from the Danish Twin Registry. The age at which intelligence was measured differed between the samples. We used three methods of genetic analysis to examine the relationship between intelligence and lifespan: we calculated the proportion of the more intelligent twins who outlived their co-twin; we regressed within-twin-pair lifespan differences on within-twin-pair intelligence differences; and we used the resulting regression coefficients to model the additive genetic covariance. We conducted a meta-analysis of the regression coefficients across the three samples. Results: The combined (and all three individual samples) showed a small positive phenotypic correlation between intelligence and lifespan. In the combined sample observed r = .12 (95% confidence interval .06 to .18). The additive genetic covariance model supported a genetic relationship between intelligence and lifespan. In the combined sample the genetic contribution to the covariance was 95%; in the US study, 84%; in the Swedish study, 86%, and in the Danish study, 85%. Conclusions: The finding of common genetic effects between lifespan and intelligence has important implications for public health, and for those interested in the genetics of intelligence, lifespan or inequalities in health outcomes including lifespan. PMID:26213105
Roso, V M; Schenkel, F S; Miller, S P; Schaeffer, L R
2005-08-01
Breed additive, dominance, and epistatic loss effects are of concern in the genetic evaluation of a multibreed population. Multiple regression equations used for fitting these effects may show a high degree of multicollinearity among predictor variables. Typically, when strong linear relationships exist, the regression coefficients have large SE and are sensitive to changes in the data file and to the addition or deletion of variables in the model. Generalized ridge regression methods were applied to obtain stable estimates of direct and maternal breed additive, dominance, and epistatic loss effects in the presence of multicollinearity among predictor variables. Preweaning weight gains of beef calves in Ontario, Canada, from 1986 to 1999 were analyzed. The genetic model included fixed direct and maternal breed additive, dominance, and epistatic loss effects, fixed environmental effects of age of the calf, contemporary group, and age of the dam x sex of the calf, random additive direct and maternal genetic effects, and random maternal permanent environment effect. The degree and the nature of the multicollinearity were identified and ridge regression methods were used as an alternative to ordinary least squares (LS). Ridge parameters were obtained using two different objective methods: 1) generalized ridge estimator of Hoerl and Kennard (R1); and 2) bootstrap in combination with cross-validation (R2). Both ridge regression methods outperformed the LS estimator with respect to mean squared error of predictions (MSEP) and variance inflation factors (VIF) computed over 100 bootstrap samples. The MSEP of R1 and R2 were similar, and they were 3% less than the MSEP of LS. The average VIF of LS, R1, and R2 were equal to 26.81, 6.10, and 4.18, respectively. Ridge regression methods were particularly effective in decreasing the multicollinearity involving predictor variables of breed additive effects. Because of a high degree of confounding between estimates of maternal dominance and direct epistatic loss effects, it was not possible to compare the relative importance of these effects with a high level of confidence. The inclusion of epistatic loss effects in the additive-dominance model did not cause noticeable reranking of sires, dams, and calves based on across-breed EBV. More precise estimates of breed effects as a result of this study may result in more stable across-breed estimated breeding values over the years.
Laminar Flame Velocity and Temperature Exponent of Diluted DME-Air Mixture
NASA Astrophysics Data System (ADS)
Naseer Mohammed, Abdul; Anwar, Muzammil; Juhany, Khalid A.; Mohammad, Akram
2017-03-01
In this paper, the laminar flame velocity and temperature exponent diluted dimethyl ether (DME) air mixtures are reported. Laminar premixed mixture of DME-air with volumetric dilutions of carbon dioxides (CO2) and nitrogen (N2) are considered. Experiments were conducted using a preheated mesoscale high aspect-ratio diverging channel with inlet dimensions of 25 mm × 2 mm. In this method, flame velocities are extracted from planar flames that were stabilized near adiabatic conditions inside the channel. The flame velocities are then plotted against the ratio of mixture temperature and the initial reference temperature. A non-linear power law regression is observed suitable. This regression analysis gives the laminar flame velocity at the initial reference temperature and temperature exponent. Decrease in the laminar flame velocity and increase in temperature exponent is observed for CO2 and N2 diluted mixtures. The addition of CO2 has profound influence when compared to N2 addition on both flame velocity and temperature exponent. Numerical prediction of the similar mixture using a detailed reaction mechanism is obtained. The computational mechanism predicts higher magnitudes for laminar flame velocity and smaller magnitudes of temperature exponent compared to experimental data.
An Optimization of Inventory Demand Forecasting in University Healthcare Centre
NASA Astrophysics Data System (ADS)
Bon, A. T.; Ng, T. K.
2017-01-01
Healthcare industry becomes an important field for human beings nowadays as it concerns about one’s health. With that, forecasting demand for health services is an important step in managerial decision making for all healthcare organizations. Hence, a case study was conducted in University Health Centre to collect historical demand data of Panadol 650mg for 68 months from January 2009 until August 2014. The aim of the research is to optimize the overall inventory demand through forecasting techniques. Quantitative forecasting or time series forecasting model was used in the case study to forecast future data as a function of past data. Furthermore, the data pattern needs to be identified first before applying the forecasting techniques. Trend is the data pattern and then ten forecasting techniques are applied using Risk Simulator Software. Lastly, the best forecasting techniques will be find out with the least forecasting error. Among the ten forecasting techniques include single moving average, single exponential smoothing, double moving average, double exponential smoothing, regression, Holt-Winter’s additive, Seasonal additive, Holt-Winter’s multiplicative, seasonal multiplicative and Autoregressive Integrated Moving Average (ARIMA). According to the forecasting accuracy measurement, the best forecasting technique is regression analysis.
Salvatore, Stefania; Bramness, Jørgen Gustav; Reid, Malcolm J; Thomas, Kevin Victor; Harman, Christopher; Røislien, Jo
2015-01-01
Wastewater-based epidemiology (WBE) is a new methodology for estimating the drug load in a population. Simple summary statistics and specification tests have typically been used to analyze WBE data, comparing differences between weekday and weekend loads. Such standard statistical methods may, however, overlook important nuanced information in the data. In this study, we apply functional data analysis (FDA) to WBE data and compare the results to those obtained from more traditional summary measures. We analysed temporal WBE data from 42 European cities, using sewage samples collected daily for one week in March 2013. For each city, the main temporal features of two selected drugs were extracted using functional principal component (FPC) analysis, along with simpler measures such as the area under the curve (AUC). The individual cities' scores on each of the temporal FPCs were then used as outcome variables in multiple linear regression analysis with various city and country characteristics as predictors. The results were compared to those of functional analysis of variance (FANOVA). The three first FPCs explained more than 99% of the temporal variation. The first component (FPC1) represented the level of the drug load, while the second and third temporal components represented the level and the timing of a weekend peak. AUC was highly correlated with FPC1, but other temporal characteristic were not captured by the simple summary measures. FANOVA was less flexible than the FPCA-based regression, and even showed concordance results. Geographical location was the main predictor for the general level of the drug load. FDA of WBE data extracts more detailed information about drug load patterns during the week which are not identified by more traditional statistical methods. Results also suggest that regression based on FPC results is a valuable addition to FANOVA for estimating associations between temporal patterns and covariate information.
Bennett, Bradley C; Husby, Chad E
2008-03-28
Botanical pharmacopoeias are non-random subsets of floras, with some taxonomic groups over- or under-represented. Moerman [Moerman, D.E., 1979. Symbols and selectivity: a statistical analysis of Native American medical ethnobotany, Journal of Ethnopharmacology 1, 111-119] introduced linear regression/residual analysis to examine these patterns. However, regression, the commonly-employed analysis, suffers from several statistical flaws. We use contingency table and binomial analyses to examine patterns of Shuar medicinal plant use (from Amazonian Ecuador). We first analyzed the Shuar data using Moerman's approach, modified to better meet requirements of linear regression analysis. Second, we assessed the exact randomization contingency table test for goodness of fit. Third, we developed a binomial model to test for non-random selection of plants in individual families. Modified regression models (which accommodated assumptions of linear regression) reduced R(2) to from 0.59 to 0.38, but did not eliminate all problems associated with regression analyses. Contingency table analyses revealed that the entire flora departs from the null model of equal proportions of medicinal plants in all families. In the binomial analysis, only 10 angiosperm families (of 115) differed significantly from the null model. These 10 families are largely responsible for patterns seen at higher taxonomic levels. Contingency table and binomial analyses offer an easy and statistically valid alternative to the regression approach.
Public Reporting of MRI of the Lumbar Spine for Low Back Pain and Changes in Clinical Documentation.
Flug, Jonathan A; Lind, Kimberly E
2017-12-01
OP-8 is the Medicare imaging efficiency metric for MRI of the lumbar spine for low back pain in the outpatient hospital. We studied trends in exclusion criteria coding over time by site of service after implementation of OP-8 to evaluate provider's response to public reporting. We conducted a secondary data analysis using the Medicare Limited Data Set 5% sample for beneficiaries with MRI lumbar spine and lower back pain during 2009 to 2014. We evaluated the association between excluding condition prevalence and site by using generalized estimating equations regression. We produced model-based estimates of excluding condition prevalence by site and year. As a sensitivity analysis, we repeated the analysis while including additional conditions in the outcome measure. We included 285,911 MRIs of the lumbar spine for low back pain. Generalized estimating equations regression found that outpatient hospitals had a higher proportion of MRIs with at least one excluding condition documented compared with outpatient clinics (P < .05), but increases in excluding condition prevalence were similar across all sites during 2009 to 2014. Our results were not sensitive to the inclusion of additional conditions. Documentation of excluding conditions and other clinically reasonable exclusions for OP-8 increased over time for outpatient hospitals and clinics. Increases in documentation of comorbidities may not translate to actual improvement in imaging appropriateness for low back pain. When accounting for all relevant conditions, the proportion of patients with low back pain considered uncomplicated and being measured by OP-8 would be small, reflecting a small proportion of patients with low back pain. Copyright © 2017 American College of Radiology. Published by Elsevier Inc. All rights reserved.
Mancio, Jennifer; Leal, Cátia; Ferreira, Marta; Norton, Pedro; Lunet, Nuno
2018-04-27
Recent studies suggested that the relation between night-shift work and prostate cancer may differ between rotating and fixed schedules. We aimed to quantify the independent association between night-shift work and prostate cancer, for rotating and fixed schedules. We searched MEDLINE for studies assessing the association of night-shift work, by rotating or fixed schedules, with prostate cancer. We computed summary relative risk (RR) estimates with 95% confidence intervals (95% CI) using the inverse variance method and quantified heterogeneity using the I 2 statistic. Meta-regression analysis was used to compare the summary RR estimates for rotating and fixed schedules, while reducing heterogeneity. A total of nine studies assessed the effect of rotating and, in addition, four of them provided the effect of fixed night-shift work, in relation to daytime workers. Rotating night-shift work was associated with a significantly increased risk of prostate cancer (RR = 1.06, 95% CI of 1.01 to 1.12; I 2 = 50%), but not fixed night-shift work (RR of 1.01, 95% CI of 0.81 to 1.26; I 2 = 33%). In meta-regression model including study design, type of population, and control of confounding, the summary RR was 20% higher for rotating vs. fixed schedule, with heterogeneity fully explained by these variables. This is the first meta-analysis suggesting that an increased risk of prostate cancer may be restricted to workers with rotating night shifts. However, the association was weak and additional studies are needed to further clarify this relation before it can be translated into measures for risk reduction in occupational settings.
Yu, Yan; Xie, Zhilan; Wang, Jihan; Chen, Chu; Du, Shuli; Chen, Peng; Li, Bin; Jin, Tianbo; Zhao, Heping
2016-12-01
The proportion of alcohol-induced osteonecrosis of the femoral head (ONFH) in all ONFH patients was 30.7%, with males prevailing among the ONFH patients in mainland China (70.1%). Matrix metalloproteinase 2 (MMP2), a member of the MMP gene family, encodes the enzyme MMP2, which can promote osteoclast migration, attachment, and bone matrix degradation. In this case-control study, we aimed to investigate the association between MMP2 and the alcohol-induced ONFH in Chinese males.In total, 299 patients with alcohol-induced ONFH and 396 healthy controls were recruited for a case-control association study. Five single-nucleotide polymorphisms within the MMP2 locus were genotyped and examined for their correlation with the risk of alcohol-induced ONFH and treatment response using Pearson χ test and unconditional logistic regression analysis. We identified 3 risk alleles for carriers: the allele "T" of rs243849 increased the risk of alcohol-induced ONFH in the allele model, the log-additive model without adjustment, and the log-additive model with adjustment for age. Conversely, the genotypes "CC" in rs7201 and "CC" in rs243832 decreased the risk of alcohol-induced ONFH, as revealed by the recessive model. After the Bonferroni multiple adjustment, no significant association was found. Furthermore, the haplotype analysis showed that the "TT" haplotype of MMP2 was more frequent among patients with alcohol-induced ONFH by unconditional logistic regression analysis adjusted for age.In conclusion, there may be an association between MMP2 and the risk of alcohol-induced ONFH in North-Chinese males. However, studies on larger populations are needed to confirm this hypothesis; these data may provide a theoretical foundation for future studies.
The Relationship between Social Capital in Hospitals and Physician Job Satisfaction
Ommen, Oliver; Driller, Elke; Köhler, Thorsten; Kowalski, Christoph; Ernstmann, Nicole; Neumann, Melanie; Steffen, Petra; Pfaff, Holger
2009-01-01
Background Job satisfaction in the hospital is an important predictor for many significant management ratios. Acceptance in professional life or high workload are known as important predictors for job satisfaction. The influence of social capital in hospitals on job satisfaction within the health care system, however, remains to be determined. Thus, this article aimed at analysing the relationship between overall job satisfaction of physicians and social capital in hospitals. Methods The results of this study are based upon questionnaires sent by mail to 454 physicians working in the field of patient care in 4 different German hospitals in 2002. 277 clinicians responded to the poll, for a response rate of 61%. Analysis was performed using three linear regression models with physician overall job satisfaction as the dependent variable and age, gender, professional experience, workload, and social capital as independent variables. Results The first regression model explained nearly 9% of the variance of job satisfaction. Whereas job satisfaction increased slightly with age, gender and professional experience were not identified as significant factors to explain the variance. Setting up a second model with the addition of subjectively-perceived workload to the analysis, the explained variance increased to 18% and job satisfaction decreased significantly with increasing workload. The third model including social capital in hospital explained 36% of the variance with social capital, professional experience and workload as significant factors. Conclusion This analysis demonstrated that the social capital of an organisation, in addition to professional experience and workload, represents a significant predictor of overall job satisfaction of physicians working in the field of patient care. Trust, mutual understanding, shared aims, and ethical values are qualities of social capital that unify members of social networks and communities and enable them to act cooperatively. PMID:19445692
Exploring public databases to characterize urban flood risks in Amsterdam
NASA Astrophysics Data System (ADS)
Gaitan, Santiago; ten Veldhuis, Marie-claire; van de Giesen, Nick
2015-04-01
Cities worldwide are challenged by increasing urban flood risks. Precise and realistic measures are required to decide upon investment to reduce their impacts. Obvious flooding factors affecting flood risk include sewer systems performance and urban topography. However, currently implemented sewer and topographic models do not provide realistic predictions of local flooding occurrence during heavy rain events. Assessing other factors such as spatially distributed rainfall and socioeconomic characteristics may help to explain probability and impacts of urban flooding. Several public databases were analyzed: complaints about flooding made by citizens, rainfall depths (15 min and 100 Ha spatio-temporal resolution), grids describing number of inhabitants, income, and housing price (1Ha and 25Ha resolution); and buildings age. Data analysis was done using Python and GIS programming, and included spatial indexing of data, cluster analysis, and multivariate regression on the complaints. Complaints were used as a proxy to characterize flooding impacts. The cluster analysis, run for all the variables except the complaints, grouped part of the grid-cells of central Amsterdam into a highly differentiated group, covering 10% of the analyzed area, and accounting for 25% of registered complaints. The configuration of the analyzed variables in central Amsterdam coincides with a high complaint count. Remaining complaints were evenly dispersed along other groups. An adjusted R2 of 0.38 in the multivariate regression suggests that explaining power can improve if additional variables are considered. While rainfall intensity explained 4% of the incidence of complaints, population density and building age significantly explained around 20% each. Data mining of public databases proved to be a valuable tool to identify factors explaining variability in occurrence of urban pluvial flooding, though additional variables must be considered to fully explain flood risk variability.
Jha, Dilip Kumar; Vinithkumar, Nambali Valsalan; Sahu, Biraja Kumar; Dheenan, Palaiya Sukumaran; Das, Apurba Kumar; Begum, Mehmuna; Devi, Marimuthu Prashanthi; Kirubagaran, Ramalingam
2015-07-15
Chidiyatappu Bay is one of the least disturbed marine environments of Andaman & Nicobar Islands, the union territory of India. Oceanic flushing from southeast and northwest direction is prevalent in this bay. Further, anthropogenic activity is minimal in the adjoining environment. Considering the pristine nature of this bay, seawater samples collected from 12 sampling stations covering three seasons were analyzed. Principal Component Analysis (PCA) revealed 69.9% of total variance and exhibited strong factor loading for nitrite, chlorophyll a and phaeophytin. In addition, analysis of variance (ANOVA-one way), regression analysis, box-whisker plots and Geographical Information System based hot spot analysis further simplified and supported multivariate results. The results obtained are important to establish reference conditions for comparative study with other similar ecosystems in the region. Copyright © 2015 Elsevier Ltd. All rights reserved.
The Precision Efficacy Analysis for Regression Sample Size Method.
ERIC Educational Resources Information Center
Brooks, Gordon P.; Barcikowski, Robert S.
The general purpose of this study was to examine the efficiency of the Precision Efficacy Analysis for Regression (PEAR) method for choosing appropriate sample sizes in regression studies used for precision. The PEAR method, which is based on the algebraic manipulation of an accepted cross-validity formula, essentially uses an effect size to…
Effect of Contact Damage on the Strength of Ceramic Materials.
1982-10-01
variables that are important to erosion, and a multivariate , linear regression analysis is used to fit the data to the dimensional analysis. The...of Equations 7 and 8 by a multivariable regression analysis (room tem- perature data) Exponent Regression Standard error Computed coefficient of...1980) 593. WEAVER, Proc. Brit. Ceram. Soc. 22 (1973) 125. 39. P. W. BRIDGMAN, "Dimensional Analaysis ", (Yale 18. R. W. RICE, S. W. FREIMAN and P. F
Parametric Methods for Dynamic 11C-Phenytoin PET Studies.
Mansor, Syahir; Yaqub, Maqsood; Boellaard, Ronald; Froklage, Femke E; de Vries, Anke; Bakker, Esther D M; Voskuyl, Rob A; Eriksson, Jonas; Schwarte, Lothar A; Verbeek, Joost; Windhorst, Albert D; Lammertsma, Adriaan A
2017-03-01
In this study, the performance of various methods for generating quantitative parametric images of dynamic 11 C-phenytoin PET studies was evaluated. Methods: Double-baseline 60-min dynamic 11 C-phenytoin PET studies, including online arterial sampling, were acquired for 6 healthy subjects. Parametric images were generated using Logan plot analysis, a basis function method, and spectral analysis. Parametric distribution volume (V T ) and influx rate ( K 1 ) were compared with those obtained from nonlinear regression analysis of time-activity curves. In addition, global and regional test-retest (TRT) variability was determined for parametric K 1 and V T values. Results: Biases in V T observed with all parametric methods were less than 5%. For K 1 , spectral analysis showed a negative bias of 16%. The mean TRT variabilities of V T and K 1 were less than 10% for all methods. Shortening the scan duration to 45 min provided similar V T and K 1 with comparable TRT performance compared with 60-min data. Conclusion: Among the various parametric methods tested, the basis function method provided parametric V T and K 1 values with the least bias compared with nonlinear regression data and showed TRT variabilities lower than 5%, also for smaller volume-of-interest sizes (i.e., higher noise levels) and shorter scan duration. © 2017 by the Society of Nuclear Medicine and Molecular Imaging.
Psychosocial factors influencing smokeless tobacco use by teen-age military dependents.
Lee, S; Raker, T; Chisick, M C
1994-02-01
Using bivariate and logistic regression analysis, we explored psychosocial correlates of smokeless tobacco (SLT) use in a sample of 2,257 teenage military dependents. We built separate regression models for males and females to explain triers and users of SLT. Results show female and male triers share five factors regarding SLT use--parental and peer approval, trying smoking, relatives using SLT, and athletic team membership. Male trial of SLT was additionally associated with race, difficulty in purchasing SLT, relatives who smoke, current smoking, and belief that SLT can cause mouth cancer. Male use of SLT was associated with race, seeing a dentist regularly, SLT counseling by a dentist, parental approval, trying and current smoking, and grade level. In all models, trying smoking was the strongest explanatory variable. Relatives and peers exert considerable influence on SLT use. Few triers or users had received SLT counseling from their dentist despite high dental utilization rates.
Analysis of selective chopper radiometer data
NASA Technical Reports Server (NTRS)
Roe, J.; Hovland, D.; Wilcox, R.
1983-01-01
Data from SCR-B on Nimbus 5 have been processed to yield global, orbital temperatures at 10, 5, 2, 1, and 0.4 mb for the period January 1977 through April 1978 under the current task. In addition gridded values at 10 deg latitude by 20 deg longitude were prepared by space-time interpolation for the period January 1975 through April 1978. Temperature retrieval was based on regression of radiances against Meteorological Rocket Network data, with regressions recomputed at approximately six-month intervals. This data now completes a consistent time series from April 1970 to April 1978 for all available radiance data from SCR A and SCR B on Nimbus 4 and 5. The processing details for the current period are discussed, but is also applicable to the previous data periods. The accuracy of the temperature retrievals for each 6-month period for the entire eight years is given in the Appendices.
Enterprise systems in financial sector - an application in precious metal trading forecasting
NASA Astrophysics Data System (ADS)
Chen, Xiaozhu; Fang, Yiwei
2013-11-01
The use of enterprise systems has become increasingly popular in the financial service industry. This paper discusses the applications of enterprise systems in the financial sectors and presents an application in gold price forecasting. We carefully examine the impacts of a few most widely assumed factors that have significant impact on the long-term gold price using statistical regression techniques. The analysis on our proposed linear regression mode indicates that the United States ultra scale of M2 money supply has been the most important catalyst for the rising price of gold, and the CRB index upward trend has also been the weighty factor for pushing up the gold price. In addition, the gold price has a low negative correlation with the Dow Jones Industrial Average, and low positive correlations with the US dollar index and the gold ETFs holdings.
Casemix funding for a specialist paediatrics hospital: a hedonic regression approach.
Bridges, J F; Hanson, R M
2000-01-01
This paper inquires into the effects that Diagnosis Related Groups (DRGs) have had on the ability to explain patient-level costs in a specialist paediatrics hospital. Two hedonic models are estimated using 1996/97 New Children's Hospital (NCH) patient level cost data, one with and one without a casemix index (CMI). The results show that the inclusion of a casemix index as an explanatory variable leads to a better accounting of cost. The full hedonic model is then used to simulate a funding model for the 1997/98 NCH cost data. These costs are highly correlated with the actual costs reported for that year. In addition, univariate regression indicates that there has been inflation in costs in the order of 4.8% between the two years. In conclusion, hedonic analysis can provide valuable evidence for the design of funding models that account for casemix.
Using exogenous variables in testing for monotonic trends in hydrologic time series
Alley, William M.
1988-01-01
One approach that has been used in performing a nonparametric test for monotonic trend in a hydrologic time series consists of a two-stage analysis. First, a regression equation is estimated for the variable being tested as a function of an exogenous variable. A nonparametric trend test such as the Kendall test is then performed on the residuals from the equation. By analogy to stagewise regression and through Monte Carlo experiments, it is demonstrated that this approach will tend to underestimate the magnitude of the trend and to result in some loss in power as a result of ignoring the interaction between the exogenous variable and time. An alternative approach, referred to as the adjusted variable Kendall test, is demonstrated to generally have increased statistical power and to provide more reliable estimates of the trend slope. In addition, the utility of including an exogenous variable in a trend test is examined under selected conditions.
Effects of export concentration on CO2 emissions in developed countries: an empirical analysis.
Apergis, Nicholas; Can, Muhlis; Gozgor, Giray; Lau, Chi Keung Marco
2018-03-08
This paper provides the evidence on the short- and the long-run effects of the export product concentration on the level of CO 2 emissions in 19 developed (high-income) economies, spanning the period 1962-2010. To this end, the paper makes use of the nonlinear panel unit root and cointegration tests with multiple endogenous structural breaks. It also considers the mean group estimations, the autoregressive distributed lag model, and the panel quantile regression estimations. The findings illustrate that the environmental Kuznets curve (EKC) hypothesis is valid in the panel dataset of 19 developed economies. In addition, it documents that a higher level of the product concentration of exports leads to lower CO 2 emissions. The results from the panel quantile regressions also indicate that the effect of the export product concentration upon the per capita CO 2 emissions is relatively high at the higher quantiles.
Use of ocean color scanner data in water quality mapping
NASA Technical Reports Server (NTRS)
Khorram, S.
1981-01-01
Remotely sensed data, in combination with in situ data, are used in assessing water quality parameters within the San Francisco Bay-Delta. The parameters include suspended solids, chlorophyll, and turbidity. Regression models are developed between each of the water quality parameter measurements and the Ocean Color Scanner (OCS) data. The models are then extended to the entire study area for mapping water quality parameters. The results include a series of color-coded maps, each pertaining to one of the water quality parameters, and the statistical analysis of the OCS data and regression models. It is found that concurrently collected OCS data and surface truth measurements are highly useful in mapping the selected water quality parameters and locating areas having relatively high biological activity. In addition, it is found to be virtually impossible, at least within this test site, to locate such areas on U-2 color and color-infrared photography.
Rousseau, Vincent; Aubé, Caroline
2010-01-01
This study investigated whether both supervisor and coworker support may be positively related to affective commitment to the organization on one hand; and on the other hand, it examined the moderating effect of job resource adequacy and ambient conditions on these relationships. The sample included 215 participants working within a health care organization. Results of regression analysis showed that supervisor and coworker support have an additive effect on affective commitment. Hierarchical regression analyses indicated that supervisor and coworker support are more strongly related to affective commitment when job resource adequacy is high. Furthermore, ambient conditions moderate the relationship between supervisor support and affective commitment in such a way that favorable ambient conditions strengthen this relationship. Overall, these findings reinforce the importance of taking into account contingent factors in the study of antecedents of affective commitment to the organization.
Influence of prolonged static stretching on motor unit firing properties.
Ye, Xin; Beck, Travis W; Wages, Nathan P
2016-05-01
The purpose of this study was to examine the influence of a stretching intervention on motor control strategy of the biceps brachii muscle. Ten men performed twelve 100-s passive static stretches of the biceps brachii. Before and after the intervention, isometric strength was tested during maximal voluntary contractions (MVCs) of the elbow flexors. Subjects also performed trapezoid isometric contractions at 30% and 70% of MVC. Surface electromyographic signals from the submaximal contractions were decomposed into individual motor unit action potential trains. Linear regression analysis was used to examine the relationship between motor unit mean firing rate and recruitment threshold. The stretching intervention caused significant decreases in y-intercepts of the linear regression lines. In addition, linear slopes at both intensities remained unchanged. Despite reduced motor unit firing rates following the stretches, the motor control scheme remained unchanged. © 2016 Wiley Periodicals, Inc.
Poor methodological quality and reporting standards of systematic reviews in burn care management.
Wasiak, Jason; Tyack, Zephanie; Ware, Robert; Goodwin, Nicholas; Faggion, Clovis M
2017-10-01
The methodological and reporting quality of burn-specific systematic reviews has not been established. The aim of this study was to evaluate the methodological quality of systematic reviews in burn care management. Computerised searches were performed in Ovid MEDLINE, Ovid EMBASE and The Cochrane Library through to February 2016 for systematic reviews relevant to burn care using medical subject and free-text terms such as 'burn', 'systematic review' or 'meta-analysis'. Additional studies were identified by hand-searching five discipline-specific journals. Two authors independently screened papers, extracted and evaluated methodological quality using the 11-item A Measurement Tool to Assess Systematic Reviews (AMSTAR) tool and reporting quality using the 27-item Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) checklist. Characteristics of systematic reviews associated with methodological and reporting quality were identified. Descriptive statistics and linear regression identified features associated with improved methodological quality. A total of 60 systematic reviews met the inclusion criteria. Six of the 11 AMSTAR items reporting on 'a priori' design, duplicate study selection, grey literature, included/excluded studies, publication bias and conflict of interest were reported in less than 50% of the systematic reviews. Of the 27 items listed for PRISMA, 13 items reporting on introduction, methods, results and the discussion were addressed in less than 50% of systematic reviews. Multivariable analyses showed that systematic reviews associated with higher methodological or reporting quality incorporated a meta-analysis (AMSTAR regression coefficient 2.1; 95% CI: 1.1, 3.1; PRISMA regression coefficient 6·3; 95% CI: 3·8, 8·7) were published in the Cochrane library (AMSTAR regression coefficient 2·9; 95% CI: 1·6, 4·2; PRISMA regression coefficient 6·1; 95% CI: 3·1, 9·2) and included a randomised control trial (AMSTAR regression coefficient 1·4; 95%CI: 0·4, 2·4; PRISMA regression coefficient 3·4; 95% CI: 0·9, 5·8). The methodological and reporting quality of systematic reviews in burn care requires further improvement with stricter adherence by authors to the PRISMA checklist and AMSTAR tool. © 2016 Medicalhelplines.com Inc and John Wiley & Sons Ltd.
Common pitfalls in statistical analysis: Linear regression analysis
Aggarwal, Rakesh; Ranganathan, Priya
2017-01-01
In a previous article in this series, we explained correlation analysis which describes the strength of relationship between two continuous variables. In this article, we deal with linear regression analysis which predicts the value of one continuous variable from another. We also discuss the assumptions and pitfalls associated with this analysis. PMID:28447022
NASA Astrophysics Data System (ADS)
Gürcan, Eser Kemal
2017-04-01
The most commonly used methods for analyzing time-dependent data are multivariate analysis of variance (MANOVA) and nonlinear regression models. The aim of this study was to compare some MANOVA techniques and nonlinear mixed modeling approach for investigation of growth differentiation in female and male Japanese quail. Weekly individual body weight data of 352 male and 335 female quail from hatch to 8 weeks of age were used to perform analyses. It is possible to say that when all the analyses are evaluated, the nonlinear mixed modeling is superior to the other techniques because it also reveals the individual variation. In addition, the profile analysis also provides important information.
Group Additivity Determination for Oxygenates, Oxonium Ions, and Oxygen-Containing Carbenium Ions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dellon, Lauren D.; Sung, Chun-Yi; Robichaud, David J.
Bio-oil produced from biomass fast pyrolysis often requires catalytic upgrading to remove oxygen and acidic species over zeolite catalysts. The elementary reactions in the mechanism for this process involve carbenium and oxonium ions. In order to develop a detailed kinetic model for the catalytic upgrading of biomass, rate constants are required for these elementary reactions. The parameters in the Arrhenius equation can be related to thermodynamic properties through structure-reactivity relationships, such as the Evans-Polanyi relationship. For this relationship, enthalpies of formation of each species are required, which can be reasonably estimated using group additivity. However, the literature previously lacked groupmore » additivity values for oxygenates, oxonium ions, and oxygen-containing carbenium ions. In this work, 71 group additivity values for these types of groups were regressed, 65 of which had not been reported previously and six of which were newly estimated based on regression in the context of the 65 new groups. Heats of formation based on atomization enthalpy calculations for a set of reference molecules and isodesmic reactions for a small set of larger species for which experimental data was available were used to demonstrate the accuracy of the Gaussian-4 quantum mechanical method in estimating enthalpies of formation for species involving the moieties of interest. Isodesmic reactions for a total of 195 species were constructed from the reference molecules to calculate enthalpies of formation that were used to regress the group additivity values. The results showed an average deviation of 1.95 kcal/mol between the values calculated from Gaussian-4 and isodesmic reactions versus those calculated from the group additivity values that were newly regressed. Importantly, the new groups enhance the database for group additivity values, especially those involving oxonium ions.« less
Sun, Xian-Jun; Li, Yan-Liang; Wang, Long-Gang; Liu, Li-Qing; Ma, Heng; Hou, Wen-Hong; Yu, Jin-Ming
2017-12-01
Microtubule-associated serine/threonine kinase like (Mastl) is deregulated in a number of types of human malignancy and may be a kinase target for cancer treatment. The aim of the present study was to determine the Mastl expression in gastric cancer and to clarify its clinical and prognostic significance. Immunohistochemistry was performed on a cohort of 126 postoperative gastric cancer samples to detect the expression of Mastl and two epithelial to mesenchymal transition (EMT) markers, epithelial-cadherin and Vimentin. The χ 2 test, Kaplan-Meier estimator analysis and Cox's regression model were used to analyze the data. Upregulated Mastl protein expression was observed in the gastric cancer tissues compared with that in the adjacent non-cancerous gastric tissues. Increased Mastl expression was identified in 54/126 (42.9%) gastric cancer samples, and was significantly associated with lymph node metastasis, tumor relapse, EMT status and poor overall survival. Additional analysis demonstrated that the Mastl expression level stratified the patient outcome in stage III, but not stage II tumor subgroups. Cox's regression analysis revealed that increased Mastl expression was an independent prognostic factor for patients with gastric cancer. Mastl expression may be a valuable prognostic marker and a potential target for patients with gastric cancer.
Lim, Seokchun; Jordan, Sumanas W; Jain, Umang; Kim, John Y S
2014-08-01
Studies that evaluate the predictors and causes of unplanned re-operation in outpatient plastic surgery. This study retrospectively reviewed the American College of Surgeons National Surgical Quality Improvement Program (ACS-NSQIP) and identified all plastic surgery outpatient cases performed in 2011. Multiple logistic regression analysis was utilised to identify independent risk factors and causes of unplanned reoperations. Of the 6749 outpatient plastic surgery cases identified, there were 125 (1.9%) unplanned re-operations (UR). Regression analysis demonstrated that body mass index (BMI, OR = 1.041, 95% CI = 1.019-1.065), preoperative open wound/wound infection (OR = 3.498, 95% CI = 1.593-7.678), American Society of Anesthesiologists (ASA) class 3 (OR = 2.235, 95% CI = 1.048-4.765), and total work relative value units (RVU, OR = 1.014, 95% CI = 1.005-1.024) were significantly predictive of UR. Additionally, the presence of any complication was significantly associated with UR (OR = 15.065, 95% CI = 5.705-39.781). In an era of outcomes-driven medicine, unplanned re-operation is a critical quality indicator for ambulatory plastic surgery facilities. The identified risk factors will aid in surgical planning and risk adjustment.
Searching for a neurologic injury's Wechsler Adult Intelligence Scale-Third Edition profile.
Gonçalves, Marta A; Moura, Octávio; Castro-Caldas, Alexandre; Simões, Mário R
2017-01-01
This study aimed to investigate the presence of a Wechsler Adult Intelligence Scale-Third Edition (WAIS-III) cognitive profile in a Portuguese neurologic injured sample. The Portuguese WAIS-III was administered to 81 mixed neurologic patients and 81 healthy matched controls selected from the Portuguese standardization sample. Although the mixed neurologic injury group performed significantly lower than the healthy controls for the majority of the WAIS-III scores (i.e., composite measures, discrepancies, and subtests), the mean scores were within the normal range and, therefore, at risk of being unobserved in a clinical evaluation. ROC curves analysis showed poor to acceptable diagnostic accuracy for the WAIS-III composite measures and subtests (Working Memory Index and Digit Span revealed the highest accuracy for discriminating between participants, respectively). Multiple regression analysis showed that both literacy and the presence of brain injury were significant predictors for all of the composite measures. In addition, multiple regression analysis also showed that literacy, age of injury onset, and years of survival predicted all seven composite measures for the mixed neurologic injured group. Despite the failure to find a WAIS-III cognitive profile for mixed neurologic patients, the results showed a significant influence of brain lesion and literacy in the performance of the WAIS-III.
The impact of gun control (Bill C-51) on suicide in Canada.
Leenaars, Antoon A; Moksony, Ferenc; Lester, David; Wenckstern, Susanne
2003-01-01
Suicide is a multiply determined behavior, calling for diverse prevention efforts. Gun control has been proposed as an important component of society's response, and an opportunity for studying the effects of legislative gun control laws on suicide rates was provided by Canada's Criminal Law Amendment Act of 1977 (Bill C-51). This article reviews previous studies of the impact of this act on the total population of Canada and subpopulations by age and gender and, in addition, presents the results of 2 new studies: a different method of analysis, an interrupted time-series analysis, and the results of a multiple regression analysis that controls for some social variables. It appears that Bill C-51 may have had an impact on suicide rates, even after controls for social variables.
Quality of life in breast cancer patients--a quantile regression analysis.
Pourhoseingholi, Mohamad Amin; Safaee, Azadeh; Moghimi-Dehkordi, Bijan; Zeighami, Bahram; Faghihzadeh, Soghrat; Tabatabaee, Hamid Reza; Pourhoseingholi, Asma
2008-01-01
Quality of life study has an important role in health care especially in chronic diseases, in clinical judgment and in medical resources supplying. Statistical tools like linear regression are widely used to assess the predictors of quality of life. But when the response is not normal the results are misleading. The aim of this study is to determine the predictors of quality of life in breast cancer patients, using quantile regression model and compare to linear regression. A cross-sectional study conducted on 119 breast cancer patients that admitted and treated in chemotherapy ward of Namazi hospital in Shiraz. We used QLQ-C30 questionnaire to assessment quality of life in these patients. A quantile regression was employed to assess the assocciated factors and the results were compared to linear regression. All analysis carried out using SAS. The mean score for the global health status for breast cancer patients was 64.92+/-11.42. Linear regression showed that only grade of tumor, occupational status, menopausal status, financial difficulties and dyspnea were statistically significant. In spite of linear regression, financial difficulties were not significant in quantile regression analysis and dyspnea was only significant for first quartile. Also emotion functioning and duration of disease statistically predicted the QOL score in the third quartile. The results have demonstrated that using quantile regression leads to better interpretation and richer inference about predictors of the breast cancer patient quality of life.
Ahearn, Elizabeth A.
2004-01-01
Multiple linear-regression equations were developed to estimate the magnitudes of floods in Connecticut for recurrence intervals ranging from 2 to 500 years. The equations can be used for nonurban, unregulated stream sites in Connecticut with drainage areas ranging from about 2 to 715 square miles. Flood-frequency data and hydrologic characteristics from 70 streamflow-gaging stations and the upstream drainage basins were used to develop the equations. The hydrologic characteristics?drainage area, mean basin elevation, and 24-hour rainfall?are used in the equations to estimate the magnitude of floods. Average standard errors of prediction for the equations are 31.8, 32.7, 34.4, 35.9, 37.6 and 45.0 percent for the 2-, 10-, 25-, 50-, 100-, and 500-year recurrence intervals, respectively. Simplified equations using only one hydrologic characteristic?drainage area?also were developed. The regression analysis is based on generalized least-squares regression techniques. Observed flows (log-Pearson Type III analysis of the annual maximum flows) from five streamflow-gaging stations in urban basins in Connecticut were compared to flows estimated from national three-parameter and seven-parameter urban regression equations. The comparison shows that the three- and seven- parameter equations used in conjunction with the new statewide equations generally provide reasonable estimates of flood flows for urban sites in Connecticut, although a national urban flood-frequency study indicated that the three-parameter equations significantly underestimated flood flows in many regions of the country. Verification of the accuracy of the three-parameter or seven-parameter national regression equations using new data from Connecticut stations was beyond the scope of this study. A technique for calculating flood flows at streamflow-gaging stations using a weighted average also is described. Two estimates of flood flows?one estimate based on the log-Pearson Type III analyses of the annual maximum flows at the gaging station, and the other estimate from the regression equation?are weighted together based on the years of record at the gaging station and the equivalent years of record value determined from the regression. Weighted averages of flood flows for the 2-, 10-, 25-, 50-, 100-, and 500-year recurrence intervals are tabulated for the 70 streamflow-gaging stations used in the regression analysis. Generally, weighted averages give the most accurate estimate of flood flows at gaging stations. An evaluation of the Connecticut's streamflow-gaging network was performed to determine whether the spatial coverage and range of geographic and hydrologic conditions are adequately represented for transferring flood characteristics from gaged to ungaged sites. Fifty-one of 54 stations in the current (2004) network support one or more flood needs of federal, state, and local agencies. Twenty-five of 54 stations in the current network are considered high-priority stations by the U.S. Geological Survey because of their contribution to the longterm understanding of floods, and their application for regionalflood analysis. Enhancements to the network to improve overall effectiveness for regionalization can be made by increasing the spatial coverage of gaging stations, establishing stations in regions of the state that are not well-represented, and adding stations in basins with drainage area sizes not represented. Additionally, the usefulness of the network for characterizing floods can be maintained and improved by continuing operation at the current stations because flood flows can be more accurately estimated at stations with continuous, long-term record.
Martin, R C; Sawrie, S M; Roth, D L; Gilliam, F G; Faught, E; Morawetz, R B; Kuzniecky, R
1998-10-01
To characterize patterns of base rate change on measures of verbal and visual memory after anterior temporal lobectomy (ATL) using a newly developed regression-based outcome methodology that accounts for effects of practice and regression towards the mean, and to comment on the predictive utility of baseline memory measures on postoperative memory outcome. Memory change was operationalized using regression-based change norms in a group of left (n = 53) and right (n = 48) ATL patients. All patients were administered tests of episodic verbal (prose recall, list learning) and visual (figure reproduction) memory, and semantic memory before and after ATL. ATL patients displayed a wide range of memory outcome across verbal and visual memory domains. Significant performance declines were noted for 25-50% of left ATL patients on verbal semantic and episodic memory tasks, while one-third of right ATL patients displayed significant declines in immediate and delayed episodic prose recall. Significant performance improvement was noted in an additional one-third of right ATL patients on delayed prose recall. Base rate change was similar between the two ATL groups across immediate and delayed visual memory. Approximately one-fourth of all patients displayed clinically meaningful losses on the visual memory task following surgery. Robust relationships between preoperative memory measures and nonstandardized change scores were attenuated or reversed using standardized memory outcome techniques. Our results demonstrated substantial group variability in memory outcome for ATL patients. These results extend previous research by incorporating known effects of practice and regression to the mean when addressing meaningful neuropsychological change following epilepsy surgery. Our findings also suggest that future neuropsychological outcome studies should take steps towards controlling for regression-to-the-mean before drawing predictive conclusions.
Kohn, Michael S.; Stevens, Michael R.; Harden, Tessa M.; Godaire, Jeanne E.; Klinger, Ralph E.; Mommandi, Amanullah
2016-09-09
The U.S. Geological Survey (USGS), in cooperation with the Colorado Department of Transportation, developed regional-regression equations for estimating the 50-, 20-, 10-, 4-, 2-, 1-, 0.5-, 0.2-percent annual exceedance-probability discharge (AEPD) for natural streamflow in eastern Colorado. A total of 188 streamgages, consisting of 6,536 years of record and a mean of approximately 35 years of record per streamgage, were used to develop the peak-streamflow regional-regression equations. The estimated AEPDs for each streamgage were computed using the USGS software program PeakFQ. The AEPDs were determined using systematic data through water year 2013. Based on previous studies conducted in Colorado and neighboring States and on the availability of data, 72 characteristics (57 basin and 15 climatic characteristics) were evaluated as candidate explanatory variables in the regression analysis. Paleoflood and non-exceedance bound ages were established based on reconnaissance-level methods. Multiple lines of evidence were used at each streamgage to arrive at a conclusion (age estimate) to add a higher degree of certainty to reconnaissance-level estimates. Paleoflood or nonexceedance bound evidence was documented at 41 streamgages, and 3 streamgages had previously collected paleoflood data.To determine the peak discharge of a paleoflood or non-exceedanc bound, two different hydraulic models were used.The mean standard error of prediction (SEP) for all 8 AEPDs was reduced approximately 25 percent compared to the previous flood-frequency study. For paleoflood data to be effective in reducing the SEP in eastern Colorado, a larger ratio than 44 of 188 (23 percent) streamgages would need paleoflood data and that paleoflood data would need to increase the record length by more than 25 years for the 1-percent AEPD. The greatest reduction in SEP for the peak-streamflow regional-regression equations was observed when additional new basin characteristics were included in the peak-streamflow regional-regression equations and when eastern Colorado was divided into two separate hydrologic regions. To make further reductions in the uncertainties of the peak-streamflow regional-regression equations in the Foothills and Plains hydrologic regions, additional streamgages or crest-stage gages are needed to collect peak-streamflow data on natural streams in eastern Colorado.Generalized-Least Squares regression was used to compute the final peak-streamflow regional-regression equations for peak-streamflow. Dividing eastern Colorado into two new individual regions at –104° longitude resulted in peak-streamflow regional-regression equations with the smallest SEP. The new hydrologic region located between –104° longitude and the Kansas-Nebraska State line will be designated the Plains hydrologic region and the hydrologic region comprising the rest of eastern Colorado located west of the –104° longitude and east of the Rocky Mountains and below 7,500 feet in the South Platte River Basin and below 9,000 feet in the Arkansas River Basin will be designated the Foothills hydrologic region.
The microcomputer scientific software series 2: general linear model--regression.
Harold M. Rauscher
1983-01-01
The general linear model regression (GLMR) program provides the microcomputer user with a sophisticated regression analysis capability. The output provides a regression ANOVA table, estimators of the regression model coefficients, their confidence intervals, confidence intervals around the predicted Y-values, residuals for plotting, a check for multicollinearity, a...
Siegrist, Michael; Connor, Melanie; Keller, Carmen
2012-08-01
In 2005, Swiss citizens endorsed a moratorium on gene technology, resulting in the prohibition of the commercial cultivation of genetically modified crops and the growth of genetically modified animals until 2013. However, scientific research was not affected by this moratorium, and in 2008, GMO field experiments were conducted that allowed us to examine the factors that influence their acceptance by the public. In this study, trust and confidence items were analyzed using principal component analysis. The analysis revealed the following three factors: "economy/health and environment" (value similarity based trust), "trust and honesty of industry and scientists" (value similarity based trust), and "competence" (confidence). The results of a regression analysis showed that all the three factors significantly influenced the acceptance of GM field experiments. Furthermore, risk communication scholars have suggested that fairness also plays an important role in the acceptance of environmental hazards. We, therefore, included measures for outcome fairness and procedural fairness in our model. However, the impact of fairness may be moderated by moral conviction. That is, fairness may be significant for people for whom GMO is not an important issue, but not for people for whom GMO is an important issue. The regression analysis showed that, in addition to the trust and confidence factors, moral conviction, outcome fairness, and procedural fairness were significant predictors. The results suggest that the influence of procedural fairness is even stronger for persons having high moral convictions compared with persons having low moral convictions. © 2012 Society for Risk Analysis.
Imai, Chisato; Hashizume, Masahiro
2015-03-01
Time series analysis is suitable for investigations of relatively direct and short-term effects of exposures on outcomes. In environmental epidemiology studies, this method has been one of the standard approaches to assess impacts of environmental factors on acute non-infectious diseases (e.g. cardiovascular deaths), with conventionally generalized linear or additive models (GLM and GAM). However, the same analysis practices are often observed with infectious diseases despite of the substantial differences from non-infectious diseases that may result in analytical challenges. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines, systematic review was conducted to elucidate important issues in assessing the associations between environmental factors and infectious diseases using time series analysis with GLM and GAM. Published studies on the associations between weather factors and malaria, cholera, dengue, and influenza were targeted. Our review raised issues regarding the estimation of susceptible population and exposure lag times, the adequacy of seasonal adjustments, the presence of strong autocorrelations, and the lack of a smaller observation time unit of outcomes (i.e. daily data). These concerns may be attributable to features specific to infectious diseases, such as transmission among individuals and complicated causal mechanisms. The consequence of not taking adequate measures to address these issues is distortion of the appropriate risk quantifications of exposures factors. Future studies should pay careful attention to details and examine alternative models or methods that improve studies using time series regression analysis for environmental determinants of infectious diseases.
USAF (United States Air Force) Stability and Control DATCOM (Data Compendium)
1978-04-01
regression analysis involves the study of a group of variables to determine their effect on a given parameter. Because of the empirical nature of this...regression analysis of mathematical statistics. In general, a regression analysis involves the study of a group of variables to determine their effect on a...Excperiment, OSR TN 58-114, MIT Fluid Dynamics Research Group Rapt. 57-5, 1957. (U) 90. Kennet, H., and Ashley, H.: Review of Unsteady Aerodynamic Studies in
Tokunaga, Makoto; Watanabe, Susumu; Sonoda, Shigeru
2017-09-01
Multiple linear regression analysis is often used to predict the outcome of stroke rehabilitation. However, the predictive accuracy may not be satisfactory. The objective of this study was to elucidate the predictive accuracy of a method of calculating motor Functional Independence Measure (mFIM) at discharge from mFIM effectiveness predicted by multiple regression analysis. The subjects were 505 patients with stroke who were hospitalized in a convalescent rehabilitation hospital. The formula "mFIM at discharge = mFIM effectiveness × (91 points - mFIM at admission) + mFIM at admission" was used. By including the predicted mFIM effectiveness obtained through multiple regression analysis in this formula, we obtained the predicted mFIM at discharge (A). We also used multiple regression analysis to directly predict mFIM at discharge (B). The correlation between the predicted and the measured values of mFIM at discharge was compared between A and B. The correlation coefficients were .916 for A and .878 for B. Calculating mFIM at discharge from mFIM effectiveness predicted by multiple regression analysis had a higher degree of predictive accuracy of mFIM at discharge than that directly predicted. Copyright © 2017 National Stroke Association. Published by Elsevier Inc. All rights reserved.
Parametric study and performance analysis of hybrid rocket motors with double-tube configuration
NASA Astrophysics Data System (ADS)
Yu, Nanjia; Zhao, Bo; Lorente, Arnau Pons; Wang, Jue
2017-03-01
The practical implementation of hybrid rocket motors has historically been hampered by the slow regression rate of the solid fuel. In recent years, the research on advanced injector designs has achieved notable results in the enhancement of the regression rate and combustion efficiency of hybrid rockets. Following this path, this work studies a new configuration called double-tube characterized by injecting the gaseous oxidizer through a head end injector and an inner tube with injector holes distributed along the motor longitudinal axis. This design has demonstrated a significant potential for improving the performance of hybrid rockets by means of a better mixing of the species achieved through a customized injection of the oxidizer. Indeed, the CFD analysis of the double-tube configuration has revealed that this design may increase the regression rate over 50% with respect to the same motor with a conventional axial showerhead injector. However, in order to fully exploit the advantages of the double-tube concept, it is necessary to acquire a deeper understanding of the influence of the different design parameters in the overall performance. In this way, a parametric study is carried out taking into account the variation of the oxidizer mass flux rate, the ratio of oxidizer mass flow rate injected through the inner tube to the total oxidizer mass flow rate, and injection angle. The data for the analysis have been gathered from a large series of three-dimensional numerical simulations that considered the changes in the design parameters. The propellant combination adopted consists of gaseous oxygen as oxidizer and high-density polyethylene as solid fuel. Furthermore, the numerical model comprises Navier-Stokes equations, k-ε turbulence model, eddy-dissipation combustion model and solid-fuel pyrolysis, which is computed through user-defined functions. This numerical model was previously validated by analyzing the computational and experimental results obtained for conventional hybrid rocket designs. In addition, a performance analysis is conducted in order to evaluate the influence in the performance provoked by the possible growth of the diameter of the inner fuel grain holes during the motor operation. The latter phenomenon is known as burn through holes. Finally, after a statistical analysis of the data, a regression rate expression as a function of the design parameters is obtained.
John W. Edwards; Susan C. Loeb; David C. Guynn
1994-01-01
Multiple regression and use-availability analyses are two methods for examining habitat selection. Use-availability analysis is commonly used to evaluate macrohabitat selection whereas multiple regression analysis can be used to determine microhabitat selection. We compared these techniques using behavioral observations (n = 5534) and telemetry locations (n = 2089) of...
Anderson, Carl A; McRae, Allan F; Visscher, Peter M
2006-07-01
Standard quantitative trait loci (QTL) mapping techniques commonly assume that the trait is both fully observed and normally distributed. When considering survival or age-at-onset traits these assumptions are often incorrect. Methods have been developed to map QTL for survival traits; however, they are both computationally intensive and not available in standard genome analysis software packages. We propose a grouped linear regression method for the analysis of continuous survival data. Using simulation we compare this method to both the Cox and Weibull proportional hazards models and a standard linear regression method that ignores censoring. The grouped linear regression method is of equivalent power to both the Cox and Weibull proportional hazards methods and is significantly better than the standard linear regression method when censored observations are present. The method is also robust to the proportion of censored individuals and the underlying distribution of the trait. On the basis of linear regression methodology, the grouped linear regression model is computationally simple and fast and can be implemented readily in freely available statistical software.
The impact of female education on fertility desires in Bangladesh.
Matin, K A
1985-01-01
The objective of this study is to determine whether in Bangladesh a wife's educational level significantly affects her desire for more children, or whether Bangladesh has not yet reached the threshold of socioeconomic development where education has any effect on the number of additional children desired. The analysis was carried out in a rural-urban context and also at different stages in the woman's life cycle, as defined by parity at time of interview. 2-3 live births were considered early stage, and 4 or more were defined as later stage. The analysis was carried out on 2791 women selected from the Bangladesh Fertility Survey of 1975-76 who at the time of interview met the criteria of being married, fecund, parity of 2 or more, and having a husband present in the household. The method used to perform the analysis was the Maximum Likelihood estimation method for the truncated dependent variable developed by J. Tobin. The method is a hybrid of probit analysis and classical regression and was chosen because the dependent variable, number of additional children desired, had a clustering of values at the lower limit of O. Separate regressions were run by life cycle stage (parity) and rural-urban background. The effect of variables such as the education of the husband, age, age at first marriage, work status of wife, number of living male children, and infant mortality were held constant in order to find the impact of education on desired fertility. The study found that the number of additional children desired by rural wives was greater than that desired by urban wives at all stages of the life cycle. The mean number of additional children desired by a rural wife at the early stage of the life cycle was .57 and in the later stage .15. For urban women the corresponding values were .29 and .12. The study found that the educational level of both the wife and the husband and the age of the wife at first marriage were not significant. The age of the woman was not significant in urban areas but was negatively significant.in rural areas. The percent of dead children was positively significant in both urban and rural areas for early stage women but for later stage women only in rural areas. In rural areas also the number of living male children was positively significant. The author concludes that Bangladesh is not at the developmental level where education has a significant effect on desire for additional children and that desired fertility is more influenced by antecedent factors such as income, infant mortality, and preference for male children. He recommends that measures be taken to improve socioeconomic levels and provide employment opportunities for women. He also suggests that additional analysis be done to identify the factors important in determining fertility decisions in urban areas and that separate policies be adopted for rural and urban areas.
ERIC Educational Resources Information Center
Duke, Joshua M.; Sassoon, David M.
2017-01-01
The concept of negative externality is central to the teaching of environmental economics, but corrective taxes are almost always regressive. How exactly might governments return externality-correcting tax revenue to overcome regressivity and not alter marginal incentives? In addition, there is a desire to achieve a double dividend in the use of…
Xiao, Yunhua; Xu, YongDong; Dong, Weiling; Liang, Yili; Fan, Fenliang; Zhang, Xiaoxia; Zhang, Xian; Niu, Jiaojiao; Ma, Liyuan; She, Siyuan; He, Zhili; Liu, Xueduan; Yin, Huaqun
2015-12-01
This study used an artificial enrichment microbial consortium to examine the effects of different substrate conditions on microbial diversity, composition, and function (e.g., zinc leaching efficiency) through adding pyrite (SP group), chalcopyrite (SC group), or both (SPC group) in sphalerite bioleaching systems. 16S rRNA gene sequencing analysis showed that microbial community structures and compositions dramatically changed with additions of pyrite or chalcopyrite during the sphalerite bioleaching process. Shannon diversity index showed a significantly increase in the SP (1.460), SC (1.476), and SPC (1.341) groups compared with control (sphalerite group, 0.624) on day 30, meanwhile, zinc leaching efficiencies were enhanced by about 13.4, 2.9, and 13.2%, respectively. Also, additions of pyrite or chalcopyrite could increase electric potential (ORP) and the concentrations of Fe3+ and H+, which were the main factors shaping microbial community structures by Mantel test analysis. Linear regression analysis showed that ORP, Fe3+ concentration, and pH were significantly correlated to zinc leaching efficiency and microbial diversity. In addition, we found that leaching efficiency showed a positive and significant relationship with microbial diversity. In conclusion, our results showed that the complicated substrates could significantly enhance microbial diversity and activity of function.
Kendrick, Sarah K; Zheng, Qi; Garbett, Nichola C; Brock, Guy N
2017-01-01
DSC is used to determine thermally-induced conformational changes of biomolecules within a blood plasma sample. Recent research has indicated that DSC curves (or thermograms) may have different characteristics based on disease status and, thus, may be useful as a monitoring and diagnostic tool for some diseases. Since thermograms are curves measured over a range of temperature values, they are considered functional data. In this paper we apply functional data analysis techniques to analyze differential scanning calorimetry (DSC) data from individuals from the Lupus Family Registry and Repository (LFRR). The aim was to assess the effect of lupus disease status as well as additional covariates on the thermogram profiles, and use FD analysis methods to create models for classifying lupus vs. control patients on the basis of the thermogram curves. Thermograms were collected for 300 lupus patients and 300 controls without lupus who were matched with diseased individuals based on sex, race, and age. First, functional regression with a functional response (DSC) and categorical predictor (disease status) was used to determine how thermogram curve structure varied according to disease status and other covariates including sex, race, and year of birth. Next, functional logistic regression with disease status as the response and functional principal component analysis (FPCA) scores as the predictors was used to model the effect of thermogram structure on disease status prediction. The prediction accuracy for patients with Osteoarthritis and Rheumatoid Arthritis but without Lupus was also calculated to determine the ability of the classifier to differentiate between Lupus and other diseases. Data were divided 1000 times into separate 2/3 training and 1/3 test data for evaluation of predictions. Finally, derivatives of thermogram curves were included in the models to determine whether they aided in prediction of disease status. Functional regression with thermogram as a functional response and disease status as predictor showed a clear separation in thermogram curve structure between cases and controls. The logistic regression model with FPCA scores as the predictors gave the most accurate results with a mean 79.22% correct classification rate with a mean sensitivity = 79.70%, and specificity = 81.48%. The model correctly classified OA and RA patients without Lupus as controls at a rate of 75.92% on average with a mean sensitivity = 79.70% and specificity = 77.6%. Regression models including FPCA scores for derivative curves did not perform as well, nor did regression models including covariates. Changes in thermograms observed in the disease state likely reflect covalent modifications of plasma proteins or changes in large protein-protein interacting networks resulting in the stabilization of plasma proteins towards thermal denaturation. By relating functional principal components from thermograms to disease status, our Functional Principal Component Analysis model provides results that are more easily interpretable compared to prior studies. Further, the model could also potentially be coupled with other biomarkers to improve diagnostic classification for lupus.
Burnout does not help predict depression among French school teachers.
Bianchi, Renzo; Schonfeld, Irvin Sam; Laurent, Eric
2015-11-01
Burnout has been viewed as a phase in the development of depression. However, supportive research is scarce. We examined whether burnout predicted depression among French school teachers. We conducted a 2-wave, 21-month study involving 627 teachers (73% female) working in French primary and secondary schools. Burnout was assessed with the Maslach Burnout Inventory and depression with the 9-item depression module of the Patient Health Questionnaire (PHQ-9). The PHQ-9 grades depressive symptom severity and provides a provisional diagnosis of major depression. Depression was treated both as a continuous and categorical variable using linear and logistic regression analyses. We controlled for gender, age, and length of employment. Controlling for baseline depressive symptoms, linear regression analysis showed that burnout symptoms at time 1 (T1) did not predict depressive symptoms at time 2 (T2). Baseline depressive symptoms accounted for about 88% of the association between T1 burnout and T2 depressive symptoms. Only baseline depressive symptoms predicted depressive symptoms at follow-up. Similarly, logistic regression analysis revealed that burnout symptoms at T1 did not predict incident cases of major depression at T2 when depressive symptoms at T1 were included in the predictive model. Only baseline depressive symptoms predicted cases of major depression at follow-up. This study does not support the view that burnout is a phase in the development of depression. Assessing burnout symptoms in addition to "classical" depressive symptoms may not always improve our ability to predict future depression.
Starck, Tuomo; Nikkinen, Juha; Rahko, Jukka; Remes, Jukka; Hurtig, Tuula; Haapsamo, Helena; Jussila, Katja; Kuusikko-Gauffin, Sanna; Mattila, Marja-Leena; Jansson-Verkasalo, Eira; Pauls, David L; Ebeling, Hanna; Moilanen, Irma; Tervonen, Osmo; Kiviniemi, Vesa J
2013-01-01
In resting state functional magnetic resonance imaging (fMRI) studies of autism spectrum disorders (ASDs) decreased frontal-posterior functional connectivity is a persistent finding. However, the picture of the default mode network (DMN) hypoconnectivity remains incomplete. In addition, the functional connectivity analyses have been shown to be susceptible even to subtle motion. DMN hypoconnectivity in ASD has been specifically called for re-evaluation with stringent motion correction, which we aimed to conduct by so-called scrubbing. A rich set of default mode subnetworks can be obtained with high dimensional group independent component analysis (ICA) which can potentially provide more detailed view of the connectivity alterations. We compared the DMN connectivity in high-functioning adolescents with ASDs to typically developing controls using ICA dual-regression with decompositions from typical to high dimensionality. Dual-regression analysis within DMN subnetworks did not reveal alterations but connectivity between anterior and posterior DMN subnetworks was decreased in ASD. The results were very similar with and without motion scrubbing thus indicating the efficacy of the conventional motion correction methods combined with ICA dual-regression. Specific dissociation between DMN subnetworks was revealed on high ICA dimensionality, where networks centered at the medial prefrontal cortex and retrosplenial cortex showed weakened coupling in adolescents with ASDs compared to typically developing control participants. Generally the results speak for disruption in the anterior-posterior DMN interplay on the network level whereas local functional connectivity in DMN seems relatively unaltered.
Ochi, H; Ikuma, I; Toda, H; Shimada, T; Morioka, S; Moriyama, K
1989-12-01
In order to determine whether isovolumic relaxation period (IRP) reflects left ventricular relaxation under different afterload conditions, 17 anesthetized, open chest dogs were studied, and the left ventricular pressure decay time constant (T) was calculated. In 12 dogs, angiotensin II and nitroprusside were administered, with the heart rate constant at 90 beats/min. Multiple linear regression analysis showed that the aortic dicrotic notch pressure (AoDNP) and T were major determinants of IRP, while left ventricular end-diastolic pressure was a minor determinant. Multiple linear regression analysis, correlating T with IRP and AoDNP, did not further improve the correlation coefficient compared with that between T and IRP. We concluded that correction of the IRP by AoDNP is not necessary to predict T from additional multiple linear regression. The effects of ascending aortic constriction or angiotensin II on IRP were examined in five dogs, after pretreatment with propranolol. Aortic constriction caused a significant decrease in IRP and T, while angiotensin II produced a significant increase in IRP and T. IRP was affected by the change of afterload. However, the IRP and T values were always altered in the same direction. These results demonstrate that IRP is substituted for T and it reflects left ventricular relaxation even in different afterload conditions. We conclude that IRP is a simple parameter easily used to evaluate left ventricular relaxation in clinical situations.
Tax amnesties, justice perceptions, and filing behavior: a simulation study.
Rechberger, Silvia; Hartner, Martina; Kirchler, Erich; Hämmerle, Franziska
2010-04-01
A simulation study demonstrates the influence of perceived justice of a tax amnesty on subsequent tax compliance. In addition, it investigates how the amnesty is perceived to serve the punishment objectives retribution (i.e., giving offenders what they "deserve") and value restoration (i.e., restoring the values violated by tax evasion). Hierarchical regression analysis revealed the expected positive influence of justice on subsequent tax compliance. However, when the influence of punishment objectives was controlled for, the influence of justice disappeared, while retribution and value restoration showed positive effects on post-amnesty tax compliance.
Exploring the Diversification Discount: A Focus on High-Technology Target Firms
2003-03-01
26 3. Breusch - Pagan Test Analysis ……………………………………………………..39 ix AFIT/GCA/ENV/03-01 Abstract When firms choose to acquire...of fit test , the Durbin-Watson test , and the Breusch - Pagan test respectively. A further discussion of the validation of the three regression...addition, the Breusch - Pagan test was employed to objectively test the assumption (Neter, 1996). The test yielded a p-value of 0.990, again
A new method for inferring carbon monoxide concentrations from gas filter radiometer data
NASA Technical Reports Server (NTRS)
Wallio, H. A.; Reichle, H. G., Jr.; Casas, J. C.; Gormsen, B. B.
1981-01-01
A method for inferring carbon monoxide concentrations from gas filter radiometer data is presented. The technique can closely approximate the results of more costly line-by-line radiative transfer calculations over a wide range of altitudes, ground temperatures, and carbon monoxide concentrations. The technique can also be used over a larger range of conditions than those used for the regression analysis. Because the influence of the carbon monoxide mixing ratio requires only addition, multiplication and a minimum of logic, the method can be implemented on very small computers or microprocessors.
Tanpitukpongse, Teerath P.; Mazurowski, Maciej A.; Ikhena, John; Petrella, Jeffrey R.
2016-01-01
Background and Purpose To assess prognostic efficacy of individual versus combined regional volumetrics in two commercially-available brain volumetric software packages for predicting conversion of patients with mild cognitive impairment to Alzheimer's disease. Materials and Methods Data was obtained through the Alzheimer's Disease Neuroimaging Initiative. 192 subjects (mean age 74.8 years, 39% female) diagnosed with mild cognitive impairment at baseline were studied. All had T1WI MRI sequences at baseline and 3-year clinical follow-up. Analysis was performed with NeuroQuant® and Neuroreader™. Receiver operating characteristic curves assessing the prognostic efficacy of each software package were generated using a univariable approach employing individual regional brain volumes, as well as two multivariable approaches (multiple regression and random forest), combining multiple volumes. Results On univariable analysis of 11 NeuroQuant® and 11 Neuroreader™ regional volumes, hippocampal volume had the highest area under the curve for both software packages (0.69 NeuroQuant®, 0.68 Neuroreader™), and was not significantly different (p > 0.05) between packages. Multivariable analysis did not increase the area under the curve for either package (0.63 logistic regression, 0.60 random forest NeuroQuant®; 0.65 logistic regression, 0.62 random forest Neuroreader™). Conclusion Of the multiple regional volume measures available in FDA-cleared brain volumetric software packages, hippocampal volume remains the best single predictor of conversion of mild cognitive impairment to Alzheimer's disease at 3-year follow-up. Combining volumetrics did not add additional prognostic efficacy. Therefore, future prognostic studies in MCI, combining such tools with demographic and other biomarker measures, are justified in using hippocampal volume as the only volumetric biomarker. PMID:28057634
A model to predict progression in brain-injured patients.
Tommasino, N; Forteza, D; Godino, M; Mizraji, R; Alvarez, I
2014-11-01
The study of brain death (BD) epidemiology and the acute brain injury (ABI) progression profile is important to improve public health programs, organ procurement strategies, and intensive care unit (ICU) protocols. The purpose of this study was to analyze the ABI progression profile among patients admitted to ICUs with a Glasgow Coma Score (GCS) ≤8, as well as establishing a prediction model of probability of death and BD. This was a retrospective analysis of prospective data that included all brain-injured patients with GCS ≤8 admitted to a total of four public and private ICUs in Uruguay (N = 1447). The independent predictor factors of death and BD were studied using logistic regression analysis. A hierarchical model consisting of 2 nested logit regression models was then created. With these models, the probabilities of death, BD, and death by cardiorespiratory arrest were analyzed. In the first regression, we observed that as the GCS decreased and age increased, the probability of death rose. Each additional year of age increased the probability of death by 0.014. In the second model, however, BD risk decreased with each year of age. The presence of swelling, mass effect, and/or space-occupying lesion increased BD risk for the same given GCS. In the presence of injuries compatible with intracranial hypertension, age behaved as a protective factor that reduced the probability of BD. Based on the analysis of the local epidemiology, a model to predict the probability of death and BD can be developed. The organ potential donation of a country, region, or hospital can be predicted on the basis of this model, customizing it to each specific situation.
Kapke, G E; Watson, G; Sheffler, S; Hunt, D; Frederick, C
1997-01-01
Several assays for quantification of DNA have been developed and are currently used in research and clinical laboratories. However, comparison of assay results has been difficult owing to the use of different standards and units of measurements as well as differences between assays in dynamic range and quantification limits. Although a few studies have compared results generated by different assays, there has been no consensus on conversion factors and thorough analysis has been precluded by small sample size and limited dynamic range studied. In this study, we have compared the Chiron branched DNA (bDNA) and Abbott liquid hybridization assays for quantification of hepatitis B virus (HBV) DNA in clinical specimens and have derived conversion factors to facilitate comparison of assay results. Additivity and variance stabilizing (AVAS) regression, a form of non-linear regression analysis, was performed on assay results for specimens from HBV clinical trials. Our results show that there is a strong linear relationship (R2 = 0.96) between log Chiron and log Abbott assay results. Conversion factors derived from regression analyses were found to be non-constant and ranged from 6-40. Analysis of paired assay results below and above each assay's limit of quantification (LOQ) indicated that a significantly (P < 0.01) larger proportion of observations were below the Abbott assay LOQ but above the Chiron assay LOQ, indicating that the Chiron assay is significantly more sensitive than the Abbott assay. Testing of replicate specimens showed that the Chiron assay consistently yielded lower per cent coefficients of variance (% CVs) than the Abbott assay, indicating that the Chiron assay provides superior precision.
Chen, Jinyun; Pande, Mala; Huang, Yu-Jing; Wei, Chongjuan; Amos, Christopher I; Talseth-Palmer, Bente A; Meldrum, Cliff J; Chen, Wei V; Gorlov, Ivan P; Lynch, Patrick M; Scott, Rodney J; Frazier, Marsha L
2013-02-01
Heterogeneity in age of onset of colorectal cancer in individuals with mutations in DNA mismatch repair genes (Lynch syndrome) suggests the influence of other lifestyle and genetic modifiers. We hypothesized that genes regulating the cell cycle influence the observed heterogeneity as cell cycle-related genes respond to DNA damage by arresting the cell cycle to provide time for repair and induce transcription of genes that facilitate repair. We examined the association of 1456 single nucleotide polymorphisms (SNPs) in 128 cell cycle-related genes and 31 DNA repair-related genes in 485 non-Hispanic white participants with Lynch syndrome to determine whether there are SNPs associated with age of onset of colorectal cancer. Genotyping was performed on an Illumina GoldenGate platform, and data were analyzed using Kaplan-Meier survival analysis, Cox regression analysis and classification and regression tree (CART) methods. Ten SNPs were independently significant in a multivariable Cox proportional hazards regression model after correcting for multiple comparisons (P < 5 × 10(-4)). Furthermore, risk modeling using CART analysis defined combinations of genotypes for these SNPs with which subjects could be classified into low-risk, moderate-risk and high-risk groups that had median ages of colorectal cancer onset of 63, 50 and 42 years, respectively. The age-associated risk of colorectal cancer in the high-risk group was more than four times the risk in the low-risk group (hazard ratio = 4.67, 95% CI = 3.16-6.92). The additional genetic markers identified may help in refining risk groups for more tailored screening and follow-up of non-Hispanic white patients with Lynch syndrome.
Chen, Jinyun; Pande, Mala
2013-01-01
Heterogeneity in age of onset of colorectal cancer in individuals with mutations in DNA mismatch repair genes (Lynch syndrome) suggests the influence of other lifestyle and genetic modifiers. We hypothesized that genes regulating the cell cycle influence the observed heterogeneity as cell cycle–related genes respond to DNA damage by arresting the cell cycle to provide time for repair and induce transcription of genes that facilitate repair. We examined the association of 1456 single nucleotide polymorphisms (SNPs) in 128 cell cycle–related genes and 31 DNA repair–related genes in 485 non-Hispanic white participants with Lynch syndrome to determine whether there are SNPs associated with age of onset of colorectal cancer. Genotyping was performed on an Illumina GoldenGate platform, and data were analyzed using Kaplan–Meier survival analysis, Cox regression analysis and classification and regression tree (CART) methods. Ten SNPs were independently significant in a multivariable Cox proportional hazards regression model after correcting for multiple comparisons (P < 5×10–4). Furthermore, risk modeling using CART analysis defined combinations of genotypes for these SNPs with which subjects could be classified into low-risk, moderate-risk and high-risk groups that had median ages of colorectal cancer onset of 63, 50 and 42 years, respectively. The age-associated risk of colorectal cancer in the high-risk group was more than four times the risk in the low-risk group (hazard ratio = 4.67, 95% CI = 3.16–6.92). The additional genetic markers identified may help in refining risk groups for more tailored screening and follow-up of non-Hispanic white patients with Lynch syndrome. PMID:23125224
Determinants of single family residential water use across scales in four western US cities.
Chang, Heejun; Bonnette, Matthew Ryan; Stoker, Philip; Crow-Miller, Britt; Wentz, Elizabeth
2017-10-15
A growing body of literature examines urban water sustainability with increasing evidence that locally-based physical and social spatial interactions contribute to water use. These studies however are based on single-city analysis and often fail to consider whether these interactions occur more generally. We examine a multi-city comparison using a common set of spatially-explicit water, socioeconomic, and biophysical data. We investigate the relative importance of variables for explaining the variations of single family residential (SFR) water uses at Census Block Group (CBG) and Census Tract (CT) scales in four representative western US cities - Austin, Phoenix, Portland, and Salt Lake City, - which cover a wide range of climate and development density. We used both ordinary least squares regression and spatial error regression models to identify the influence of spatial dependence on water use patterns. Our results show that older downtown areas show lower water use than newer suburban areas in all four cities. Tax assessed value and building age are the main determinants of SFR water use across the four cities regardless of the scale. Impervious surface area becomes an important variable for summer water use in all cities, and it is important in all seasons for arid environments such as Phoenix. CT level analysis shows better model predictability than CBG analysis. In all cities, seasons, and spatial scales, spatial error regression models better explain the variations of SFR water use. Such a spatially-varying relationship of urban water consumption provides additional evidence for the need to integrate urban land use planning and municipal water planning. Copyright © 2017 Elsevier B.V. All rights reserved.
Magnitude and Frequency of Floods on Nontidal Streams in Delaware
Ries, Kernell G.; Dillow, Jonathan J.A.
2006-01-01
Reliable estimates of the magnitude and frequency of annual peak flows are required for the economical and safe design of transportation and water-conveyance structures. This report, done in cooperation with the Delaware Department of Transportation (DelDOT) and the Delaware Geological Survey (DGS), presents methods for estimating the magnitude and frequency of floods on nontidal streams in Delaware at locations where streamgaging stations monitor streamflow continuously and at ungaged sites. Methods are presented for estimating the magnitude of floods for return frequencies ranging from 2 through 500 years. These methods are applicable to watersheds exhibiting a full range of urban development conditions. The report also describes StreamStats, a web application that makes it easy to obtain flood-frequency estimates for user-selected locations on Delaware streams. Flood-frequency estimates for ungaged sites are obtained through a process known as regionalization, using statistical regression analysis, where information determined for a group of streamgaging stations within a region forms the basis for estimates for ungaged sites within the region. One hundred and sixteen streamgaging stations in and near Delaware with at least 10 years of non-regulated annual peak-flow data available were used in the regional analysis. Estimates for gaged sites are obtained by combining the station peak-flow statistics (mean, standard deviation, and skew) and peak-flow estimates with regional estimates of skew and flood-frequency magnitudes. Example flood-frequency estimate calculations using the methods presented in the report are given for: (1) ungaged sites, (2) gaged locations, (3) sites upstream or downstream from a gaged location, and (4) sites between gaged locations. Regional regression equations applicable to ungaged sites in the Piedmont and Coastal Plain Physiographic Provinces of Delaware are presented. The equations incorporate drainage area, forest cover, impervious area, basin storage, housing density, soil type A, and mean basin slope as explanatory variables, and have average standard errors of prediction ranging from 28 to 72 percent. Additional regression equations that incorporate drainage area and housing density as explanatory variables are presented for use in defining the effects of urbanization on peak-flow estimates throughout Delaware for the 2-year through 500-year recurrence intervals, along with suggestions for their appropriate use in predicting development-affected peak flows. Additional topics associated with the analyses performed during the study are also discussed, including: (1) the availability and description of more than 30 basin and climatic characteristics considered during the development of the regional regression equations; (2) the treatment of increasing trends in the annual peak-flow series identified at 18 gaged sites, with respect to their relations with maximum 24-hour precipitation and housing density, and their use in the regional analysis; (3) calculation of the 90-percent confidence interval associated with peak-flow estimates from the regional regression equations; and (4) a comparison of flood-frequency estimates at gages used in a previous study, highlighting the effects of various improved analytical techniques.
Nonparametric instrumental regression with non-convex constraints
NASA Astrophysics Data System (ADS)
Grasmair, M.; Scherzer, O.; Vanhems, A.
2013-03-01
This paper considers the nonparametric regression model with an additive error that is dependent on the explanatory variables. As is common in empirical studies in epidemiology and economics, it also supposes that valid instrumental variables are observed. A classical example in microeconomics considers the consumer demand function as a function of the price of goods and the income, both variables often considered as endogenous. In this framework, the economic theory also imposes shape restrictions on the demand function, such as integrability conditions. Motivated by this illustration in microeconomics, we study an estimator of a nonparametric constrained regression function using instrumental variables by means of Tikhonov regularization. We derive rates of convergence for the regularized model both in a deterministic and stochastic setting under the assumption that the true regression function satisfies a projected source condition including, because of the non-convexity of the imposed constraints, an additional smallness condition.
[Mapping environmental vulnerability from ETM + data in the Yellow River Mouth Area].
Wang, Rui-Yan; Yu, Zhen-Wen; Xia, Yan-Ling; Wang, Xiang-Feng; Zhao, Geng-Xing; Jiang, Shu-Qian
2013-10-01
The environmental vulnerability retrieval is important to support continuing data. The spatial distribution of regional environmental vulnerability was got through remote sensing retrieval. In view of soil and vegetation, the environmental vulnerability evaluation index system was built, and the environmental vulnerability of sampling points was calculated by the AHP-fuzzy method, then the correlation between the sampling points environmental vulnerability and ETM + spectral reflectance ratio including some kinds of conversion data was analyzed to determine the sensitive spectral parameters. Based on that, models of correlation analysis, traditional regression, BP neural network and support vector regression were taken to explain the quantitative relationship between the spectral reflectance and the environmental vulnerability. With this model, the environmental vulnerability distribution was retrieved in the Yellow River Mouth Area. The results showed that the correlation between the environmental vulnerability and the spring NDVI, the September NDVI and the spring brightness was better than others, so they were selected as the sensitive spectral parameters. The model precision result showed that in addition to the support vector model, the other model reached the significant level. While all the multi-variable regression was better than all one-variable regression, and the model accuracy of BP neural network was the best. This study will serve as a reliable theoretical reference for the large spatial scale environmental vulnerability estimation based on remote sensing data.
NASA Astrophysics Data System (ADS)
Li, Chengen; Cai, Guobiao; Tian, Hui
2016-06-01
This paper is aimed to analyse the combustion characteristics of hybrid rocket motor with multi-section swirl injection by simulating the combustion flow field. Numerical combustion flow field and combustion performance parameters are obtained through three-dimensional numerical simulations based on a steady numerical model proposed in this paper. The hybrid rocket motor adopts 98% hydrogen peroxide and polyethylene as the propellants. Multiple injection sections are set along the axis of the solid fuel grain, and the oxidizer enters the combustion chamber by means of tangential injection via the injector ports in the injection sections. Simulation results indicate that the combustion flow field structure of the hybrid rocket motor could be improved by multi-section swirl injection method. The transformation of the combustion flow field can greatly increase the fuel regression rate and the combustion efficiency. The average fuel regression rate of the motor with multi-section swirl injection is improved by 8.37 times compared with that of the motor with conventional head-end irrotational injection. The combustion efficiency is increased to 95.73%. Besides, the simulation results also indicate that (1) the additional injection sections can increase the fuel regression rate and the combustion efficiency; (2) the upstream offset of the injection sections reduces the combustion efficiency; and (3) the fuel regression rate and the combustion efficiency decrease with the reduction of the number of injector ports in each injection section.
Masselot, Pierre; Chebana, Fateh; Bélanger, Diane; St-Hilaire, André; Abdous, Belkacem; Gosselin, Pierre; Ouarda, Taha B M J
2018-07-01
In environmental epidemiology studies, health response data (e.g. hospitalization or mortality) are often noisy because of hospital organization and other social factors. The noise in the data can hide the true signal related to the exposure. The signal can be unveiled by performing a temporal aggregation on health data and then using it as the response in regression analysis. From aggregated series, a general methodology is introduced to account for the particularities of an aggregated response in a regression setting. This methodology can be used with usually applied regression models in weather-related health studies, such as generalized additive models (GAM) and distributed lag nonlinear models (DLNM). In particular, the residuals are modelled using an autoregressive-moving average (ARMA) model to account for the temporal dependence. The proposed methodology is illustrated by modelling the influence of temperature on cardiovascular mortality in Canada. A comparison with classical DLNMs is provided and several aggregation methods are compared. Results show that there is an increase in the fit quality when the response is aggregated, and that the estimated relationship focuses more on the outcome over several days than the classical DLNM. More precisely, among various investigated aggregation schemes, it was found that an aggregation with an asymmetric Epanechnikov kernel is more suited for studying the temperature-mortality relationship. Copyright © 2018. Published by Elsevier B.V.
Bütof, Rebecca; Hofheinz, Frank; Zöphel, Klaus; Stadelmann, Tobias; Schmollack, Julia; Jentsch, Christina; Löck, Steffen; Kotzerke, Jörg; Baumann, Michael; van den Hoff, Jörg
2015-08-01
Despite ongoing efforts to develop new treatment options, the prognosis for patients with inoperable esophageal carcinoma is still poor and the reliability of individual therapy outcome prediction based on clinical parameters is not convincing. The aim of this work was to investigate whether PET can provide independent prognostic information in such a patient group and whether the tumor-to-blood standardized uptake ratio (SUR) can improve the prognostic value of tracer uptake values. (18)F-FDG PET/CT was performed in 130 consecutive patients (mean age ± SD, 63 ± 11 y; 113 men, 17 women) with newly diagnosed esophageal cancer before definitive radiochemotherapy. In the PET images, the metabolically active tumor volume (MTV) of the primary tumor was delineated with an adaptive threshold method. The blood standardized uptake value (SUV) was determined by manually delineating the aorta in the low-dose CT. SUR values were computed as the ratio of tumor SUV and blood SUV. Uptake values were scan-time-corrected to 60 min after injection. Univariate Cox regression and Kaplan-Meier analysis with respect to overall survival (OS), distant metastases-free survival (DM), and locoregional tumor control (LRC) was performed. Additionally, a multivariate Cox regression including clinically relevant parameters was performed. In multivariate Cox regression with respect to OS, including T stage, N stage, and smoking state, MTV- and SUR-based parameters were significant prognostic factors for OS with similar effect size. Multivariate analysis with respect to DM revealed smoking state, MTV, and all SUR-based parameters as significant prognostic factors. The highest hazard ratios (HRs) were found for scan-time-corrected maximum SUR (HR = 3.9) and mean SUR (HR = 4.4). None of the PET parameters was associated with LRC. Univariate Cox regression with respect to LRC revealed a significant effect only for N stage greater than 0 (P = 0.048). PET provides independent prognostic information for OS and DM but not for LRC in patients with locally advanced esophageal carcinoma treated with definitive radiochemotherapy in addition to clinical parameters. Among the investigated uptake-based parameters, only SUR was an independent prognostic factor for OS and DM. These results suggest that the prognostic value of tracer uptake can be improved when characterized by SUR instead of SUV. Further investigations are required to confirm these preliminary results. © 2015 by the Society of Nuclear Medicine and Molecular Imaging, Inc.
Prediction by regression and intrarange data scatter in surface-process studies
Toy, T.J.; Osterkamp, W.R.; Renard, K.G.
1993-01-01
Modeling is a major component of contemporary earth science, and regression analysis occupies a central position in the parameterization, calibration, and validation of geomorphic and hydrologic models. Although this methodology can be used in many ways, we are primarily concerned with the prediction of values for one variable from another variable. Examination of the literature reveals considerable inconsistency in the presentation of the results of regression analysis and the occurrence of patterns in the scatter of data points about the regression line. Both circumstances confound utilization and evaluation of the models. Statisticians are well aware of various problems associated with the use of regression analysis and offer improved practices; often, however, their guidelines are not followed. After a review of the aforementioned circumstances and until standard criteria for model evaluation become established, we recommend, as a minimum, inclusion of scatter diagrams, the standard error of the estimate, and sample size in reporting the results of regression analyses for most surface-process studies. ?? 1993 Springer-Verlag.
NASA Astrophysics Data System (ADS)
Nishidate, Izumi; Wiswadarma, Aditya; Hase, Yota; Tanaka, Noriyuki; Maeda, Takaaki; Niizeki, Kyuichi; Aizu, Yoshihisa
2011-08-01
In order to visualize melanin and blood concentrations and oxygen saturation in human skin tissue, a simple imaging technique based on multispectral diffuse reflectance images acquired at six wavelengths (500, 520, 540, 560, 580 and 600nm) was developed. The technique utilizes multiple regression analysis aided by Monte Carlo simulation for diffuse reflectance spectra. Using the absorbance spectrum as a response variable and the extinction coefficients of melanin, oxygenated hemoglobin, and deoxygenated hemoglobin as predictor variables, multiple regression analysis provides regression coefficients. Concentrations of melanin and total blood are then determined from the regression coefficients using conversion vectors that are deduced numerically in advance, while oxygen saturation is obtained directly from the regression coefficients. Experiments with a tissue-like agar gel phantom validated the method. In vivo experiments with human skin of the human hand during upper limb occlusion and of the inner forearm exposed to UV irradiation demonstrated the ability of the method to evaluate physiological reactions of human skin tissue.
Kadiyala, Akhil; Kaur, Devinder; Kumar, Ashok
2013-02-01
The present study developed a novel approach to modeling indoor air quality (IAQ) of a public transportation bus by the development of hybrid genetic-algorithm-based neural networks (also known as evolutionary neural networks) with input variables optimized from using the regression trees, referred as the GART approach. This study validated the applicability of the GART modeling approach in solving complex nonlinear systems by accurately predicting the monitored contaminants of carbon dioxide (CO2), carbon monoxide (CO), nitric oxide (NO), sulfur dioxide (SO2), 0.3-0.4 microm sized particle numbers, 0.4-0.5 microm sized particle numbers, particulate matter (PM) concentrations less than 1.0 microm (PM10), and PM concentrations less than 2.5 microm (PM2.5) inside a public transportation bus operating on 20% grade biodiesel in Toledo, OH. First, the important variables affecting each monitored in-bus contaminant were determined using regression trees. Second, the analysis of variance was used as a complimentary sensitivity analysis to the regression tree results to determine a subset of statistically significant variables affecting each monitored in-bus contaminant. Finally, the identified subsets of statistically significant variables were used as inputs to develop three artificial neural network (ANN) models. The models developed were regression tree-based back-propagation network (BPN-RT), regression tree-based radial basis function network (RBFN-RT), and GART models. Performance measures were used to validate the predictive capacity of the developed IAQ models. The results from this approach were compared with the results obtained from using a theoretical approach and a generalized practicable approach to modeling IAQ that included the consideration of additional independent variables when developing the aforementioned ANN models. The hybrid GART models were able to capture majority of the variance in the monitored in-bus contaminants. The genetic-algorithm-based neural network IAQ models outperformed the traditional ANN methods of the back-propagation and the radial basis function networks. The novelty of this research is the development of a novel approach to modeling vehicular indoor air quality by integration of the advanced methods of genetic algorithms, regression trees, and the analysis of variance for the monitored in-vehicle gaseous and particulate matter contaminants, and comparing the results obtained from using the developed approach with conventional artificial intelligence techniques of back propagation networks and radial basis function networks. This study validated the newly developed approach using holdout and threefold cross-validation methods. These results are of great interest to scientists, researchers, and the public in understanding the various aspects of modeling an indoor microenvironment. This methodology can easily be extended to other fields of study also.
CADDIS Volume 4. Data Analysis: PECBO Appendix - R Scripts for Non-Parametric Regressions
Script for computing nonparametric regression analysis. Overview of using scripts to infer environmental conditions from biological observations, statistically estimating species-environment relationships, statistical scripts.
Jain, Nikhil; Phillips, Frank M; Khan, Safdar N
2018-04-01
A retrospective, economic analysis. The objective of this article is to analyze the distribution of 90-day payments, sources of variation, and reimbursement for complications and readmissions for primary ≥3-level posterior lumbar fusion (PLF) from Medicare data. A secondary objective was to identify risk factors for complications. Bundled payments represent a single payment system to cover all costs associated with a single episode of care, typically over 90 days. The dollar amount spent on different health service providers and the variation in payments for ≥3-level PLF have not been analyzed from a bundled perspective. Administrative claims data were used to study 90-day Medicare (2005-2012) reimbursements for primary ≥3-level PLF for deformity and degenerative conditions of the lumbar spine. Distribution of payments, sources of variation, and reimbursements for managing complications were studied using linear regression models. Risk factors for complications were studied by stepwise multiple-variable logistic regression analysis. Hospital payments comprised 73.8% share of total 90-day payment. Adjusted analysis identified several factors for variation in index hospital payments. The average 90-day Medicare payment for all multilevel PLFs without complications was $35,878 per patient. The additional average cost of treating complications with/without revision surgery within 90 days period ranged from $17,284 to $68,963. A 90-day bundle for ≥3-level PLF with readmission ranges from $88,648 (3 levels) to $117,215 (8+ levels). Rates and risk factors for complications were also identified. The average 90-day payment per patient from Medicare was $35,878 with several factors such as levels of surgery, comorbidities, and development of complications influencing the cost. The study also identifies the risks and costs associated with complications and readmissions and emphasize the significant effect these would have on bundled payments (additional burden of up to 192% the cost of an average uncomplicated procedure over 90 days). Level 3.
B-type natriuretic peptides and mortality after stroke: a systematic review and meta-analysis.
García-Berrocoso, Teresa; Giralt, Dolors; Bustamante, Alejandro; Etgen, Thorleif; Jensen, Jesper K; Sharma, Jagdish C; Shibazaki, Kensaku; Saritas, Ayhan; Chen, Xingyong; Whiteley, William N; Montaner, Joan
2013-12-03
To measure the association of B-type natriuretic peptide (BNP) and N-terminal fragment of BNP (NT-proBNP) with all-cause mortality after stroke, and to evaluate the additional predictive value of BNP/NT-proBNP over clinical information. Suitable studies for meta-analysis were found by searching MEDLINE and EMBASE databases until October 26, 2012. Weighted mean differences measured effect size; meta-regression and publication bias were assessed. Individual participant data were used to estimate effects by logistic regression and to evaluate BNP/NT-proBNP additional predictive value by area under the receiver operating characteristic curves, and integrated discrimination improvement and categorical net reclassification improvement indexes. Literature-based meta-analysis included 3,498 stroke patients from 16 studies and revealed that BNP/NT-proBNP levels were 255.78 pg/mL (95% confidence interval [CI] 105.10-406.47, p = 0.001) higher in patients who died; publication bias entailed the loss of this association. Individual participant data analysis comprised 2,258 stroke patients. After normalization of the data, patients in the highest quartile had double the risk of death after adjustment for clinical variables (NIH Stroke Scale score, age, sex) (odds ratio 2.30, 95% CI 1.32-4.01 for BNP; and odds ratio 2.63, 95% CI 1.75-3.94 for NT-proBNP). Only NT-proBNP showed a slight added value to clinical prognostic variables, increasing discrimination by 0.028 points (integrated discrimination improvement index; p < 0.001) and reclassifying 8.1% of patients into correct risk mortality categories (net reclassification improvement index; p = 0.003). Neither etiology nor time from onset to death affected the association of BNP/NT-proBNP with mortality. BNPs are associated with poststroke mortality independent of NIH Stroke Scale score, age, and sex. However, their translation to clinical practice seems difficult because BNP/NT-proBNP add only minor predictive value to clinical information.
Beyond Reading Alone: The Relationship Between Aural Literacy And Asthma Management
Rosenfeld, Lindsay; Rudd, Rima; Emmons, Karen M.; Acevedo-García, Dolores; Martin, Laurie; Buka, Stephen
2010-01-01
Objectives To examine the relationship between literacy and asthma management with a focus on the oral exchange. Methods Study participants, all of whom reported asthma, were drawn from the New England Family Study (NEFS), an examination of links between education and health. NEFS data included reading, oral (speaking), and aural (listening) literacy measures. An additional survey was conducted with this group of study participants related to asthma issues, particularly asthma management. Data analysis focused on bivariate and multivariable logistic regression. Results In bivariate logistic regression models exploring aural literacy, there was a statistically significant association between those participants with lower aural literacy skills and less successful asthma management (OR:4.37, 95%CI:1.11, 17.32). In multivariable logistic regression analyses, controlling for gender, income, and race in separate models (one-at-a-time), there remained a statistically significant association between those participants with lower aural literacy skills and less successful asthma management. Conclusion Lower aural literacy skills seem to complicate asthma management capabilities. Practice Implications Greater attention to the oral exchange, in particular the listening skills highlighted by aural literacy, as well as other related literacy skills may help us develop strategies for clear communication related to asthma management. PMID:20399060
Semiparametric regression analysis of interval-censored competing risks data.
Mao, Lu; Lin, Dan-Yu; Zeng, Donglin
2017-09-01
Interval-censored competing risks data arise when each study subject may experience an event or failure from one of several causes and the failure time is not observed directly but rather is known to lie in an interval between two examinations. We formulate the effects of possibly time-varying (external) covariates on the cumulative incidence or sub-distribution function of competing risks (i.e., the marginal probability of failure from a specific cause) through a broad class of semiparametric regression models that captures both proportional and non-proportional hazards structures for the sub-distribution. We allow each subject to have an arbitrary number of examinations and accommodate missing information on the cause of failure. We consider nonparametric maximum likelihood estimation and devise a fast and stable EM-type algorithm for its computation. We then establish the consistency, asymptotic normality, and semiparametric efficiency of the resulting estimators for the regression parameters by appealing to modern empirical process theory. In addition, we show through extensive simulation studies that the proposed methods perform well in realistic situations. Finally, we provide an application to a study on HIV-1 infection with different viral subtypes. © 2017, The International Biometric Society.
NASA Technical Reports Server (NTRS)
Patnaik, Surya N.; Guptill, James D.; Hopkins, Dale A.; Lavelle, Thomas M.
2000-01-01
The NASA Engine Performance Program (NEPP) can configure and analyze almost any type of gas turbine engine that can be generated through the interconnection of a set of standard physical components. In addition, the code can optimize engine performance by changing adjustable variables under a set of constraints. However, for engine cycle problems at certain operating points, the NEPP code can encounter difficulties: nonconvergence in the currently implemented Powell's optimization algorithm and deficiencies in the Newton-Raphson solver during engine balancing. A project was undertaken to correct these deficiencies. Nonconvergence was avoided through a cascade optimization strategy, and deficiencies associated with engine balancing were eliminated through neural network and linear regression methods. An approximation-interspersed cascade strategy was used to optimize the engine's operation over its flight envelope. Replacement of Powell's algorithm by the cascade strategy improved the optimization segment of the NEPP code. The performance of the linear regression and neural network methods as alternative engine analyzers was found to be satisfactory. This report considers two examples-a supersonic mixed-flow turbofan engine and a subsonic waverotor-topped engine-to illustrate the results, and it discusses insights gained from the improved version of the NEPP code.
NASA Technical Reports Server (NTRS)
Jones, Harrison P.; Branston, Detrick D.; Jones, Patricia B.; Popescu, Miruna D.
2002-01-01
An earlier study compared NASA/NSO Spectromagnetograph (SPM) data with spacecraft measurements of total solar irradiance (TSI) variations over a 1.5 year period in the declining phase of solar cycle 22. This paper extends the analysis to an eight-year period which also spans the rising and early maximum phases of cycle 23. The conclusions of the earlier work appear to be robust: three factors (sunspots, strong unipolar regions, and strong mixed polarity regions) describe most of the variation in the SPM record, but only the first two are associated with TSI. Additionally, the residuals of a linear multiple regression of TSI against SPM observations over the entire eight-year period show an unexplained, increasing, linear time variation with a rate of about 0.05 W m(exp -2) per year. Separate regressions for the periods before and after 1996 January 01 show no unexplained trends but differ substantially in regression parameters. This behavior may reflect a solar source of TSI variations beyond sunspots and faculae but more plausibly results from uncompensated non-solar effects in one or both of the TSI and SPM data sets.
Hoos, Anne B.; Patel, Anant R.
1996-01-01
Model-adjustment procedures were applied to the combined data bases of storm-runoff quality for Chattanooga, Knoxville, and Nashville, Tennessee, to improve predictive accuracy for storm-runoff quality for urban watersheds in these three cities and throughout Middle and East Tennessee. Data for 45 storms at 15 different sites (five sites in each city) constitute the data base. Comparison of observed values of storm-runoff load and event-mean concentration to the predicted values from the regional regression models for 10 constituents shows prediction errors, as large as 806,000 percent. Model-adjustment procedures, which combine the regional model predictions with local data, are applied to improve predictive accuracy. Standard error of estimate after model adjustment ranges from 67 to 322 percent. Calibration results may be biased due to sampling error in the Tennessee data base. The relatively large values of standard error of estimate for some of the constituent models, although representing significant reduction (at least 50 percent) in prediction error compared to estimation with unadjusted regional models, may be unacceptable for some applications. The user may wish to collect additional local data for these constituents and repeat the analysis, or calibrate an independent local regression model.
Kovalska, M P; Bürki, E; Schoetzau, A; Orguel, S F; Orguel, S; Grieshaber, M C
2011-04-01
The distinction of real progression from test variability in visual field (VF) series may be based on clinical judgment, on trend analysis based on follow-up of test parameters over time, or on identification of a significant change related to the mean of baseline exams (event analysis). The aim of this study was to compare a new population-based method (Octopus field analysis, OFA) with classic regression analyses and clinical judgment for detecting glaucomatous VF changes. 240 VF series of 240 patients with at least 9 consecutive examinations available were included into this study. They were independently classified by two experienced investigators. The results of such a classification served as a reference for comparison for the following statistical tests: (a) t-test global, (b) r-test global, (c) regression analysis of 10 VF clusters and (d) point-wise linear regression analysis. 32.5 % of the VF series were classified as progressive by the investigators. The sensitivity and specificity were 89.7 % and 92.0 % for r-test, and 73.1 % and 93.8 % for the t-test, respectively. In the point-wise linear regression analysis, the specificity was comparable (89.5 % versus 92 %), but the sensitivity was clearly lower than in the r-test (22.4 % versus 89.7 %) at a significance level of p = 0.01. A regression analysis for the 10 VF clusters showed a markedly higher sensitivity for the r-test (37.7 %) than the t-test (14.1 %) at a similar specificity (88.3 % versus 93.8 %) for a significant trend (p = 0.005). In regard to the cluster distribution, the paracentral clusters and the superior nasal hemifield progressed most frequently. The population-based regression analysis seems to be superior to the trend analysis in detecting VF progression in glaucoma, and may eliminate the drawbacks of the event analysis. Further, it may assist the clinician in the evaluation of VF series and may allow better visualization of the correlation between function and structure owing to VF clusters. © Georg Thieme Verlag KG Stuttgart · New York.
ERIC Educational Resources Information Center
Berenson, Mark L.
2013-01-01
There is consensus in the statistical literature that severe departures from its assumptions invalidate the use of regression modeling for purposes of inference. The assumptions of regression modeling are usually evaluated subjectively through visual, graphic displays in a residual analysis but such an approach, taken alone, may be insufficient…
L.R. Grosenbaugh
1967-01-01
Describes an expansible computerized system that provides data needed in regression or covariance analysis of as many as 50 variables, 8 of which may be dependent. Alternatively, it can screen variously generated combinations of independent variables to find the regression with the smallest mean-squared-residual, which will be fitted if desired. The user can easily...
The 2011 heat wave in Greater Houston: Effects of land use on temperature.
Zhou, Weihe; Ji, Shuang; Chen, Tsun-Hsuan; Hou, Yi; Zhang, Kai
2014-11-01
Effects of land use on temperatures during severe heat waves have been rarely studied. This paper examines land use-temperature associations during the 2011 heat wave in Greater Houston. We obtained high resolution of satellite-derived land use data from the US National Land Cover Database, and temperature observations at 138 weather stations from Weather Underground, Inc (WU) during the August of 2011, which was the hottest month in Houston since 1889. Land use regression and quantile regression methods were applied to the monthly averages of daily maximum/mean/minimum temperatures and 114 land use-related predictors. Although selected variables vary with temperature metric, distance to the coastline consistently appears among all models. Other variables are generally related to high developed intensity, open water or wetlands. In addition, our quantile regression analysis shows that distance to the coastline and high developed intensity areas have larger impacts on daily average temperatures at higher quantiles, and open water area has greater impacts on daily minimum temperatures at lower quantiles. By utilizing both land use regression and quantile regression on a recent heat wave in one of the largest US metropolitan areas, this paper provides a new perspective on the impacts of land use on temperatures. Our models can provide estimates of heat exposures for epidemiological studies, and our findings can be combined with demographic variables, air conditioning and relevant diseases information to identify 'hot spots' of population vulnerability for public health interventions to reduce heat-related health effects during heat waves. Copyright © 2014 Elsevier Inc. All rights reserved.
Mohan, Atiharsh; Singh, Preet Mohinder; Malviya, Deepak; Arya, Sunil Kumar; Singh, Dinesh Kumar
2012-01-01
Background: Epidural volume extension (EVE) is claimed to increase the block height and decrease the dose requirement for intrathecal drug. However, almost all studies have been done in obstetric population and none actually compares the effect of additional drugs added to epidural volume. Materials and Methods: Seventy-five (ASA I and II) patients scheduled for lower abdominal surgery were randomly divided into three groups. All groups received intrathecal 10 mg bupivacaine; two groups received additional 10 ml of normal saline epidurally with 25 mg tramadol or 25 mg of fentanyl. Groups were than compared for maximal block height, rate of sensory block regression to T10, and motor block regression to Bromage scale of 0. Time to first analgesia and adverse effects were also compared among the three groups. Materials and Methods: Seventy-five (ASA I and II) patients scheduled for lower abdominal surgery were randomly divided into three groups. All groups received intrathecal 10 mg bupivacaine; two groups received additional 10 ml of normal saline epidurally with 25 mg tramadol or 25 mg of fentanyl. Groups were than compared for maximal block height, rate of sensory block regression to T10, and motor block regression to Bromage scale of 0. Time to first analgesia and adverse effects were also compared among the three groups. Results: Groups with EVE had statistically significant higher block height, with a significant faster regression that the control group. However, both fentanyl and tramadol groups were inseparable in respect to motor or sensory block regression. Fentanyl group had maximal time to first analgesia, followed by tramadol and control groups. Hemodynamic alterations were also more common in EVE groups. Conclusion: EVE can increase the block height significantly, but it seems to be limited only to the physical property of additional volume in epidural space and fentanyl or tramadol do not seem to differ in their ability to alter block properties. PMID:25885615
Applications of statistics to medical science, III. Correlation and regression.
Watanabe, Hiroshi
2012-01-01
In this third part of a series surveying medical statistics, the concepts of correlation and regression are reviewed. In particular, methods of linear regression and logistic regression are discussed. Arguments related to survival analysis will be made in a subsequent paper.
Effects of urban form on the urban heat island effect based on spatial regression model.
Yin, Chaohui; Yuan, Man; Lu, Youpeng; Huang, Yaping; Liu, Yanfang
2018-09-01
The urban heat island (UHI) effect is becoming more of a concern with the accelerated process of urbanization. However, few studies have examined the effect of urban form on land surface temperature (LST) especially from an urban planning perspective. This paper used spatial regression model to investigate the effects of both land use composition and urban form on LST in Wuhan City, China, based on the regulatory planning management unit. Landsat ETM+ image data was used to estimate LST. Land use composition was calculated by impervious surface area proportion, vegetated area proportion, and water proportion, while urban form indicators included sky view factor (SVF), building density, and floor area ratio (FAR). We first tested for spatial autocorrelation of urban LST, which confirmed that a traditional regression method would be invalid. A spatial error model (SEM) was chosen because its parameters were better than a spatial lag model (SLM). The results showed that urban form metrics should be the focus for mitigation efforts of UHI effects. In addition, analysis of the relationship between urban form and UHI effect based on the regulatory planning management unit was helpful for promoting corresponding UHI effect mitigation rules in practice. Finally, the spatial regression model was recommended to be an appropriate method for dealing with problems related to the urban thermal environment. Results suggested that the impact of urbanization on the UHI effect can be mitigated not only by balancing various land use types, but also by optimizing urban form, which is even more effective. This research expands the scientific understanding of effects of urban form on UHI by explicitly analyzing indicators closely related to urban detailed planning at the level of regulatory planning management unit. In addition, it may provide important insights and effective regulation measures for urban planners to mitigate future UHI effects. Copyright © 2018 Elsevier B.V. All rights reserved.
Kanamori, Shogo; Castro, Marcia C; Sow, Seydou; Matsuno, Rui; Cissokho, Alioune; Jimba, Masamine
2016-01-01
The 5S method is a lean management tool for workplace organization, with 5S being an abbreviation for five Japanese words that translate to English as Sort, Set in Order, Shine, Standardize, and Sustain. In Senegal, the 5S intervention program was implemented in 10 health centers in two regions between 2011 and 2014. To identify the impact of the 5S intervention program on the satisfaction of clients (patients and caretakers) who visited the health centers. A standardized 5S intervention protocol was implemented in the health centers using a quasi-experimental separate pre-post samples design (four intervention and three control health facilities). A questionnaire with 10 five-point Likert items was used to measure client satisfaction. Linear regression analysis was conducted to identify the intervention's effect on the client satisfaction scores, represented by an equally weighted average of the 10 Likert items (Cronbach's alpha=0.83). Additional regression analyses were conducted to identify the intervention's effect on the scores of each Likert item. Backward stepwise linear regression ( n= 1,928) indicated a statistically significant effect of the 5S intervention, represented by an increase of 0.19 points in the client satisfaction scores in the intervention group, 6 to 8 months after the intervention ( p= 0.014). Additional regression analyses showed significant score increases of 0.44 ( p= 0.002), 0.14 ( p= 0.002), 0.06 ( p= 0.019), and 0.17 ( p= 0.044) points on four items, which, respectively were healthcare staff members' communication, explanations about illnesses or cases, and consultation duration, and clients' overall satisfaction. The 5S has the potential to improve client satisfaction at resource-poor health facilities and could therefore be recommended as a strategic option for improving the quality of healthcare service in low- and middle-income countries. To explore more effective intervention modalities, further studies need to address the mechanisms by which 5S leads to attitude changes in healthcare staff.
NASA Astrophysics Data System (ADS)
Styborski, Jeremy A.
This project was started in the interest of supplementing existing data on additives to composite solid propellants. The study on the addition of iron and aluminum nanoparticles to composite AP/HTPB propellants was conducted at the Combustion and Energy Systems Laboratory at RPI in the new strand-burner experiment setup. For this study, a large literature review was conducted on history of solid propellant combustion modeling and the empirical results of tests on binders, plasticizers, AP particle size, and additives. The study focused on the addition of nano-scale aluminum and iron in small concentrations to AP/HTPB solid propellants with an average AP particle size of 200 microns. Replacing 1% of the propellant's AP with 40-60 nm aluminum particles produced no change in combustive behavior. The addition of 1% 60-80 nm iron particles produced a significant increase in burn rate, although the increase was lesser at higher pressures. These results are summarized in Table 2. The increase in the burn rate at all pressures due to the addition of iron nanoparticles warranted further study on the effect of concentration of iron. Tests conducted at 10 atm showed that the mean regression rate varied with iron concentration, peaking at 1% and 3%. Regardless of the iron concentration, the regression rate was higher than the baseline AP/HTPB propellants. These results are summarized in Table 3.
Demirjian's method in the estimation of age: A study on human third molars.
Lewis, Amitha J; Boaz, Karen; Nagesh, K R; Srikant, N; Gupta, Neha; Nandita, K P; Manaktala, Nidhi
2015-01-01
The primary aim of the following study is to estimate the chronological age based on the stages of third molar development following the eight stages (A to H) method of Demirjian et al. (along with two modifications-Orhan) and secondary aim is to compare third molar development with sex and age. The sample consisted of 115 orthopantomograms from South Indian subjects with known chronological age and gender. Multiple regression analysis was performed with chronological age as the dependable variable and third molar root development as independent variable. All the statistical analysis was performed using the SPSS 11.0 package (IBM ® Corporation). Statistically no significant differences were found in third molar development between males and females. Depending on the available number of wisdom teeth in an individual, R (2) varied for males from 0.21 to 0.48 and for females from 0.16 to 0.38. New equations were derived for estimating the chronological age. The chronological age of a South Indian individual between 14 and 22 years may be estimated based on the regression formulae. However, additional studies with a larger study population must be conducted to meet the need for population-based information on third molar development.
Predictors of employer satisfaction: technical and non-technical skills.
Danielson, Jared A; Wu, Tsui-Feng; Fales-Williams, Amanda J; Kirk, Ryan A; Preast, Vanessa A
2012-01-01
Employers of 2007-2009 graduates from Iowa State University College of Veterinary Medicine were asked to respond to a survey regarding their overall satisfaction with their new employees as well as their new employees' preparation in several technical and non-technical skill areas. Seventy-five responses contained complete data and were used in the analysis. Four technical skill areas (data collection, data interpretation, planning, and taking action) and five non-technical skill areas (interpersonal skills, ability to deal with legal issues, business skills, making referrals, and problem solving) were identified. All of the skill area subscales listed above had appropriate reliability (Cronbach's alpha>0.70) and were positively and significantly correlated with overall employer satisfaction. Results of two simultaneous regression analyses indicated that of the four technical skill areas, taking action is the most salient predictor of employer satisfaction. Of the five non-technical skill areas, interpersonal skills, business skills, making referrals, and problem solving were the most important skills in predicting employer satisfaction. Hierarchical regression analysis revealed that all technical skills explained 25% of the variation in employer satisfaction; non-technical skills explained an additional 42% of the variation in employer satisfaction.
A New Global Regression Analysis Method for the Prediction of Wind Tunnel Model Weight Corrections
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred; Bridge, Thomas M.; Amaya, Max A.
2014-01-01
A new global regression analysis method is discussed that predicts wind tunnel model weight corrections for strain-gage balance loads during a wind tunnel test. The method determines corrections by combining "wind-on" model attitude measurements with least squares estimates of the model weight and center of gravity coordinates that are obtained from "wind-off" data points. The method treats the least squares fit of the model weight separate from the fit of the center of gravity coordinates. Therefore, it performs two fits of "wind- off" data points and uses the least squares estimator of the model weight as an input for the fit of the center of gravity coordinates. Explicit equations for the least squares estimators of the weight and center of gravity coordinates are derived that simplify the implementation of the method in the data system software of a wind tunnel. In addition, recommendations for sets of "wind-off" data points are made that take typical model support system constraints into account. Explicit equations of the confidence intervals on the model weight and center of gravity coordinates and two different error analyses of the model weight prediction are also discussed in the appendices of the paper.
Belilovsky, Eugene; Gkirtzou, Katerina; Misyrlis, Michail; Konova, Anna B; Honorio, Jean; Alia-Klein, Nelly; Goldstein, Rita Z; Samaras, Dimitris; Blaschko, Matthew B
2015-12-01
We explore various sparse regularization techniques for analyzing fMRI data, such as the ℓ1 norm (often called LASSO in the context of a squared loss function), elastic net, and the recently introduced k-support norm. Employing sparsity regularization allows us to handle the curse of dimensionality, a problem commonly found in fMRI analysis. In this work we consider sparse regularization in both the regression and classification settings. We perform experiments on fMRI scans from cocaine-addicted as well as healthy control subjects. We show that in many cases, use of the k-support norm leads to better predictive performance, solution stability, and interpretability as compared to other standard approaches. We additionally analyze the advantages of using the absolute loss function versus the standard squared loss which leads to significantly better predictive performance for the regularization methods tested in almost all cases. Our results support the use of the k-support norm for fMRI analysis and on the clinical side, the generalizability of the I-RISA model of cocaine addiction. Copyright © 2015 Elsevier Ltd. All rights reserved.
A flexible count data regression model for risk analysis.
Guikema, Seth D; Coffelt, Jeremy P; Goffelt, Jeremy P
2008-02-01
In many cases, risk and reliability analyses involve estimating the probabilities of discrete events such as hardware failures and occurrences of disease or death. There is often additional information in the form of explanatory variables that can be used to help estimate the likelihood of different numbers of events in the future through the use of an appropriate regression model, such as a generalized linear model. However, existing generalized linear models (GLM) are limited in their ability to handle the types of variance structures often encountered in using count data in risk and reliability analysis. In particular, standard models cannot handle both underdispersed data (variance less than the mean) and overdispersed data (variance greater than the mean) in a single coherent modeling framework. This article presents a new GLM based on a reformulation of the Conway-Maxwell Poisson (COM) distribution that is useful for both underdispersed and overdispersed count data and demonstrates this model by applying it to the assessment of electric power system reliability. The results show that the proposed COM GLM can provide as good of fits to data as the commonly used existing models for overdispered data sets while outperforming these commonly used models for underdispersed data sets.
Foong, Hui Foh; Hamid, Tengku Aizan; Ibrahim, Rahimah; Haron, Sharifah Azizah
2018-04-01
Research has found that depression in later life is associated with cognitive impairment. Thus, the mechanism to reduce the effect of depression on cognitive function is warranted. In this paper, we intend to examine whether intrinsic religiosity mediates the association between depression and cognitive function. The study included 2322 nationally representative community-dwelling elderly in Malaysia, randomly selected through a multi-stage proportional cluster random sampling from Peninsular Malaysia. The elderly were surveyed on socio-demographic information, cognitive function, depression and intrinsic religiosity. A four-step moderated hierarchical regression analysis was employed to test the moderating effect. Statistical analyses were performed using SPSS (version 15.0). Bivariate analyses showed that both depression and intrinsic religiosity had significant relationships with cognitive function. In addition, four-step moderated hierarchical regression analysis revealed that the intrinsic religiosity moderated the association between depression and cognitive function, after controlling for selected socio-demographic characteristics. Intrinsic religiosity might reduce the negative effect of depression on cognitive function. Professionals who are working with depressed older adults should seek ways to improve their intrinsic religiosity as one of the strategies to prevent cognitive impairment.
NASA Astrophysics Data System (ADS)
Reinhardt, Katja; Samimi, Cyrus
2018-01-01
While climatological data of high spatial resolution are largely available in most developed countries, the network of climatological stations in many other regions of the world still constitutes large gaps. Especially for those regions, interpolation methods are important tools to fill these gaps and to improve the data base indispensible for climatological research. Over the last years, new hybrid methods of machine learning and geostatistics have been developed which provide innovative prospects in spatial predictive modelling. This study will focus on evaluating the performance of 12 different interpolation methods for the wind components \\overrightarrow{u} and \\overrightarrow{v} in a mountainous region of Central Asia. Thereby, a special focus will be on applying new hybrid methods on spatial interpolation of wind data. This study is the first evaluating and comparing the performance of several of these hybrid methods. The overall aim of this study is to determine whether an optimal interpolation method exists, which can equally be applied for all pressure levels, or whether different interpolation methods have to be used for the different pressure levels. Deterministic (inverse distance weighting) and geostatistical interpolation methods (ordinary kriging) were explored, which take into account only the initial values of \\overrightarrow{u} and \\overrightarrow{v} . In addition, more complex methods (generalized additive model, support vector machine and neural networks as single methods and as hybrid methods as well as regression-kriging) that consider additional variables were applied. The analysis of the error indices revealed that regression-kriging provided the most accurate interpolation results for both wind components and all pressure heights. At 200 and 500 hPa, regression-kriging is followed by the different kinds of neural networks and support vector machines and for 850 hPa it is followed by the different types of support vector machine and ordinary kriging. Overall, explanatory variables improve the interpolation results.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Park, In Ja; Kim, Dae Yong; Kim, Hee Cheol
2015-07-01
Objective: To explore the role of adjuvant chemotherapy for patients with ypT0-2N0 rectal cancer treated by preoperative chemoradiation therapy (PCRT) and radical resection. Patients and Methods: A national consortium of 10 institutions was formed, and patients with ypT0-2N0 mid- and low-rectal cancer after PCRT and radical resection from 2004 to 2009 were included. Patients were categorized into 2 groups according to receipt of additional adjuvant chemotherapy: Adj CTx (+) versus Adj CTx (−). Propensity scores were calculated and used to perform matched and adjusted analyses comparing relapse-free survival (RFS) between treatment groups while controlling for potential confounding. Results: A totalmore » of 1016 patients, who met the selection criteria, were evaluated. Of these, 106 (10.4%) did not receive adjuvant chemotherapy. There was no overall improvement in 5-year RFS as a result of adjuvant chemotherapy [91.6% for Adj CTx (+) vs 87.5% for Adj CTx (−), P=.18]. There were no differences in 5-year local recurrence and distant metastasis rate between the 2 groups. In patients who show moderate, minimal, or no regression in tumor regression grade, however, possible association of adjuvant chemotherapy with RFS would be considered (hazard ratio 0.35; 95% confidence interval 0.14-0.88; P=.03). Cox regression analysis after propensity score matching failed to show that addition of adjuvant chemotherapy was associated with improved RFS (hazard ratio 0.81; 95% confidence interval 0.39-1.70; P=.58). Conclusions: Adjuvant chemotherapy seemed to not influence the RFS of patients with ypT0-2N0 rectal cancer after PCRT followed by radical resection. Thus, the addition of adjuvant chemotherapy needs to be weighed against its oncologic benefits.« less
Non-homogeneous hybrid rocket fuel for enhanced regression rates utilizing partial entrainment
NASA Astrophysics Data System (ADS)
Boronowsky, Kenny
A concept was developed and tested to enhance the performance and regression rate of hydroxyl terminated polybutadiene (HTPB), a commonly used hybrid rocket fuel. By adding small nodules of paraffin into the HTPB fuel, a non-homogeneous mixture was created resulting in increased regression rates. The goal was to develop a fuel with a simplified single core geometry and a tailorable regression rate. The new fuel would benefit from the structural stability of HTPB yet not suffer from the large void fraction representative of typical HTPB core geometries. Regression rates were compared between traditional HTPB single core grains, 85% HTPB mixed with 15% (by weight) paraffin cores, 70% HTPB mixed with 30% paraffin cores, and plain paraffin single core grains. Each fuel combination was tested at oxidizer flow rates, ranging from 0.9 - 3.3 g/s of gaseous oxygen, in a small scale hybrid test rocket and average regression rates were measured. While large uncertainties were present in the experimental setup, the overall data showed that the regression rate was enhanced as paraffin concentration increased. While further testing would be required at larger scales of interest, the trends are encouraging. Inclusion of paraffin nodules in the HTPB grain may produce a greater advantage than other more noxious additives in current use. In addition, it may lead to safer rocket motors with higher integrated thrust due to the decreased void fraction.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seong W. Lee
During this reporting period, the literature survey including the gasifier temperature measurement literature, the ultrasonic application and its background study in cleaning application, and spray coating process are completed. The gasifier simulator (cold model) testing has been successfully conducted. Four factors (blower voltage, ultrasonic application, injection time intervals, particle weight) were considered as significant factors that affect the temperature measurement. The Analysis of Variance (ANOVA) was applied to analyze the test data. The analysis shows that all four factors are significant to the temperature measurements in the gasifier simulator (cold model). The regression analysis for the case with the normalizedmore » room temperature shows that linear model fits the temperature data with 82% accuracy (18% error). The regression analysis for the case without the normalized room temperature shows 72.5% accuracy (27.5% error). The nonlinear regression analysis indicates a better fit than that of the linear regression. The nonlinear regression model's accuracy is 88.7% (11.3% error) for normalized room temperature case, which is better than the linear regression analysis. The hot model thermocouple sleeve design and fabrication are completed. The gasifier simulator (hot model) design and the fabrication are completed. The system tests of the gasifier simulator (hot model) have been conducted and some modifications have been made. Based on the system tests and results analysis, the gasifier simulator (hot model) has met the proposed design requirement and the ready for system test. The ultrasonic cleaning method is under evaluation and will be further studied for the gasifier simulator (hot model) application. The progress of this project has been on schedule.« less
The Economic Value of Mangroves: A Meta-Analysis
Marwa Salem; D. Evan Mercer
2012-01-01
This paper presents a synthesis of the mangrove ecosystem valuation literature through a meta-regression analysis. The main contribution of this study is that it is the first meta-analysis focusing solely on mangrove forests, whereas previous studies have included different types of wetlands. The number of studies included in the regression analysis is 44 for a total...
2014-01-01
Background Support vector regression (SVR) and Gaussian process regression (GPR) were used for the analysis of electroanalytical experimental data to estimate diffusion coefficients. Results For simulated cyclic voltammograms based on the EC, Eqr, and EqrC mechanisms these regression algorithms in combination with nonlinear kernel/covariance functions yielded diffusion coefficients with higher accuracy as compared to the standard approach of calculating diffusion coefficients relying on the Nicholson-Shain equation. The level of accuracy achieved by SVR and GPR is virtually independent of the rate constants governing the respective reaction steps. Further, the reduction of high-dimensional voltammetric signals by manual selection of typical voltammetric peak features decreased the performance of both regression algorithms compared to a reduction by downsampling or principal component analysis. After training on simulated data sets, diffusion coefficients were estimated by the regression algorithms for experimental data comprising voltammetric signals for three organometallic complexes. Conclusions Estimated diffusion coefficients closely matched the values determined by the parameter fitting method, but reduced the required computational time considerably for one of the reaction mechanisms. The automated processing of voltammograms according to the regression algorithms yields better results than the conventional analysis of peak-related data. PMID:24987463
Zhang, Chao; Jia, Pengli; Yu, Liu; Xu, Chang
2018-05-01
Dose-response meta-analysis (DRMA) is widely applied to investigate the dose-specific relationship between independent and dependent variables. Such methods have been in use for over 30 years and are increasingly employed in healthcare and clinical decision-making. In this article, we give an overview of the methodology used in DRMA. We summarize the commonly used regression model and the pooled method in DRMA. We also use an example to illustrate how to employ a DRMA by these methods. Five regression models, linear regression, piecewise regression, natural polynomial regression, fractional polynomial regression, and restricted cubic spline regression, were illustrated in this article to fit the dose-response relationship. And two types of pooling approaches, that is, one-stage approach and two-stage approach are illustrated to pool the dose-response relationship across studies. The example showed similar results among these models. Several dose-response meta-analysis methods can be used for investigating the relationship between exposure level and the risk of an outcome. However the methodology of DRMA still needs to be improved. © 2018 Chinese Cochrane Center, West China Hospital of Sichuan University and John Wiley & Sons Australia, Ltd.
Mocking, R J T; Harmsen, I; Assies, J; Koeter, M W J; Ruhé, H G; Schene, A H
2016-03-15
Omega-3 polyunsaturated fatty acid (PUFA) supplementation has been proposed as (adjuvant) treatment for major depressive disorder (MDD). In the present meta-analysis, we pooled randomized placebo-controlled trials assessing the effects of omega-3 PUFA supplementation on depressive symptoms in MDD. Moreover, we performed meta-regression to test whether supplementation effects depended on eicosapentaenoic acid (EPA) or docosahexaenoic acid dose, their ratio, study duration, participants' age, percentage antidepressant users, baseline MDD symptom severity, publication year and study quality. To limit heterogeneity, we only included studies in adult patients with MDD assessed using standardized clinical interviews, and excluded studies that specifically studied perinatal/perimenopausal or comorbid MDD. Our PubMED/EMBASE search resulted in 1955 articles, from which we included 13 studies providing 1233 participants. After taking potential publication bias into account, meta-analysis showed an overall beneficial effect of omega-3 PUFAs on depressive symptoms in MDD (standardized mean difference=0.398 (0.114-0.682), P=0.006, random-effects model). As an explanation for significant heterogeneity (I(2)=73.36, P<0.001), meta-regression showed that higher EPA dose (β=0.00037 (0.00009-0.00065), P=0.009), higher percentage antidepressant users (β=0.0058 (0.00017-0.01144), P=0.044) and earlier publication year (β=-0.0735 (-0.143 to 0.004), P=0.04) were significantly associated with better outcome for PUFA supplementation. Additional sensitivity analyses were performed. In conclusion, present meta-analysis suggested a beneficial overall effect of omega-3 PUFA supplementation in MDD patients, especially for higher doses of EPA and in participants taking antidepressants. Future precision medicine trials should establish whether possible interactions between EPA and antidepressants could provide targets to improve antidepressant response and its prediction. Furthermore, potential long-term biochemical side effects of high-dosed add-on EPA supplementation should be carefully monitored.
Suzuki, Taku; Iwamoto, Takuji; Shizu, Kanae; Suzuki, Katsuji; Yamada, Harumoto; Sato, Kazuki
2017-05-01
This retrospective study was designed to investigate prognostic factors for postoperative outcomes for cubital tunnel syndrome (CubTS) using multiple logistic regression analysis with a large number of patients. Eighty-three patients with CubTS who underwent surgeries were enrolled. The following potential prognostic factors for disease severity were selected according to previous reports: sex, age, type of surgery, disease duration, body mass index, cervical lesion, presence of diabetes mellitus, Workers' Compensation status, preoperative severity, and preoperative electrodiagnostic testing. Postoperative severity of disease was assessed 2 years after surgery by Messina's criteria which is an outcome measure specifically for CubTS. Bivariate analysis was performed to select candidate prognostic factors for multiple linear regression analyses. Multiple logistic regression analysis was conducted to identify the association between postoperative severity and selected prognostic factors. Both bivariate and multiple linear regression analysis revealed only preoperative severity as an independent risk factor for poor prognosis, while other factors did not show any significant association. Although conflicting results exist regarding prognosis of CubTS, this study supports evidence from previous studies and concludes early surgical intervention portends the most favorable prognosis. Copyright © 2017 The Japanese Orthopaedic Association. Published by Elsevier B.V. All rights reserved.
Brunetti, Natale Daniele; Santoro, Francesco; De Gennaro, Luisa; Correale, Michele; Gaglione, Antonio; Di Biase, Matteo
2016-07-01
In a recent paper Singh et al. analyzed the effect of drug treatment on recurrence of takotsubo cardiomyopathy (TTC) in a comprehensive meta-analysis. The study found that recurrence rates were independent of clinic utilization of BB prescription, but inversely correlated with ACEi/ARB prescription: authors therefore conclude that ACEi/ARB rather than BB may reduce risk of recurrence. We aimed to re-analyze data reported in the study, now weighted for populations' size, in a meta-regression analysis. After multiple meta-regression analysis, we found a significant regression between rates of prescription of ACEi and rates of recurrence of TTC; regression was not statistically significant for BBs. On the bases of our re-analysis, we confirm that rates of recurrence of TTC are lower in populations of patients with higher rates of treatment with ACEi/ARB. That could not necessarily imply that ACEi may prevent recurrence of TTC, but barely that, for example, rates of recurrence are lower in cohorts more compliant with therapy or more prescribed with ACEi because more carefully followed. Randomized prospective studies are surely warranted. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
A general framework for the use of logistic regression models in meta-analysis.
Simmonds, Mark C; Higgins, Julian Pt
2016-12-01
Where individual participant data are available for every randomised trial in a meta-analysis of dichotomous event outcomes, "one-stage" random-effects logistic regression models have been proposed as a way to analyse these data. Such models can also be used even when individual participant data are not available and we have only summary contingency table data. One benefit of this one-stage regression model over conventional meta-analysis methods is that it maximises the correct binomial likelihood for the data and so does not require the common assumption that effect estimates are normally distributed. A second benefit of using this model is that it may be applied, with only minor modification, in a range of meta-analytic scenarios, including meta-regression, network meta-analyses and meta-analyses of diagnostic test accuracy. This single model can potentially replace the variety of often complex methods used in these areas. This paper considers, with a range of meta-analysis examples, how random-effects logistic regression models may be used in a number of different types of meta-analyses. This one-stage approach is compared with widely used meta-analysis methods including Bayesian network meta-analysis and the bivariate and hierarchical summary receiver operating characteristic (ROC) models for meta-analyses of diagnostic test accuracy. © The Author(s) 2014.
Examination of influential observations in penalized spline regression
NASA Astrophysics Data System (ADS)
Türkan, Semra
2013-10-01
In parametric or nonparametric regression models, the results of regression analysis are affected by some anomalous observations in the data set. Thus, detection of these observations is one of the major steps in regression analysis. These observations are precisely detected by well-known influence measures. Pena's statistic is one of them. In this study, Pena's approach is formulated for penalized spline regression in terms of ordinary residuals and leverages. The real data and artificial data are used to see illustrate the effectiveness of Pena's statistic as to Cook's distance on detecting influential observations. The results of the study clearly reveal that the proposed measure is superior to Cook's Distance to detect these observations in large data set.
Robust analysis of trends in noisy tokamak confinement data using geodesic least squares regression
DOE Office of Scientific and Technical Information (OSTI.GOV)
Verdoolaege, G., E-mail: geert.verdoolaege@ugent.be; Laboratory for Plasma Physics, Royal Military Academy, B-1000 Brussels; Shabbir, A.
Regression analysis is a very common activity in fusion science for unveiling trends and parametric dependencies, but it can be a difficult matter. We have recently developed the method of geodesic least squares (GLS) regression that is able to handle errors in all variables, is robust against data outliers and uncertainty in the regression model, and can be used with arbitrary distribution models and regression functions. We here report on first results of application of GLS to estimation of the multi-machine scaling law for the energy confinement time in tokamaks, demonstrating improved consistency of the GLS results compared to standardmore » least squares.« less
Nie, Sasa; Feng, Zhe; Tang, Li; Wang, Xiaolong; He, Yani; Fang, Jingai; Li, Suhua; Yang, Yibin; Mao, Huijuan; Jiao, Jundong; Liu, Wenhu; Cao, Ning; Wang, Wenge; Sun, Jifeng; Shao, Fengmin; Li, Wenge; He, Qiang; Jiang, Hongli; Lin, Hongli; Fu, Ping; Zhang, Xinzhou; Liu, Yinghong; Wu, Yonggui; Xi, ChunSheng; Liang, Meng; Qu, Zhijie; Zhu, Jun; Wu, Guangli; Zheng, Yali; Na, Yu; Li, Ying; Li, Wei; Cai, Guangyan; Chen, Xiangmei
2017-01-01
Risk factor studies for acute kidney injury (AKI) in China are lacking, especially those regarding non-traditional risk factors, such as laboratory indicators. All adult patients admitted to 38 tertiary and 22 secondary hospitals in China in any one month between July and December 2014 were surveyed. AKI patients were screened according to the Kidney Disease: Improving Global Outcomes' definition of AKI. Logistic regression was used to analyze the risk factors for AKI, and Cox regression was used to analyze the risk of in-hospital mortality for AKI patients; additionally, a propensity score analysis was used to reconfirm the risk factors among laboratory indicators for mortality. The morbidity of AKI was 0.97%. Independent risk factors for AKI were advancing age, male gender, hypertension, and chronic kidney disease. All-cause mortality was 16.5%. The predictors of mortality in AKI patients were advancing age, tumor, higher uric acid level and increases in Acute Physiologic Assessment and Chronic Health Evaluation II and Sequential Organ Failure Assessment scores. The hazard ratio (HR) for mortality with uric acid levels > 9.1 mg/dl compared with ≤ 5.2 mg/dl was 1.78 (95% CI: 1.23 to 2.58) for the AKI patients as a group, and was 1.73 (95% CI: 1.24 to 2.42) for a propensity score-matched set. In addition to traditional risk factors, uric acid level is an independent predictor of all-cause mortality after AKI. © 2017 The Author(s). Published by S. Karger AG, Basel.
Ou, Fengrong; Li, Kai; Gao, Qian; Liu, Dan; Li, Jinghai; Hu, Liwen; Wu, Xian; Edmiston, E Kale; Liu, Yang
2012-01-01
To investigate quality of life (QOL) and related characteristics among an urban neo-poverty population in northeast China, and to compare this population with a traditional poverty cohort. The research was a cross-sectional survey executed from June 2005 to October 2007, with a sample of 2940 individuals ages 36 to 55 in three different industrial cities of northeast China. Data were collected on QOL status and sociodemographic characteristics. QOL was assessed using the 36-item Short Form Health Survey (Chinese version). Multiple regression analysis was employed to analyze association between sociodemographic variables and QOL. The scores for QOL in the neo-poverty group were higher than those in the traditional poverty group, but lower than those in the general population. When the neo-poverty population was divided into two subgroups by age, 36-45 years and 46-55 years, the differences in QOL scores were not significant. However, there were significant differences in several dimensions between two subgroups according to unemployment time (<5 years and >5 years). Additionally, stepwise regression analysis indicated that disease burden, including disease and medical expenditures, was a common risk factor for declining QOL in the neo-poverty group. Despite some limitations, this study provides initial evidence that the QOL of the urban neo-poverty population lies between that of the general population and traditional poverty. QOL of the neo-poverty group approached QOL of the traditional poverty group with increased unemployment years. In addition to decreased income, disease burden is the most important factor influencing QOL status in urban neo-poverty.
NASA Astrophysics Data System (ADS)
Muehlbauer, J. D.; Doyle, M. W.; Tockner, K.
2011-12-01
This presentation will present the results of a meta-analysis on river-floodplain carbon/energy subsidies. This analysis combines data from the existing body of literature (ca. 100 studies) to determine a "stream signature:" a regression equation that fits the decline in aquatic-derived energy in terrestrial predator food webs as a function of distance from the river. The nature of this decay curve and its implications for river/riparian ecological dynamics will be desrcibed. Variation in this metric due to the influence of stream order, river bank characteristics, and channel geomorphology will be assessed. In addition, the implications of variation in the stream signature for terrestrial aquatic food webs under different geomorphic and anthropogenic scenarios will be discussed.
Effects of macroeconomic trends on social security spending due to sickness and disability.
Khan, Jahangir; Gerdtham, Ulf-G; Jansson, Bjarne
2004-11-01
We analyzed the relationship between macroeconomic conditions, measured as unemployment rate and social security spending, from 4 social security schemes and total spending due to sickness and disability. We obtained aggregated panel data from 13 Organization for Economic Cooperation and Development member countries for 1980-1996. We used regression analysis and fixed effect models to examine spending on sickness benefits, disability pensions, occupational-injury benefits, survivor's pensions, and total spending. A decline in unemployment increased sickness benefits spending and reduced disability pension spending. These effects reversed direction after 4 years of unemployment. Inclusion of mortality rate as an additional variable in the analysis did not affect the findings. Macroeconomic conditions influence some reimbursements from social security schemes but not total spending.