Element enrichment factor calculation using grain-size distribution and functional data regression.
Sierra, C; Ordóñez, C; Saavedra, A; Gallego, J R
2015-01-01
In environmental geochemistry studies it is common practice to normalize element concentrations in order to remove the effect of grain size. Linear regression with respect to a particular grain size or conservative element is a widely used method of normalization. In this paper, the utility of functional linear regression, in which the grain-size curve is the independent variable and the concentration of pollutant the dependent variable, is analyzed and applied to detrital sediment. After implementing functional linear regression and classical linear regression models to normalize and calculate enrichment factors, we concluded that the former regression technique has some advantages over the latter. First, functional linear regression directly considers the grain-size distribution of the samples as the explanatory variable. Second, as the regression coefficients are not constant values but functions depending on the grain size, it is easier to comprehend the relationship between grain size and pollutant concentration. Third, regularization can be introduced into the model in order to establish equilibrium between reliability of the data and smoothness of the solutions. Copyright © 2014 Elsevier Ltd. All rights reserved.
Linear regression crash prediction models : issues and proposed solutions.
DOT National Transportation Integrated Search
2010-05-01
The paper develops a linear regression model approach that can be applied to : crash data to predict vehicle crashes. The proposed approach involves novice data aggregation : to satisfy linear regression assumptions; namely error structure normality ...
A Technique of Fuzzy C-Mean in Multiple Linear Regression Model toward Paddy Yield
NASA Astrophysics Data System (ADS)
Syazwan Wahab, Nur; Saifullah Rusiman, Mohd; Mohamad, Mahathir; Amira Azmi, Nur; Che Him, Norziha; Ghazali Kamardan, M.; Ali, Maselan
2018-04-01
In this paper, we propose a hybrid model which is a combination of multiple linear regression model and fuzzy c-means method. This research involved a relationship between 20 variates of the top soil that are analyzed prior to planting of paddy yields at standard fertilizer rates. Data used were from the multi-location trials for rice carried out by MARDI at major paddy granary in Peninsular Malaysia during the period from 2009 to 2012. Missing observations were estimated using mean estimation techniques. The data were analyzed using multiple linear regression model and a combination of multiple linear regression model and fuzzy c-means method. Analysis of normality and multicollinearity indicate that the data is normally scattered without multicollinearity among independent variables. Analysis of fuzzy c-means cluster the yield of paddy into two clusters before the multiple linear regression model can be used. The comparison between two method indicate that the hybrid of multiple linear regression model and fuzzy c-means method outperform the multiple linear regression model with lower value of mean square error.
Who Will Win?: Predicting the Presidential Election Using Linear Regression
ERIC Educational Resources Information Center
Lamb, John H.
2007-01-01
This article outlines a linear regression activity that engages learners, uses technology, and fosters cooperation. Students generated least-squares linear regression equations using TI-83 Plus[TM] graphing calculators, Microsoft[C] Excel, and paper-and-pencil calculations using derived normal equations to predict the 2004 presidential election.…
An Expert System for the Evaluation of Cost Models
1990-09-01
contrast to the condition of equal error variance, called homoscedasticity. (Reference: Applied Linear Regression Models by John Neter - page 423...normal. (Reference: Applied Linear Regression Models by John Neter - page 125) Click Here to continue -> Autocorrelation Click Here for the index - Index...over time. Error terms correlated over time are said to be autocorrelated or serially correlated. (REFERENCE: Applied Linear Regression Models by John
Data Transformations for Inference with Linear Regression: Clarifications and Recommendations
ERIC Educational Resources Information Center
Pek, Jolynn; Wong, Octavia; Wong, C. M.
2017-01-01
Data transformations have been promoted as a popular and easy-to-implement remedy to address the assumption of normally distributed errors (in the population) in linear regression. However, the application of data transformations introduces non-ignorable complexities which should be fully appreciated before their implementation. This paper adds to…
Identifying the Factors That Influence Change in SEBD Using Logistic Regression Analysis
ERIC Educational Resources Information Center
Camilleri, Liberato; Cefai, Carmel
2013-01-01
Multiple linear regression and ANOVA models are widely used in applications since they provide effective statistical tools for assessing the relationship between a continuous dependent variable and several predictors. However these models rely heavily on linearity and normality assumptions and they do not accommodate categorical dependent…
Anderson, Carl A; McRae, Allan F; Visscher, Peter M
2006-07-01
Standard quantitative trait loci (QTL) mapping techniques commonly assume that the trait is both fully observed and normally distributed. When considering survival or age-at-onset traits these assumptions are often incorrect. Methods have been developed to map QTL for survival traits; however, they are both computationally intensive and not available in standard genome analysis software packages. We propose a grouped linear regression method for the analysis of continuous survival data. Using simulation we compare this method to both the Cox and Weibull proportional hazards models and a standard linear regression method that ignores censoring. The grouped linear regression method is of equivalent power to both the Cox and Weibull proportional hazards methods and is significantly better than the standard linear regression method when censored observations are present. The method is also robust to the proportion of censored individuals and the underlying distribution of the trait. On the basis of linear regression methodology, the grouped linear regression model is computationally simple and fast and can be implemented readily in freely available statistical software.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Callister, Stephen J.; Barry, Richard C.; Adkins, Joshua N.
2006-02-01
Central tendency, linear regression, locally weighted regression, and quantile techniques were investigated for normalization of peptide abundance measurements obtained from high-throughput liquid chromatography-Fourier transform ion cyclotron resonance mass spectrometry (LC-FTICR MS). Arbitrary abundances of peptides were obtained from three sample sets, including a standard protein sample, two Deinococcus radiodurans samples taken from different growth phases, and two mouse striatum samples from control and methamphetamine-stressed mice (strain C57BL/6). The selected normalization techniques were evaluated in both the absence and presence of biological variability by estimating extraneous variability prior to and following normalization. Prior to normalization, replicate runs from each sample setmore » were observed to be statistically different, while following normalization replicate runs were no longer statistically different. Although all techniques reduced systematic bias, assigned ranks among the techniques revealed significant trends. For most LC-FTICR MS analyses, linear regression normalization ranked either first or second among the four techniques, suggesting that this technique was more generally suitable for reducing systematic biases.« less
On the null distribution of Bayes factors in linear regression
USDA-ARS?s Scientific Manuscript database
We show that under the null, the 2 log (Bayes factor) is asymptotically distributed as a weighted sum of chi-squared random variables with a shifted mean. This claim holds for Bayesian multi-linear regression with a family of conjugate priors, namely, the normal-inverse-gamma prior, the g-prior, and...
Ernst, Anja F; Albers, Casper J
2017-01-01
Misconceptions about the assumptions behind the standard linear regression model are widespread and dangerous. These lead to using linear regression when inappropriate, and to employing alternative procedures with less statistical power when unnecessary. Our systematic literature review investigated employment and reporting of assumption checks in twelve clinical psychology journals. Findings indicate that normality of the variables themselves, rather than of the errors, was wrongfully held for a necessary assumption in 4% of papers that use regression. Furthermore, 92% of all papers using linear regression were unclear about their assumption checks, violating APA-recommendations. This paper appeals for a heightened awareness for and increased transparency in the reporting of statistical assumption checking.
Ernst, Anja F.
2017-01-01
Misconceptions about the assumptions behind the standard linear regression model are widespread and dangerous. These lead to using linear regression when inappropriate, and to employing alternative procedures with less statistical power when unnecessary. Our systematic literature review investigated employment and reporting of assumption checks in twelve clinical psychology journals. Findings indicate that normality of the variables themselves, rather than of the errors, was wrongfully held for a necessary assumption in 4% of papers that use regression. Furthermore, 92% of all papers using linear regression were unclear about their assumption checks, violating APA-recommendations. This paper appeals for a heightened awareness for and increased transparency in the reporting of statistical assumption checking. PMID:28533971
Linear regression analysis of survival data with missing censoring indicators.
Wang, Qihua; Dinse, Gregg E
2011-04-01
Linear regression analysis has been studied extensively in a random censorship setting, but typically all of the censoring indicators are assumed to be observed. In this paper, we develop synthetic data methods for estimating regression parameters in a linear model when some censoring indicators are missing. We define estimators based on regression calibration, imputation, and inverse probability weighting techniques, and we prove all three estimators are asymptotically normal. The finite-sample performance of each estimator is evaluated via simulation. We illustrate our methods by assessing the effects of sex and age on the time to non-ambulatory progression for patients in a brain cancer clinical trial.
ERIC Educational Resources Information Center
Osborne, Jason W.
2013-01-01
Osborne and Waters (2002) focused on checking some of the assumptions of multiple linear regression. In a critique of that paper, Williams, Grajales, and Kurkiewicz correctly clarify that regression models estimated using ordinary least squares require the assumption of normally distributed errors, but not the assumption of normally distributed…
Are your covariates under control? How normalization can re-introduce covariate effects.
Pain, Oliver; Dudbridge, Frank; Ronald, Angelica
2018-04-30
Many statistical tests rely on the assumption that the residuals of a model are normally distributed. Rank-based inverse normal transformation (INT) of the dependent variable is one of the most popular approaches to satisfy the normality assumption. When covariates are included in the analysis, a common approach is to first adjust for the covariates and then normalize the residuals. This study investigated the effect of regressing covariates against the dependent variable and then applying rank-based INT to the residuals. The correlation between the dependent variable and covariates at each stage of processing was assessed. An alternative approach was tested in which rank-based INT was applied to the dependent variable before regressing covariates. Analyses based on both simulated and real data examples demonstrated that applying rank-based INT to the dependent variable residuals after regressing out covariates re-introduces a linear correlation between the dependent variable and covariates, increasing type-I errors and reducing power. On the other hand, when rank-based INT was applied prior to controlling for covariate effects, residuals were normally distributed and linearly uncorrelated with covariates. This latter approach is therefore recommended in situations were normality of the dependent variable is required.
Notes on power of normality tests of error terms in regression models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Střelec, Luboš
2015-03-10
Normality is one of the basic assumptions in applying statistical procedures. For example in linear regression most of the inferential procedures are based on the assumption of normality, i.e. the disturbance vector is assumed to be normally distributed. Failure to assess non-normality of the error terms may lead to incorrect results of usual statistical inference techniques such as t-test or F-test. Thus, error terms should be normally distributed in order to allow us to make exact inferences. As a consequence, normally distributed stochastic errors are necessary in order to make a not misleading inferences which explains a necessity and importancemore » of robust tests of normality. Therefore, the aim of this contribution is to discuss normality testing of error terms in regression models. In this contribution, we introduce the general RT class of robust tests for normality, and present and discuss the trade-off between power and robustness of selected classical and robust normality tests of error terms in regression models.« less
Su, Liyun; Zhao, Yanyong; Yan, Tianshun; Li, Fenglan
2012-01-01
Multivariate local polynomial fitting is applied to the multivariate linear heteroscedastic regression model. Firstly, the local polynomial fitting is applied to estimate heteroscedastic function, then the coefficients of regression model are obtained by using generalized least squares method. One noteworthy feature of our approach is that we avoid the testing for heteroscedasticity by improving the traditional two-stage method. Due to non-parametric technique of local polynomial estimation, it is unnecessary to know the form of heteroscedastic function. Therefore, we can improve the estimation precision, when the heteroscedastic function is unknown. Furthermore, we verify that the regression coefficients is asymptotic normal based on numerical simulations and normal Q-Q plots of residuals. Finally, the simulation results and the local polynomial estimation of real data indicate that our approach is surely effective in finite-sample situations.
Functional Relationships and Regression Analysis.
ERIC Educational Resources Information Center
Preece, Peter F. W.
1978-01-01
Using a degenerate multivariate normal model for the distribution of organismic variables, the form of least-squares regression analysis required to estimate a linear functional relationship between variables is derived. It is suggested that the two conventional regression lines may be considered to describe functional, not merely statistical,…
Biostatistics Series Module 6: Correlation and Linear Regression.
Hazra, Avijit; Gogtay, Nithya
2016-01-01
Correlation and linear regression are the most commonly used techniques for quantifying the association between two numeric variables. Correlation quantifies the strength of the linear relationship between paired variables, expressing this as a correlation coefficient. If both variables x and y are normally distributed, we calculate Pearson's correlation coefficient ( r ). If normality assumption is not met for one or both variables in a correlation analysis, a rank correlation coefficient, such as Spearman's rho (ρ) may be calculated. A hypothesis test of correlation tests whether the linear relationship between the two variables holds in the underlying population, in which case it returns a P < 0.05. A 95% confidence interval of the correlation coefficient can also be calculated for an idea of the correlation in the population. The value r 2 denotes the proportion of the variability of the dependent variable y that can be attributed to its linear relation with the independent variable x and is called the coefficient of determination. Linear regression is a technique that attempts to link two correlated variables x and y in the form of a mathematical equation ( y = a + bx ), such that given the value of one variable the other may be predicted. In general, the method of least squares is applied to obtain the equation of the regression line. Correlation and linear regression analysis are based on certain assumptions pertaining to the data sets. If these assumptions are not met, misleading conclusions may be drawn. The first assumption is that of linear relationship between the two variables. A scatter plot is essential before embarking on any correlation-regression analysis to show that this is indeed the case. Outliers or clustering within data sets can distort the correlation coefficient value. Finally, it is vital to remember that though strong correlation can be a pointer toward causation, the two are not synonymous.
Biostatistics Series Module 6: Correlation and Linear Regression
Hazra, Avijit; Gogtay, Nithya
2016-01-01
Correlation and linear regression are the most commonly used techniques for quantifying the association between two numeric variables. Correlation quantifies the strength of the linear relationship between paired variables, expressing this as a correlation coefficient. If both variables x and y are normally distributed, we calculate Pearson's correlation coefficient (r). If normality assumption is not met for one or both variables in a correlation analysis, a rank correlation coefficient, such as Spearman's rho (ρ) may be calculated. A hypothesis test of correlation tests whether the linear relationship between the two variables holds in the underlying population, in which case it returns a P < 0.05. A 95% confidence interval of the correlation coefficient can also be calculated for an idea of the correlation in the population. The value r2 denotes the proportion of the variability of the dependent variable y that can be attributed to its linear relation with the independent variable x and is called the coefficient of determination. Linear regression is a technique that attempts to link two correlated variables x and y in the form of a mathematical equation (y = a + bx), such that given the value of one variable the other may be predicted. In general, the method of least squares is applied to obtain the equation of the regression line. Correlation and linear regression analysis are based on certain assumptions pertaining to the data sets. If these assumptions are not met, misleading conclusions may be drawn. The first assumption is that of linear relationship between the two variables. A scatter plot is essential before embarking on any correlation-regression analysis to show that this is indeed the case. Outliers or clustering within data sets can distort the correlation coefficient value. Finally, it is vital to remember that though strong correlation can be a pointer toward causation, the two are not synonymous. PMID:27904175
[From clinical judgment to linear regression model.
Palacios-Cruz, Lino; Pérez, Marcela; Rivas-Ruiz, Rodolfo; Talavera, Juan O
2013-01-01
When we think about mathematical models, such as linear regression model, we think that these terms are only used by those engaged in research, a notion that is far from the truth. Legendre described the first mathematical model in 1805, and Galton introduced the formal term in 1886. Linear regression is one of the most commonly used regression models in clinical practice. It is useful to predict or show the relationship between two or more variables as long as the dependent variable is quantitative and has normal distribution. Stated in another way, the regression is used to predict a measure based on the knowledge of at least one other variable. Linear regression has as it's first objective to determine the slope or inclination of the regression line: Y = a + bx, where "a" is the intercept or regression constant and it is equivalent to "Y" value when "X" equals 0 and "b" (also called slope) indicates the increase or decrease that occurs when the variable "x" increases or decreases in one unit. In the regression line, "b" is called regression coefficient. The coefficient of determination (R 2 ) indicates the importance of independent variables in the outcome.
Normality of raw data in general linear models: The most widespread myth in statistics
Kery, Marc; Hatfield, Jeff S.
2003-01-01
In years of statistical consulting for ecologists and wildlife biologists, by far the most common misconception we have come across has been the one about normality in general linear models. These comprise a very large part of the statistical models used in ecology and include t tests, simple and multiple linear regression, polynomial regression, and analysis of variance (ANOVA) and covariance (ANCOVA). There is a widely held belief that the normality assumption pertains to the raw data rather than to the model residuals. We suspect that this error may also occur in countless published studies, whenever the normality assumption is tested prior to analysis. This may lead to the use of nonparametric alternatives (if there are any), when parametric tests would indeed be appropriate, or to use of transformations of raw data, which may introduce hidden assumptions such as multiplicative effects on the natural scale in the case of log-transformed data. Our aim here is to dispel this myth. We very briefly describe relevant theory for two cases of general linear models to show that the residuals need to be normally distributed if tests requiring normality are to be used, such as t and F tests. We then give two examples demonstrating that the distribution of the response variable may be nonnormal, and yet the residuals are well behaved. We do not go into the issue of how to test normality; instead we display the distributions of response variables and residuals graphically.
Analysis of Learning Curve Fitting Techniques.
1987-09-01
1986. 15. Neter, John and others. Applied Linear Regression Models. Homewood IL: Irwin, 19-33. 16. SAS User’s Guide: Basics, Version 5 Edition. SAS... Linear Regression Techniques (15:23-52). Random errors are assumed to be normally distributed when using -# ordinary least-squares, according to Johnston...lot estimated by the improvement curve formula. For a more detailed explanation of the ordinary least-squares technique, see Neter, et. al., Applied
Quality of life in breast cancer patients--a quantile regression analysis.
Pourhoseingholi, Mohamad Amin; Safaee, Azadeh; Moghimi-Dehkordi, Bijan; Zeighami, Bahram; Faghihzadeh, Soghrat; Tabatabaee, Hamid Reza; Pourhoseingholi, Asma
2008-01-01
Quality of life study has an important role in health care especially in chronic diseases, in clinical judgment and in medical resources supplying. Statistical tools like linear regression are widely used to assess the predictors of quality of life. But when the response is not normal the results are misleading. The aim of this study is to determine the predictors of quality of life in breast cancer patients, using quantile regression model and compare to linear regression. A cross-sectional study conducted on 119 breast cancer patients that admitted and treated in chemotherapy ward of Namazi hospital in Shiraz. We used QLQ-C30 questionnaire to assessment quality of life in these patients. A quantile regression was employed to assess the assocciated factors and the results were compared to linear regression. All analysis carried out using SAS. The mean score for the global health status for breast cancer patients was 64.92+/-11.42. Linear regression showed that only grade of tumor, occupational status, menopausal status, financial difficulties and dyspnea were statistically significant. In spite of linear regression, financial difficulties were not significant in quantile regression analysis and dyspnea was only significant for first quartile. Also emotion functioning and duration of disease statistically predicted the QOL score in the third quartile. The results have demonstrated that using quantile regression leads to better interpretation and richer inference about predictors of the breast cancer patient quality of life.
High correlations between MRI brain volume measurements based on NeuroQuant® and FreeSurfer.
Ross, David E; Ochs, Alfred L; Tate, David F; Tokac, Umit; Seabaugh, John; Abildskov, Tracy J; Bigler, Erin D
2018-05-30
NeuroQuant ® (NQ) and FreeSurfer (FS) are commonly used computer-automated programs for measuring MRI brain volume. Previously they were reported to have high intermethod reliabilities but often large intermethod effect size differences. We hypothesized that linear transformations could be used to reduce the large effect sizes. This study was an extension of our previously reported study. We performed NQ and FS brain volume measurements on 60 subjects (including normal controls, patients with traumatic brain injury, and patients with Alzheimer's disease). We used two statistical approaches in parallel to develop methods for transforming FS volumes into NQ volumes: traditional linear regression, and Bayesian linear regression. For both methods, we used regression analyses to develop linear transformations of the FS volumes to make them more similar to the NQ volumes. The FS-to-NQ transformations based on traditional linear regression resulted in effect sizes which were small to moderate. The transformations based on Bayesian linear regression resulted in all effect sizes being trivially small. To our knowledge, this is the first report describing a method for transforming FS to NQ data so as to achieve high reliability and low effect size differences. Machine learning methods like Bayesian regression may be more useful than traditional methods. Copyright © 2018 Elsevier B.V. All rights reserved.
Modeling absolute differences in life expectancy with a censored skew-normal regression approach
Clough-Gorr, Kerri; Zwahlen, Marcel
2015-01-01
Parameter estimates from commonly used multivariable parametric survival regression models do not directly quantify differences in years of life expectancy. Gaussian linear regression models give results in terms of absolute mean differences, but are not appropriate in modeling life expectancy, because in many situations time to death has a negative skewed distribution. A regression approach using a skew-normal distribution would be an alternative to parametric survival models in the modeling of life expectancy, because parameter estimates can be interpreted in terms of survival time differences while allowing for skewness of the distribution. In this paper we show how to use the skew-normal regression so that censored and left-truncated observations are accounted for. With this we model differences in life expectancy using data from the Swiss National Cohort Study and from official life expectancy estimates and compare the results with those derived from commonly used survival regression models. We conclude that a censored skew-normal survival regression approach for left-truncated observations can be used to model differences in life expectancy across covariates of interest. PMID:26339544
A comparison of methods for the analysis of binomial clustered outcomes in behavioral research.
Ferrari, Alberto; Comelli, Mario
2016-12-01
In behavioral research, data consisting of a per-subject proportion of "successes" and "failures" over a finite number of trials often arise. This clustered binary data are usually non-normally distributed, which can distort inference if the usual general linear model is applied and sample size is small. A number of more advanced methods is available, but they are often technically challenging and a comparative assessment of their performances in behavioral setups has not been performed. We studied the performances of some methods applicable to the analysis of proportions; namely linear regression, Poisson regression, beta-binomial regression and Generalized Linear Mixed Models (GLMMs). We report on a simulation study evaluating power and Type I error rate of these models in hypothetical scenarios met by behavioral researchers; plus, we describe results from the application of these methods on data from real experiments. Our results show that, while GLMMs are powerful instruments for the analysis of clustered binary outcomes, beta-binomial regression can outperform them in a range of scenarios. Linear regression gave results consistent with the nominal level of significance, but was overall less powerful. Poisson regression, instead, mostly led to anticonservative inference. GLMMs and beta-binomial regression are generally more powerful than linear regression; yet linear regression is robust to model misspecification in some conditions, whereas Poisson regression suffers heavily from violations of the assumptions when used to model proportion data. We conclude providing directions to behavioral scientists dealing with clustered binary data and small sample sizes. Copyright © 2016 Elsevier B.V. All rights reserved.
Modeling Longitudinal Data Containing Non-Normal Within Subject Errors
NASA Technical Reports Server (NTRS)
Feiveson, Alan; Glenn, Nancy L.
2013-01-01
The mission of the National Aeronautics and Space Administration’s (NASA) human research program is to advance safe human spaceflight. This involves conducting experiments, collecting data, and analyzing data. The data are longitudinal and result from a relatively few number of subjects; typically 10 – 20. A longitudinal study refers to an investigation where participant outcomes and possibly treatments are collected at multiple follow-up times. Standard statistical designs such as mean regression with random effects and mixed–effects regression are inadequate for such data because the population is typically not approximately normally distributed. Hence, more advanced data analysis methods are necessary. This research focuses on four such methods for longitudinal data analysis: the recently proposed linear quantile mixed models (lqmm) by Geraci and Bottai (2013), quantile regression, multilevel mixed–effects linear regression, and robust regression. This research also provides computational algorithms for longitudinal data that scientists can directly use for human spaceflight and other longitudinal data applications, then presents statistical evidence that verifies which method is best for specific situations. This advances the study of longitudinal data in a broad range of applications including applications in the sciences, technology, engineering and mathematics fields.
The Inverted Student Density and Test Scores.
ERIC Educational Resources Information Center
Boldt, Robert F.
The inverted density is one whose contour lines are spheroidal as in the normal distribution, but whose moments differ from those of the normal in that its conditional arrays are not homoscedastic, being quadratic functions of the values of the linear regression functions. It is also platykurtic, its measure of kurtosis ranging from that of the…
Tan, Kok Chooi; Lim, Hwee San; Matjafri, Mohd Zubir; Abdullah, Khiruddin
2012-06-01
Atmospheric corrections for multi-temporal optical satellite images are necessary, especially in change detection analyses, such as normalized difference vegetation index (NDVI) rationing. Abrupt change detection analysis using remote-sensing techniques requires radiometric congruity and atmospheric correction to monitor terrestrial surfaces over time. Two atmospheric correction methods were used for this study: relative radiometric normalization and the simplified method for atmospheric correction (SMAC) in the solar spectrum. A multi-temporal data set consisting of two sets of Landsat images from the period between 1991 and 2002 of Penang Island, Malaysia, was used to compare NDVI maps, which were generated using the proposed atmospheric correction methods. Land surface temperature (LST) was retrieved using ATCOR3_T in PCI Geomatica 10.1 image processing software. Linear regression analysis was utilized to analyze the relationship between NDVI and LST. This study reveals that both of the proposed atmospheric correction methods yielded high accuracy through examination of the linear correlation coefficients. To check for the accuracy of the equation obtained through linear regression analysis for every single satellite image, 20 points were randomly chosen. The results showed that the SMAC method yielded a constant value (in terms of error) to predict the NDVI value from linear regression analysis-derived equation. The errors (average) from both proposed atmospheric correction methods were less than 10%.
Detecting trends in raptor counts: power and type I error rates of various statistical tests
Hatfield, J.S.; Gould, W.R.; Hoover, B.A.; Fuller, M.R.; Lindquist, E.L.
1996-01-01
We conducted simulations that estimated power and type I error rates of statistical tests for detecting trends in raptor population count data collected from a single monitoring site. Results of the simulations were used to help analyze count data of bald eagles (Haliaeetus leucocephalus) from 7 national forests in Michigan, Minnesota, and Wisconsin during 1980-1989. Seven statistical tests were evaluated, including simple linear regression on the log scale and linear regression with a permutation test. Using 1,000 replications each, we simulated n = 10 and n = 50 years of count data and trends ranging from -5 to 5% change/year. We evaluated the tests at 3 critical levels (alpha = 0.01, 0.05, and 0.10) for both upper- and lower-tailed tests. Exponential count data were simulated by adding sampling error with a coefficient of variation of 40% from either a log-normal or autocorrelated log-normal distribution. Not surprisingly, tests performed with 50 years of data were much more powerful than tests with 10 years of data. Positive autocorrelation inflated alpha-levels upward from their nominal levels, making the tests less conservative and more likely to reject the null hypothesis of no trend. Of the tests studied, Cox and Stuart's test and Pollard's test clearly had lower power than the others. Surprisingly, the linear regression t-test, Collins' linear regression permutation test, and the nonparametric Lehmann's and Mann's tests all had similar power in our simulations. Analyses of the count data suggested that bald eagles had increasing trends on at least 2 of the 7 national forests during 1980-1989.
Murad, Havi; Kipnis, Victor; Freedman, Laurence S
2016-10-01
Assessing interactions in linear regression models when covariates have measurement error (ME) is complex.We previously described regression calibration (RC) methods that yield consistent estimators and standard errors for interaction coefficients of normally distributed covariates having classical ME. Here we extend normal based RC (NBRC) and linear RC (LRC) methods to a non-classical ME model, and describe more efficient versions that combine estimates from the main study and internal sub-study. We apply these methods to data from the Observing Protein and Energy Nutrition (OPEN) study. Using simulations we show that (i) for normally distributed covariates efficient NBRC and LRC were nearly unbiased and performed well with sub-study size ≥200; (ii) efficient NBRC had lower MSE than efficient LRC; (iii) the naïve test for a single interaction had type I error probability close to the nominal significance level, whereas efficient NBRC and LRC were slightly anti-conservative but more powerful; (iv) for markedly non-normal covariates, efficient LRC yielded less biased estimators with smaller variance than efficient NBRC. Our simulations suggest that it is preferable to use: (i) efficient NBRC for estimating and testing interaction effects of normally distributed covariates and (ii) efficient LRC for estimating and testing interactions for markedly non-normal covariates. © The Author(s) 2013.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seong W. Lee
During this reporting period, the literature survey including the gasifier temperature measurement literature, the ultrasonic application and its background study in cleaning application, and spray coating process are completed. The gasifier simulator (cold model) testing has been successfully conducted. Four factors (blower voltage, ultrasonic application, injection time intervals, particle weight) were considered as significant factors that affect the temperature measurement. The Analysis of Variance (ANOVA) was applied to analyze the test data. The analysis shows that all four factors are significant to the temperature measurements in the gasifier simulator (cold model). The regression analysis for the case with the normalizedmore » room temperature shows that linear model fits the temperature data with 82% accuracy (18% error). The regression analysis for the case without the normalized room temperature shows 72.5% accuracy (27.5% error). The nonlinear regression analysis indicates a better fit than that of the linear regression. The nonlinear regression model's accuracy is 88.7% (11.3% error) for normalized room temperature case, which is better than the linear regression analysis. The hot model thermocouple sleeve design and fabrication are completed. The gasifier simulator (hot model) design and the fabrication are completed. The system tests of the gasifier simulator (hot model) have been conducted and some modifications have been made. Based on the system tests and results analysis, the gasifier simulator (hot model) has met the proposed design requirement and the ready for system test. The ultrasonic cleaning method is under evaluation and will be further studied for the gasifier simulator (hot model) application. The progress of this project has been on schedule.« less
Pattern Recognition Analysis of Age-Related Retinal Ganglion Cell Signatures in the Human Eye
Yoshioka, Nayuta; Zangerl, Barbara; Nivison-Smith, Lisa; Khuu, Sieu K.; Jones, Bryan W.; Pfeiffer, Rebecca L.; Marc, Robert E.; Kalloniatis, Michael
2017-01-01
Purpose To characterize macular ganglion cell layer (GCL) changes with age and provide a framework to assess changes in ocular disease. This study used data clustering to analyze macular GCL patterns from optical coherence tomography (OCT) in a large cohort of subjects without ocular disease. Methods Single eyes of 201 patients evaluated at the Centre for Eye Health (Sydney, Australia) were retrospectively enrolled (age range, 20–85); 8 × 8 grid locations obtained from Spectralis OCT macular scans were analyzed with unsupervised classification into statistically separable classes sharing common GCL thickness and change with age. The resulting classes and gridwise data were fitted with linear and segmented linear regression curves. Additionally, normalized data were analyzed to determine regression as a percentage. Accuracy of each model was examined through comparison of predicted 50-year-old equivalent macular GCL thickness for the entire cohort to a true 50-year-old reference cohort. Results Pattern recognition clustered GCL thickness across the macula into five to eight spatially concentric classes. F-test demonstrated segmented linear regression to be the most appropriate model for macular GCL change. The pattern recognition–derived and normalized model revealed less difference between the predicted macular GCL thickness and the reference cohort (average ± SD 0.19 ± 0.92 and −0.30 ± 0.61 μm) than a gridwise model (average ± SD 0.62 ± 1.43 μm). Conclusions Pattern recognition successfully identified statistically separable macular areas that undergo a segmented linear reduction with age. This regression model better predicted macular GCL thickness. The various unique spatial patterns revealed by pattern recognition combined with core GCL thickness data provide a framework to analyze GCL loss in ocular disease. PMID:28632847
Schwantes-An, Tae-Hwi; Sung, Heejong; Sabourin, Jeremy A; Justice, Cristina M; Sorant, Alexa J M; Wilson, Alexander F
2016-01-01
In this study, the effects of (a) the minor allele frequency of the single nucleotide variant (SNV), (b) the degree of departure from normality of the trait, and (c) the position of the SNVs on type I error rates were investigated in the Genetic Analysis Workshop (GAW) 19 whole exome sequence data. To test the distribution of the type I error rate, 5 simulated traits were considered: standard normal and gamma distributed traits; 2 transformed versions of the gamma trait (log 10 and rank-based inverse normal transformations); and trait Q1 provided by GAW 19. Each trait was tested with 313,340 SNVs. Tests of association were performed with simple linear regression and average type I error rates were determined for minor allele frequency classes. Rare SNVs (minor allele frequency < 0.05) showed inflated type I error rates for non-normally distributed traits that increased as the minor allele frequency decreased. The inflation of average type I error rates increased as the significance threshold decreased. Normally distributed traits did not show inflated type I error rates with respect to the minor allele frequency for rare SNVs. There was no consistent effect of transformation on the uniformity of the distribution of the location of SNVs with a type I error.
Reliability Analysis of the Gradual Degradation of Semiconductor Devices.
1983-07-20
under the heading of linear models or linear statistical models . 3 ,4 We have not used this material in this report. Assuming catastrophic failure when...assuming a catastrophic model . In this treatment we first modify our system loss formula and then proceed to the actual analysis. II. ANALYSIS OF...Failure Time 1 Ti Ti 2 T2 T2 n Tn n and are easily analyzed by simple linear regression. Since we have assumed a log normal/Arrhenius activation
Krasikova, Dina V; Le, Huy; Bachura, Eric
2018-06-01
To address a long-standing concern regarding a gap between organizational science and practice, scholars called for more intuitive and meaningful ways of communicating research results to users of academic research. In this article, we develop a common language effect size index (CLβ) that can help translate research results to practice. We demonstrate how CLβ can be computed and used to interpret the effects of continuous and categorical predictors in multiple linear regression models. We also elaborate on how the proposed CLβ index is computed and used to interpret interactions and nonlinear effects in regression models. In addition, we test the robustness of the proposed index to violations of normality and provide means for computing standard errors and constructing confidence intervals around its estimates. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
SU-F-T-130: [18F]-FDG Uptake Dose Response in Lung Correlates Linearly with Proton Therapy Dose
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, D; Titt, U; Mirkovic, D
2016-06-15
Purpose: Analysis of clinical outcomes in lung cancer patients treated with protons using 18F-FDG uptake in lung as a measure of dose response. Methods: A test case lung cancer patient was selected in an unbiased way. The test patient’s treatment planning and post treatment positron emission tomography (PET) were collected from picture archiving and communication system at the UT M.D. Anderson Cancer Center. Average computerized tomography scan was registered with post PET/CT through both rigid and deformable registrations for selected region of interest (ROI) via VelocityAI imaging informatics software. For the voxels in the ROI, a system that extracts themore » Standard Uptake Value (SUV) from PET was developed, and the corresponding relative biological effectiveness (RBE) weighted (both variable and constant) dose was computed using the Monte Carlo (MC) methods. The treatment planning system (TPS) dose was also obtained. Using histogram analysis, the voxel average normalized SUV vs. 3 different doses was obtained and linear regression fit was performed. Results: From the registration process, there were some regions that showed significant artifacts near the diaphragm and heart region, which yielded poor r-squared values when the linear regression fit was performed on normalized SUV vs. dose. Excluding these values, TPS fit yielded mean r-squared value of 0.79 (range 0.61–0.95), constant RBE fit yielded 0.79 (range 0.52–0.94), and variable RBE fit yielded 0.80 (range 0.52–0.94). Conclusion: A system that extracts SUV from PET to correlate between normalized SUV and various dose calculations was developed. A linear relation between normalized SUV and all three different doses was found.« less
USDA-ARS?s Scientific Manuscript database
Using linear regression models, we studied the main and two-way interaction effects of the predictor variables gender, age, BMI, and 64 folate/vitamin B-12/homocysteine/lipid/cholesterol-related single nucleotide polymorphisms (SNP) on log-transformed plasma homocysteine normalized by red blood cell...
The Weight of Euro Coins: Its Distribution Might Not Be as Normal as You Would Expect
ERIC Educational Resources Information Center
Shkedy, Ziv; Aerts, Marc; Callaert, Herman
2006-01-01
Classical regression models, ANOVA models and linear mixed models are just three examples (out of many) in which the normal distribution of the response is an essential assumption of the model. In this paper we use a dataset of 2000 euro coins containing information (up to the milligram) about the weight of each coin, to illustrate that the…
Iorgulescu, E; Voicu, V A; Sârbu, C; Tache, F; Albu, F; Medvedovici, A
2016-08-01
The influence of the experimental variability (instrumental repeatability, instrumental intermediate precision and sample preparation variability) and data pre-processing (normalization, peak alignment, background subtraction) on the discrimination power of multivariate data analysis methods (Principal Component Analysis -PCA- and Cluster Analysis -CA-) as well as a new algorithm based on linear regression was studied. Data used in the study were obtained through positive or negative ion monitoring electrospray mass spectrometry (+/-ESI/MS) and reversed phase liquid chromatography/UV spectrometric detection (RPLC/UV) applied to green tea extracts. Extractions in ethanol and heated water infusion were used as sample preparation procedures. The multivariate methods were directly applied to mass spectra and chromatograms, involving strictly a holistic comparison of shapes, without assignment of any structural identity to compounds. An alternative data interpretation based on linear regression analysis mutually applied to data series is also discussed. Slopes, intercepts and correlation coefficients produced by the linear regression analysis applied on pairs of very large experimental data series successfully retain information resulting from high frequency instrumental acquisition rates, obviously better defining the profiles being compared. Consequently, each type of sample or comparison between samples produces in the Cartesian space an ellipsoidal volume defined by the normal variation intervals of the slope, intercept and correlation coefficient. Distances between volumes graphically illustrates (dis)similarities between compared data. The instrumental intermediate precision had the major effect on the discrimination power of the multivariate data analysis methods. Mass spectra produced through ionization from liquid state in atmospheric pressure conditions of bulk complex mixtures resulting from extracted materials of natural origins provided an excellent data basis for multivariate analysis methods, equivalent to data resulting from chromatographic separations. The alternative evaluation of very large data series based on linear regression analysis produced information equivalent to results obtained through application of PCA an CA. Copyright © 2016 Elsevier B.V. All rights reserved.
Sandhu, Rupninder; Chollet-Hinton, Lynn; Kirk, Erin L; Midkiff, Bentley; Troester, Melissa A
2016-02-01
Complete age-related regression of mammary epithelium, often termed postmenopausal involution, is associated with decreased breast cancer risk. However, most studies have qualitatively assessed involution. We quantitatively analyzed epithelium, stroma, and adipose tissue from histologically normal breast tissue of 454 patients in the Normal Breast Study. High-resolution digital images of normal breast hematoxylin and eosin-stained slides were partitioned into epithelium, adipose tissue, and nonfatty stroma. Percentage area and nuclei per unit area (nuclear density) were calculated for each component. Quantitative data were evaluated in association with age using linear regression and cubic spline models. Stromal area decreased (P = 0.0002), and adipose tissue area increased (P < 0.0001), with an approximate 0.7% change in area for each component, until age 55 years when these area measures reached a steady state. Although epithelial area did not show linear changes with age, epithelial nuclear density decreased linearly beginning in the third decade of life. No significant age-related trends were observed for stromal or adipose nuclear density. Digital image analysis offers a high-throughput method for quantitatively measuring tissue morphometry and for objectively assessing age-related changes in adipose tissue, stroma, and epithelium. Epithelial nuclear density is a quantitative measure of age-related breast involution that begins to decline in the early premenopausal period. Copyright © 2015 Elsevier Inc. All rights reserved.
CAG repeat expansion in Huntington disease determines age at onset in a fully dominant fashion
Lee, J.-M.; Ramos, E.M.; Lee, J.-H.; Gillis, T.; Mysore, J.S.; Hayden, M.R.; Warby, S.C.; Morrison, P.; Nance, M.; Ross, C.A.; Margolis, R.L.; Squitieri, F.; Orobello, S.; Di Donato, S.; Gomez-Tortosa, E.; Ayuso, C.; Suchowersky, O.; Trent, R.J.A.; McCusker, E.; Novelletto, A.; Frontali, M.; Jones, R.; Ashizawa, T.; Frank, S.; Saint-Hilaire, M.H.; Hersch, S.M.; Rosas, H.D.; Lucente, D.; Harrison, M.B.; Zanko, A.; Abramson, R.K.; Marder, K.; Sequeiros, J.; Paulsen, J.S.; Landwehrmeyer, G.B.; Myers, R.H.; MacDonald, M.E.; Durr, Alexandra; Rosenblatt, Adam; Frati, Luigi; Perlman, Susan; Conneally, Patrick M.; Klimek, Mary Lou; Diggin, Melissa; Hadzi, Tiffany; Duckett, Ayana; Ahmed, Anwar; Allen, Paul; Ames, David; Anderson, Christine; Anderson, Karla; Anderson, Karen; Andrews, Thomasin; Ashburner, John; Axelson, Eric; Aylward, Elizabeth; Barker, Roger A.; Barth, Katrin; Barton, Stacey; Baynes, Kathleen; Bea, Alexandra; Beall, Erik; Beg, Mirza Faisal; Beglinger, Leigh J.; Biglan, Kevin; Bjork, Kristine; Blanchard, Steve; Bockholt, Jeremy; Bommu, Sudharshan Reddy; Brossman, Bradley; Burrows, Maggie; Calhoun, Vince; Carlozzi, Noelle; Chesire, Amy; Chiu, Edmond; Chua, Phyllis; Connell, R.J.; Connor, Carmela; Corey-Bloom, Jody; Craufurd, David; Cross, Stephen; Cysique, Lucette; Santos, Rachelle Dar; Davis, Jennifer; Decolongon, Joji; DiPietro, Anna; Doucette, Nicholas; Downing, Nancy; Dudler, Ann; Dunn, Steve; Ecker, Daniel; Epping, Eric A.; Erickson, Diane; Erwin, Cheryl; Evans, Ken; Factor, Stewart A.; Farias, Sarah; Fatas, Marta; Fiedorowicz, Jess; Fullam, Ruth; Furtado, Sarah; Garde, Monica Bascunana; Gehl, Carissa; Geschwind, Michael D.; Goh, Anita; Gooblar, Jon; Goodman, Anna; Griffith, Jane; Groves, Mark; Guttman, Mark; Hamilton, Joanne; Harrington, Deborah; Harris, Greg; Heaton, Robert K.; Helmer, Karl; Henneberry, Machelle; Hershey, Tamara; Herwig, Kelly; Howard, Elizabeth; Hunter, Christine; Jankovic, Joseph; Johnson, Hans; Johnson, Arik; Jones, Kathy; Juhl, Andrew; Kim, Eun Young; Kimble, Mycah; King, Pamela; Klimek, Mary Lou; Klöppel, Stefan; Koenig, Katherine; Komiti, Angela; Kumar, Rajeev; Langbehn, Douglas; Leavitt, Blair; Leserman, Anne; Lim, Kelvin; Lipe, Hillary; Lowe, Mark; Magnotta, Vincent A.; Mallonee, William M.; Mans, Nicole; Marietta, Jacquie; Marshall, Frederick; Martin, Wayne; Mason, Sarah; Matheson, Kirsty; Matson, Wayne; Mazzoni, Pietro; McDowell, William; Miedzybrodzka, Zosia; Miller, Michael; Mills, James; Miracle, Dawn; Montross, Kelsey; Moore, David; Mori, Sasumu; Moser, David J.; Moskowitz, Carol; Newman, Emily; Nopoulos, Peg; Novak, Marianne; O'Rourke, Justin; Oakes, David; Ondo, William; Orth, Michael; Panegyres, Peter; Pease, Karen; Perlman, Susan; Perlmutter, Joel; Peterson, Asa; Phillips, Michael; Pierson, Ron; Potkin, Steve; Preston, Joy; Quaid, Kimberly; Radtke, Dawn; Rae, Daniela; Rao, Stephen; Raymond, Lynn; Reading, Sarah; Ready, Rebecca; Reece, Christine; Reilmann, Ralf; Reynolds, Norm; Richardson, Kylie; Rickards, Hugh; Ro, Eunyoe; Robinson, Robert; Rodnitzky, Robert; Rogers, Ben; Rosenblatt, Adam; Rosser, Elisabeth; Rosser, Anne; Price, Kathy; Price, Kathy; Ryan, Pat; Salmon, David; Samii, Ali; Schumacher, Jamy; Schumacher, Jessica; Sendon, Jose Luis Lópenz; Shear, Paula; Sheinberg, Alanna; Shpritz, Barnett; Siedlecki, Karen; Simpson, Sheila A.; Singer, Adam; Smith, Jim; Smith, Megan; Smith, Glenn; Snyder, Pete; Song, Allen; Sran, Satwinder; Stephan, Klaas; Stober, Janice; Sü?muth, Sigurd; Suter, Greg; Tabrizi, Sarah; Tempkin, Terry; Testa, Claudia; Thompson, Sean; Thomsen, Teri; Thumma, Kelli; Toga, Arthur; Trautmann, Sonja; Tremont, Geoff; Turner, Jessica; Uc, Ergun; Vaccarino, Anthony; van Duijn, Eric; Van Walsem, Marleen; Vik, Stacie; Vonsattel, Jean Paul; Vuletich, Elizabeth; Warner, Tom; Wasserman, Paula; Wassink, Thomas; Waterman, Elijah; Weaver, Kurt; Weir, David; Welsh, Claire; Werling-Witkoske, Chris; Wesson, Melissa; Westervelt, Holly; Weydt, Patrick; Wheelock, Vicki; Williams, Kent; Williams, Janet; Wodarski, Mary; Wojcieszek, Joanne; Wood, Jessica; Wood-Siverio, Cathy; Wu, Shuhua; Yastrubetskaya, Olga; de Yebenes, Justo Garcia; Zhao, Yong Qiang; Zimbelman, Janice; Zschiegner, Roland; Aaserud, Olaf; Abbruzzese, Giovanni; Andrews, Thomasin; Andrich, Jurgin; Antczak, Jakub; Arran, Natalie; Artiga, Maria J. Saiz; Bachoud-Lévi, Anne-Catherine; Banaszkiewicz, Krysztof; di Poggio, Monica Bandettini; Bandmann, Oliver; Barbera, Miguel A.; Barker, Roger A.; Barrero, Francisco; Barth, Katrin; Bas, Jordi; Beister, Antoine; Bentivoglio, Anna Rita; Bertini, Elisabetta; Biunno, Ida; Bjørgo, Kathrine; Bjørnevoll, Inga; Bohlen, Stefan; Bonelli, Raphael M.; Bos, Reineke; Bourne, Colin; Bradbury, Alyson; Brockie, Peter; Brown, Felicity; Bruno, Stefania; Bryl, Anna; Buck, Andrea; Burg, Sabrina; Burgunder, Jean-Marc; Burns, Peter; Burrows, Liz; Busquets, Nuria; Busse, Monica; Calopa, Matilde; Carruesco, Gemma T.; Casado, Ana Gonzalez; Catena, Judit López; Chu, Carol; Ciesielska, Anna; Clapton, Jackie; Clayton, Carole; Clenaghan, Catherine; Coelho, Miguel; Connemann, Julia; Craufurd, David; Crooks, Jenny; Cubillo, Patricia Trigo; Cubo, Esther; Curtis, Adrienne; De Michele, Giuseppe; De Nicola, A.; de Souza, Jenny; de Weert, A. Marit; de Yébenes, Justo Garcia; Dekker, M.; Descals, A. Martínez; Di Maio, Luigi; Di Pietro, Anna; Dipple, Heather; Dose, Matthias; Dumas, Eve M.; Dunnett, Stephen; Ecker, Daniel; Elifani, F.; Ellison-Rose, Lynda; Elorza, Marina D.; Eschenbach, Carolin; Evans, Carole; Fairtlough, Helen; Fannemel, Madelein; Fasano, Alfonso; Fenollar, Maria; Ferrandes, Giovanna; Ferreira, Jaoquim J.; Fillingham, Kay; Finisterra, Ana Maria; Fisher, K.; Fletcher, Amy; Foster, Jillian; Foustanos, Isabella; Frech, Fernando A.; Fullam, Robert; Fullham, Ruth; Gago, Miguel; García, RocioGarcía-Ramos; García, Socorro S.; Garrett, Carolina; Gellera, Cinzia; Gill, Paul; Ginestroni, Andrea; Golding, Charlotte; Goodman, Anna; Gørvell, Per; Grant, Janet; Griguoli, A.; Gross, Diana; Guedes, Leonor; BascuñanaGuerra, Monica; Guerra, Maria Rosalia; Guerrero, Rosa; Guia, Dolores B.; Guidubaldi, Arianna; Hallam, Caroline; Hamer, Stephanie; Hammer, Kathrin; Handley, Olivia J.; Harding, Alison; Hasholt, Lis; Hedge, Reikha; Heiberg, Arvid; Heinicke, Walburgis; Held, Christine; Hernanz, Laura Casas; Herranhof, Briggitte; Herrera, Carmen Durán; Hidding, Ute; Hiivola, Heli; Hill, Susan; Hjermind, Lena. E.; Hobson, Emma; Hoffmann, Rainer; Holl, Anna Hödl; Howard, Liz; Hunt, Sarah; Huson, Susan; Ialongo, Tamara; Idiago, Jesus Miguel R.; Illmann, Torsten; Jachinska, Katarzyna; Jacopini, Gioia; Jakobsen, Oda; Jamieson, Stuart; Jamrozik, Zygmunt; Janik, Piotr; Johns, Nicola; Jones, Lesley; Jones, Una; Jurgens, Caroline K.; Kaelin, Alain; Kalbarczyk, Anna; Kershaw, Ann; Khalil, Hanan; Kieni, Janina; Klimberg, Aneta; Koivisto, Susana P.; Koppers, Kerstin; Kosinski, Christoph Michael; Krawczyk, Malgorzata; Kremer, Berry; Krysa, Wioletta; Kwiecinski, Hubert; Lahiri, Nayana; Lambeck, Johann; Lange, Herwig; Laver, Fiona; Leenders, K.L.; Levey, Jamie; Leythaeuser, Gabriele; Lezius, Franziska; Llesoy, Joan Roig; Löhle, Matthias; López, Cristobal Diez-Aja; Lorenza, Fortuna; Loria, Giovanna; Magnet, Markus; Mandich, Paola; Marchese, Roberta; Marcinkowski, Jerzy; Mariotti, Caterina; Mariscal, Natividad; Markova, Ivana; Marquard, Ralf; Martikainen, Kirsti; Martínez, Isabel Haro; Martínez-Descals, Asuncion; Martino, T.; Mason, Sarah; McKenzie, Sue; Mechi, Claudia; Mendes, Tiago; Mestre, Tiago; Middleton, Julia; Milkereit, Eva; Miller, Joanne; Miller, Julie; Minster, Sara; Möller, Jens Carsten; Monza, Daniela; Morales, Blas; Moreau, Laura V.; Moreno, Jose L. López-Sendón; Münchau, Alexander; Murch, Ann; Nielsen, Jørgen E.; Niess, Anke; Nørremølle, Anne; Novak, Marianne; O'Donovan, Kristy; Orth, Michael; Otti, Daniela; Owen, Michael; Padieu, Helene; Paganini, Marco; Painold, Annamaria; Päivärinta, Markku; Partington-Jones, Lucy; Paterski, Laurent; Paterson, Nicole; Patino, Dawn; Patton, Michael; Peinemann, Alexander; Peppa, Nadia; Perea, Maria Fuensanta Noguera; Peterson, Maria; Piacentini, Silvia; Piano, Carla; Càrdenas, Regina Pons i; Prehn, Christian; Price, Kathleen; Probst, Daniela; Quarrell, Oliver; Quiroga, Purificacion Pin; Raab, Tina; Rakowicz, Maryla; Raman, Ashok; Raymond, Lucy; Reilmann, Ralf; Reinante, Gema; Reisinger, Karin; Retterstol, Lars; Ribaï, Pascale; Riballo, Antonio V.; Ribas, Guillermo G.; Richter, Sven; Rickards, Hugh; Rinaldi, Carlo; Rissling, Ida; Ritchie, Stuart; Rivera, Susana Vázquez; Robert, Misericordia Floriach; Roca, Elvira; Romano, Silvia; Romoli, Anna Maria; Roos, Raymond A.C.; Røren, Niini; Rose, Sarah; Rosser, Elisabeth; Rosser, Anne; Rossi, Fabiana; Rothery, Jean; Rudzinska, Monika; Ruíz, Pedro J. García; Ruíz, Belan Garzon; Russo, Cinzia Valeria; Ryglewicz, Danuta; Saft, Carston; Salvatore, Elena; Sánchez, Vicenta; Sando, Sigrid Botne; Šašinková, Pavla; Sass, Christian; Scheibl, Monika; Schiefer, Johannes; Schlangen, Christiane; Schmidt, Simone; Schöggl, Helmut; Schrenk, Caroline; Schüpbach, Michael; Schuierer, Michele; Sebastián, Ana Rojo; Selimbegovic-Turkovic, Amina; Sempolowicz, Justyna; Silva, Mark; Sitek, Emilia; Slawek, Jaroslaw; Snowden, Julie; Soleti, Francesco; Soliveri, Paola; Sollom, Andrea; Soltan, Witold; Sorbi, Sandro; Sorensen, Sven Asger; Spadaro, Maria; Städtler, Michael; Stamm, Christiane; Steiner, Tanja; Stokholm, Jette; Stokke, Bodil; Stopford, Cheryl; Storch, Alexander; Straßburger, Katrin; Stubbe, Lars; Sulek, Anna; Szczudlik, Andrzej; Tabrizi, Sarah; Taylor, Rachel; Terol, Santiago Duran-Sindreu; Thomas, Gareth; Thompson, Jennifer; Thomson, Aileen; Tidswell, Katherine; Torres, Maria M. Antequera; Toscano, Jean; Townhill, Jenny; Trautmann, Sonja; Tucci, Tecla; Tuuha, Katri; Uhrova, Tereza; Valadas, Anabela; van Hout, Monique S.E.; van Oostrom, J.C.H.; van Vugt, Jeroen P.P.; vanm, Walsem Marleen R.; Vandenberghe, Wim; Verellen-Dumoulin, Christine; Vergara, Mar Ruiz; Verstappen, C.C.P.; Verstraelen, Nichola; Viladrich, Celia Mareca; Villanueva, Clara; Wahlström, Jan; Warner, Thomas; Wehus, Raghild; Weindl, Adolf; Werner, Cornelius J.; Westmoreland, Leann; Weydt, Patrick; Wiedemann, Alexandra; Wild, Edward; Wild, Sue; Witjes-Ané, Marie-Noelle; Witkowski, Grzegorz; Wójcik, Magdalena; Wolz, Martin; Wolz, Annett; Wright, Jan; Yardumian, Pam; Yates, Shona; Yudina, Elizaveta; Zaremba, Jacek; Zaugg, Sabine W.; Zdzienicka, Elzbieta; Zielonka, Daniel; Zielonka, Euginiusz; Zinzi, Paola; Zittel, Simone; Zucker, Birgrit; Adams, John; Agarwal, Pinky; Antonijevic, Irina; Beck, Christopher; Chiu, Edmond; Churchyard, Andrew; Colcher, Amy; Corey-Bloom, Jody; Dorsey, Ray; Drazinic, Carolyn; Dubinsky, Richard; Duff, Kevin; Factor, Stewart; Foroud, Tatiana; Furtado, Sarah; Giuliano, Joe; Greenamyre, Timothy; Higgins, Don; Jankovic, Joseph; Jennings, Dana; Kang, Un Jung; Kostyk, Sandra; Kumar, Rajeev; Leavitt, Blair; LeDoux, Mark; Mallonee, William; Marshall, Frederick; Mohlo, Eric; Morgan, John; Oakes, David; Panegyres, Peter; Panisset, Michel; Perlman, Susan; Perlmutter, Joel; Quaid, Kimberly; Raymond, Lynn; Revilla, Fredy; Robertson, Suzanne; Robottom, Bradley; Sanchez-Ramos, Juan; Scott, Burton; Shannon, Kathleen; Shoulson, Ira; Singer, Carlos; Tabbal, Samer; Testa, Claudia; van, Kammen Dan; Vetter, Louise; Walker, Francis; Warner, John; Weiner, illiam; Wheelock, Vicki; Yastrubetskaya, Olga; Barton, Stacey; Broyles, Janice; Clouse, Ronda; Coleman, Allison; Davis, Robert; Decolongon, Joji; DeLaRosa, Jeanene; Deuel, Lisa; Dietrich, Susan; Dubinsky, Hilary; Eaton, Ken; Erickson, Diane; Fitzpatrick, Mary Jane; Frucht, Steven; Gartner, Maureen; Goldstein, Jody; Griffith, Jane; Hickey, Charlyne; Hunt, Victoria; Jaglin, Jeana; Klimek, Mary Lou; Lindsay, Pat; Louis, Elan; Loy, Clemet; Lucarelli, Nancy; Malarick, Keith; Martin, Amanda; McInnis, Robert; Moskowitz, Carol; Muratori, Lisa; Nucifora, Frederick; O'Neill, Christine; Palao, Alicia; Peavy, Guerry; Quesada, Monica; Schmidt, Amy; Segro, Vicki; Sperin, Elaine; Suter, Greg; Tanev, Kalo; Tempkin, Teresa; Thiede, Curtis; Wasserman, Paula; Welsh, Claire; Wesson, Melissa; Zauber, Elizabeth
2012-01-01
Objective: Age at onset of diagnostic motor manifestations in Huntington disease (HD) is strongly correlated with an expanded CAG trinucleotide repeat. The length of the normal CAG repeat allele has been reported also to influence age at onset, in interaction with the expanded allele. Due to profound implications for disease mechanism and modification, we tested whether the normal allele, interaction between the expanded and normal alleles, or presence of a second expanded allele affects age at onset of HD motor signs. Methods: We modeled natural log-transformed age at onset as a function of CAG repeat lengths of expanded and normal alleles and their interaction by linear regression. Results: An apparently significant effect of interaction on age at motor onset among 4,068 subjects was dependent on a single outlier data point. A rigorous statistical analysis with a well-behaved dataset that conformed to the fundamental assumptions of linear regression (e.g., constant variance and normally distributed error) revealed significance only for the expanded CAG repeat, with no effect of the normal CAG repeat. Ten subjects with 2 expanded alleles showed an age at motor onset consistent with the length of the larger expanded allele. Conclusions: Normal allele CAG length, interaction between expanded and normal alleles, and presence of a second expanded allele do not influence age at onset of motor manifestations, indicating that the rate of HD pathogenesis leading to motor diagnosis is determined by a completely dominant action of the longest expanded allele and as yet unidentified genetic or environmental factors. Neurology® 2012;78:690–695 PMID:22323755
Browning of the landscape of interior Alaska based on 1986-2009 Landsat sensor NDVI
Rebecca A. Baird; David Verbyla; Teresa N. Hollingsworth
2012-01-01
We used a time series of 1986-2009 Landsat sensor data to compute the Normalized Difference Vegetation Index (NDVI) for 30 m pixels within the Bonanza Creek Experimental Forest of interior Alaska. Based on simple linear regression, we found significant (p
Fausti, S A; Olson, D J; Frey, R H; Henry, J A; Schaffer, H I; Phillips, D S
1995-01-01
The latency-intensity functions (LIFs) of ABRs elicited by high-frequency (8, 10, 12, and 14 kHz) toneburst stimuli were evaluated in 20 subjects with confirmed 'moderate' high-frequency sensorineural hearing loss. Wave V results from clicks and tonebursts revealed all intra- and intersession data to be reliable (p > 0.05). Linear regression curves were highly significant (p < or = 0.0001), indicating linear relationships for all stimuli analyzed. Comparisons between the linear regression curves from a previously reported normal-hearing subject group and this sensorineural hearing-impaired group showed no significant differences. This study demonstrated that tonebursts at 8, 10, and 12 kHz evoked ABRs which decreased in latency as a function of increasing intensity and that these LIFs were consistent and orderly (14 kHz was not determinable). These results will contribute information to facilitate the establishment of change criteria used to predict change in hearing during treatment with ototoxic medications.
NASA Astrophysics Data System (ADS)
Abunama, Taher; Othman, Faridah
2017-06-01
Analysing the fluctuations of wastewater inflow rates in sewage treatment plants (STPs) is essential to guarantee a sufficient treatment of wastewater before discharging it to the environment. The main objectives of this study are to statistically analyze and forecast the wastewater inflow rates into the Bandar Tun Razak STP in Kuala Lumpur, Malaysia. A time series analysis of three years’ weekly influent data (156weeks) has been conducted using the Auto-Regressive Integrated Moving Average (ARIMA) model. Various combinations of ARIMA orders (p, d, q) have been tried to select the most fitted model, which was utilized to forecast the wastewater inflow rates. The linear regression analysis was applied to testify the correlation between the observed and predicted influents. ARIMA (3, 1, 3) model was selected with the highest significance R-square and lowest normalized Bayesian Information Criterion (BIC) value, and accordingly the wastewater inflow rates were forecasted to additional 52weeks. The linear regression analysis between the observed and predicted values of the wastewater inflow rates showed a positive linear correlation with a coefficient of 0.831.
Demidenko, Eugene
2017-09-01
The exact density distribution of the nonlinear least squares estimator in the one-parameter regression model is derived in closed form and expressed through the cumulative distribution function of the standard normal variable. Several proposals to generalize this result are discussed. The exact density is extended to the estimating equation (EE) approach and the nonlinear regression with an arbitrary number of linear parameters and one intrinsically nonlinear parameter. For a very special nonlinear regression model, the derived density coincides with the distribution of the ratio of two normally distributed random variables previously obtained by Fieller (1932), unlike other approximations previously suggested by other authors. Approximations to the density of the EE estimators are discussed in the multivariate case. Numerical complications associated with the nonlinear least squares are illustrated, such as nonexistence and/or multiple solutions, as major factors contributing to poor density approximation. The nonlinear Markov-Gauss theorem is formulated based on the near exact EE density approximation.
Roldan-Valadez, Ernesto; Garcia-Ulloa, Ana Cristina; Gonzalez-Gutierrez, Omar; Martinez-Lopez, Manuel
2011-01-01
Computed-assisted three-dimensional data (3D) allows for an accurate evaluation of volumes compared with traditional measurements. An in vitro method comparison between geometric volume and 3D volumetry to obtain reference data for pituitary volumes in normal pituitary glands (PGs) and PGs containing adenomas. Prospective, transverse, analytical study. Forty-eight subjects underwent brain magnetic resonance imaging (MRI) with 3D sequencing for computer-aided volumetry. PG phantom volumes by both methods were compared. Using the best volumetric method, volumes of normal PGs and PGs with adenoma were compared. Statistical analysis used the Bland-Altman method, t-statistics, effect size and linear regression analysis. Method comparison between 3D volumetry and geometric volume revealed a lower bias and precision for 3D volumetry. A total of 27 patients exhibited normal PGs (mean age, 42.07 ± 16.17 years), although length, height, width, geometric volume and 3D volumetry were greater in women than in men. A total of 21 patients exhibited adenomas (mean age 39.62 ± 10.79 years), and length, height, width, geometric volume and 3D volumetry were greater in men than in women, with significant volumetric differences. Age did not influence pituitary volumes on linear regression analysis. Results from the present study showed that 3D volumetry was more accurate than the geometric method. In addition, the upper normal limits of PGs overlapped with lower volume limits during early stage microadenomas.
Artes, Paul H; Crabb, David P
2010-01-01
To investigate why the specificity of the Moorfields Regression Analysis (MRA) of the Heidelberg Retina Tomograph (HRT) varies with disc size, and to derive accurate normative limits for neuroretinal rim area to address this problem. Two datasets from healthy subjects (Manchester, UK, n = 88; Halifax, Nova Scotia, Canada, n = 75) were used to investigate the physiological relationship between the optic disc and neuroretinal rim area. Normative limits for rim area were derived by quantile regression (QR) and compared with those of the MRA (derived by linear regression). Logistic regression analyses were performed to quantify the association between disc size and positive classifications with the MRA, as well as with the QR-derived normative limits. In both datasets, the specificity of the MRA depended on optic disc size. The odds of observing a borderline or outside-normal-limits classification increased by approximately 10% for each 0.1 mm(2) increase in disc area (P < 0.1). The lower specificity of the MRA with large optic discs could be explained by the failure of linear regression to model the extremes of the rim area distribution (observations far from the mean). In comparison, the normative limits predicted by QR were larger for smaller discs (less specific, more sensitive), and smaller for larger discs, such that false-positive rates became independent of optic disc size. Normative limits derived by quantile regression appear to remove the size-dependence of specificity with the MRA. Because quantile regression does not rely on the restrictive assumptions of standard linear regression, it may be a more appropriate method for establishing normative limits in other clinical applications where the underlying distributions are nonnormal or have nonconstant variance.
Kilian, Reinhold; Matschinger, Herbert; Löeffler, Walter; Roick, Christiane; Angermeyer, Matthias C
2002-03-01
Transformation of the dependent cost variable is often used to solve the problems of heteroscedasticity and skewness in linear ordinary least square regression of health service cost data. However, transformation may cause difficulties in the interpretation of regression coefficients and the retransformation of predicted values. The study compares the advantages and disadvantages of different methods to estimate regression based cost functions using data on the annual costs of schizophrenia treatment. Annual costs of psychiatric service use and clinical and socio-demographic characteristics of the patients were assessed for a sample of 254 patients with a diagnosis of schizophrenia (ICD-10 F 20.0) living in Leipzig. The clinical characteristics of the participants were assessed by means of the BPRS 4.0, the GAF, and the CAN for service needs. Quality of life was measured by WHOQOL-BREF. A linear OLS regression model with non-parametric standard errors, a log-transformed OLS model and a generalized linear model with a log-link and a gamma distribution were used to estimate service costs. For the estimation of robust non-parametric standard errors, the variance estimator by White and a bootstrap estimator based on 2000 replications were employed. Models were evaluated by the comparison of the R2 and the root mean squared error (RMSE). RMSE of the log-transformed OLS model was computed with three different methods of bias-correction. The 95% confidence intervals for the differences between the RMSE were computed by means of bootstrapping. A split-sample-cross-validation procedure was used to forecast the costs for the one half of the sample on the basis of a regression equation computed for the other half of the sample. All three methods showed significant positive influences of psychiatric symptoms and met psychiatric service needs on service costs. Only the log- transformed OLS model showed a significant negative impact of age, and only the GLM shows a significant negative influences of employment status and partnership on costs. All three models provided a R2 of about.31. The Residuals of the linear OLS model revealed significant deviances from normality and homoscedasticity. The residuals of the log-transformed model are normally distributed but still heteroscedastic. The linear OLS model provided the lowest prediction error and the best forecast of the dependent cost variable. The log-transformed model provided the lowest RMSE if the heteroscedastic bias correction was used. The RMSE of the GLM with a log link and a gamma distribution was higher than those of the linear OLS model and the log-transformed OLS model. The difference between the RMSE of the linear OLS model and that of the log-transformed OLS model without bias correction was significant at the 95% level. As result of the cross-validation procedure, the linear OLS model provided the lowest RMSE followed by the log-transformed OLS model with a heteroscedastic bias correction. The GLM showed the weakest model fit again. None of the differences between the RMSE resulting form the cross- validation procedure were found to be significant. The comparison of the fit indices of the different regression models revealed that the linear OLS model provided a better fit than the log-transformed model and the GLM, but the differences between the models RMSE were not significant. Due to the small number of cases in the study the lack of significance does not sufficiently proof that the differences between the RSME for the different models are zero and the superiority of the linear OLS model can not be generalized. The lack of significant differences among the alternative estimators may reflect a lack of sample size adequate to detect important differences among the estimators employed. Further studies with larger case number are necessary to confirm the results. Specification of an adequate regression models requires a careful examination of the characteristics of the data. Estimation of standard errors and confidence intervals by nonparametric methods which are robust against deviations from the normal distribution and the homoscedasticity of residuals are suitable alternatives to the transformation of the skew distributed dependent variable. Further studies with more adequate case numbers are needed to confirm the results.
Tools for Basic Statistical Analysis
NASA Technical Reports Server (NTRS)
Luz, Paul L.
2005-01-01
Statistical Analysis Toolset is a collection of eight Microsoft Excel spreadsheet programs, each of which performs calculations pertaining to an aspect of statistical analysis. These programs present input and output data in user-friendly, menu-driven formats, with automatic execution. The following types of calculations are performed: Descriptive statistics are computed for a set of data x(i) (i = 1, 2, 3 . . . ) entered by the user. Normal Distribution Estimates will calculate the statistical value that corresponds to cumulative probability values, given a sample mean and standard deviation of the normal distribution. Normal Distribution from two Data Points will extend and generate a cumulative normal distribution for the user, given two data points and their associated probability values. Two programs perform two-way analysis of variance (ANOVA) with no replication or generalized ANOVA for two factors with four levels and three repetitions. Linear Regression-ANOVA will curvefit data to the linear equation y=f(x) and will do an ANOVA to check its significance.
Combined analysis of magnetic and gravity anomalies using normalized source strength (NSS)
NASA Astrophysics Data System (ADS)
Li, L.; Wu, Y.
2017-12-01
Gravity field and magnetic field belong to potential fields which lead inherent multi-solution. Combined analysis of magnetic and gravity anomalies based on Poisson's relation is used to determinate homology gravity and magnetic anomalies and decrease the ambiguity. The traditional combined analysis uses the linear regression of the reduction to pole (RTP) magnetic anomaly to the first order vertical derivative of the gravity anomaly, and provides the quantitative or semi-quantitative interpretation by calculating the correlation coefficient, slope and intercept. In the calculation process, due to the effect of remanent magnetization, the RTP anomaly still contains the effect of oblique magnetization. In this case the homology gravity and magnetic anomalies display irrelevant results in the linear regression calculation. The normalized source strength (NSS) can be transformed from the magnetic tensor matrix, which is insensitive to the remanence. Here we present a new combined analysis using NSS. Based on the Poisson's relation, the gravity tensor matrix can be transformed into the pseudomagnetic tensor matrix of the direction of geomagnetic field magnetization under the homologous condition. The NSS of pseudomagnetic tensor matrix and original magnetic tensor matrix are calculated and linear regression analysis is carried out. The calculated correlation coefficient, slope and intercept indicate the homology level, Poisson's ratio and the distribution of remanent respectively. We test the approach using synthetic model under complex magnetization, the results show that it can still distinguish the same source under the condition of strong remanence, and establish the Poisson's ratio. Finally, this approach is applied in China. The results demonstrated that our approach is feasible.
Can Functional Cardiac Age be Predicted from ECG in a Normal Healthy Population
NASA Technical Reports Server (NTRS)
Schlegel, Todd; Starc, Vito; Leban, Manja; Sinigoj, Petra; Vrhovec, Milos
2011-01-01
In a normal healthy population, we desired to determine the most age-dependent conventional and advanced ECG parameters. We hypothesized that changes in several ECG parameters might correlate with age and together reliably characterize the functional age of the heart. Methods: An initial study population of 313 apparently healthy subjects was ultimately reduced to 148 subjects (74 men, 84 women, in the range from 10 to 75 years of age) after exclusion criteria. In all subjects, ECG recordings (resting 5-minute 12-lead high frequency ECG) were evaluated via custom software programs to calculate up to 85 different conventional and advanced ECG parameters including beat-to-beat QT and RR variability, waveform complexity, and signal-averaged, high-frequency and spatial/spatiotemporal ECG parameters. The prediction of functional age was evaluated by multiple linear regression analysis using the best 5 univariate predictors. Results: Ignoring what were ultimately small differences between males and females, the functional age was found to be predicted (R2= 0.69, P < 0.001) from a linear combination of 5 independent variables: QRS elevation in the frontal plane (p<0.001), a new repolarization parameter QTcorr (p<0.001), mean high frequency QRS amplitude (p=0.009), the variability parameter % VLF of RRV (p=0.021) and the P-wave width (p=0.10). Here, QTcorr represents the correlation between the calculated QT and the measured QT signal. Conclusions: In apparently healthy subjects with normal conventional ECGs, functional cardiac age can be estimated by multiple linear regression analysis of mostly advanced ECG results. Because some parameters in the regression formula, such as QTcorr, high frequency QRS amplitude and P-wave width also change with disease in the same direction as with increased age, increased functional age of the heart may reflect subtle age-related pathologies in cardiac electrical function that are usually hidden on conventional ECG.
NASA Technical Reports Server (NTRS)
Mcgwire, K.; Friedl, M.; Estes, J. E.
1993-01-01
This article describes research related to sampling techniques for establishing linear relations between land surface parameters and remotely-sensed data. Predictive relations are estimated between percentage tree cover in a savanna environment and a normalized difference vegetation index (NDVI) derived from the Thematic Mapper sensor. Spatial autocorrelation in original measurements and regression residuals is examined using semi-variogram analysis at several spatial resolutions. Sampling schemes are then tested to examine the effects of autocorrelation on predictive linear models in cases of small sample sizes. Regression models between image and ground data are affected by the spatial resolution of analysis. Reducing the influence of spatial autocorrelation by enforcing minimum distances between samples may also improve empirical models which relate ground parameters to satellite data.
Testing a single regression coefficient in high dimensional linear models
Zhong, Ping-Shou; Li, Runze; Wang, Hansheng; Tsai, Chih-Ling
2017-01-01
In linear regression models with high dimensional data, the classical z-test (or t-test) for testing the significance of each single regression coefficient is no longer applicable. This is mainly because the number of covariates exceeds the sample size. In this paper, we propose a simple and novel alternative by introducing the Correlated Predictors Screening (CPS) method to control for predictors that are highly correlated with the target covariate. Accordingly, the classical ordinary least squares approach can be employed to estimate the regression coefficient associated with the target covariate. In addition, we demonstrate that the resulting estimator is consistent and asymptotically normal even if the random errors are heteroscedastic. This enables us to apply the z-test to assess the significance of each covariate. Based on the p-value obtained from testing the significance of each covariate, we further conduct multiple hypothesis testing by controlling the false discovery rate at the nominal level. Then, we show that the multiple hypothesis testing achieves consistent model selection. Simulation studies and empirical examples are presented to illustrate the finite sample performance and the usefulness of the proposed method, respectively. PMID:28663668
Testing a single regression coefficient in high dimensional linear models.
Lan, Wei; Zhong, Ping-Shou; Li, Runze; Wang, Hansheng; Tsai, Chih-Ling
2016-11-01
In linear regression models with high dimensional data, the classical z -test (or t -test) for testing the significance of each single regression coefficient is no longer applicable. This is mainly because the number of covariates exceeds the sample size. In this paper, we propose a simple and novel alternative by introducing the Correlated Predictors Screening (CPS) method to control for predictors that are highly correlated with the target covariate. Accordingly, the classical ordinary least squares approach can be employed to estimate the regression coefficient associated with the target covariate. In addition, we demonstrate that the resulting estimator is consistent and asymptotically normal even if the random errors are heteroscedastic. This enables us to apply the z -test to assess the significance of each covariate. Based on the p -value obtained from testing the significance of each covariate, we further conduct multiple hypothesis testing by controlling the false discovery rate at the nominal level. Then, we show that the multiple hypothesis testing achieves consistent model selection. Simulation studies and empirical examples are presented to illustrate the finite sample performance and the usefulness of the proposed method, respectively.
Normal reference values for bladder wall thickness on CT in a healthy population.
Fananapazir, Ghaneh; Kitich, Aleksandar; Lamba, Ramit; Stewart, Susan L; Corwin, Michael T
2018-02-01
To determine normal bladder wall thickness on CT in patients without bladder disease. Four hundred and nineteen patients presenting for trauma with normal CTs of the abdomen and pelvis were included in our retrospective study. Bladder wall thickness was assessed, and bladder volume was measured using both the ellipsoid formula and an automated technique. Patient age, gender, and body mass index were recorded. Linear regression models were created to account for bladder volume, age, gender, and body mass index, and the multiple correlation coefficient with bladder wall thickness was computed. Bladder volume and bladder wall thickness were log-transformed to achieve approximate normality and homogeneity of variance. Variables that did not contribute substantively to the model were excluded, and a parsimonious model was created and the multiple correlation coefficient was calculated. Expected bladder wall thickness was estimated for different bladder volumes, and 1.96 standard deviation above expected provided the upper limit of normal on the log scale. Age, gender, and bladder volume were associated with bladder wall thickness (p = 0.049, 0.024, and < 0.001, respectively). The linear regression model had an R 2 of 0.52. Age and gender were negligible in contribution to the model, and a parsimonious model using only volume was created for both the ellipsoid and automated volumes (R 2 = 0.52 and 0.51, respectively). Bladder wall thickness correlates with bladder wall volume. The study provides reference bladder wall thicknesses on CT utilizing both the ellipsoid formula and automated bladder volumes.
Islam, Md Rafiqul; Arslan, Iqbal; Attia, John; McEvoy, Mark; McElduff, Patrick; Basher, Ariful; Rahman, Waliur; Peel, Roseanne; Akhter, Ayesha; Akter, Shahnaz; Vashum, Khanrin P; Milton, Abul Hasnat
2013-01-01
To determine serum zinc level and other relevant biological markers in normal, prediabetic and diabetic individuals and their association with Homeostasis Model Assessment (HOMA) parameters. This cross-sectional study was conducted between March and December 2009. Any patient aged ≥ 30 years attending the medicine outpatient department of a medical university hospital in Dhaka, Bangladesh and who had a blood glucose level ordered by a physician was eligible to participate. A total of 280 participants were analysed. On fasting blood sugar results, 51% were normal, 13% had prediabetes and 36% had diabetes. Mean serum zinc level was lowest in prediabetic compared to normal and diabetic participants (mean differences were approximately 65 ppb/L and 33 ppb/L, respectively). In multiple linear regression, serum zinc level was found to be significantly lower in prediabetes than in those with normoglycemia. Beta cell function was significantly lower in prediabetes than normal participants. Adjusted linear regression for HOMA parameters did not show a statistically significant association between serum zinc level, beta cell function (P = 0.07) and insulin resistance (P = 0.08). Low serum zinc accentuated the increase in insulin resistance seen with increasing BMI. Participants with prediabetes have lower zinc levels than controls and zinc is significantly associated with beta cell function and insulin resistance. Further longitudinal population based studies are warranted and controlled trials would be valuable for establishing whether zinc supplementation in prediabetes could be a useful strategy in preventing progression to Type 2 diabetes.
Human Language Technology: Opportunities and Challenges
2005-01-01
because of the connections to and reliance on signal processing. Audio diarization critically includes indexing of speakers [12], since speaker ...to reduce inter- speaker variability in training. Standard techniques include vocal-tract length normalization, adaptation of acoustic models using...maximum likelihood linear regression (MLLR), and speaker -adaptive training based on MLLR. The acoustic models are mixtures of Gaussians, typically with
Federal Register 2010, 2011, 2012, 2013, 2014
2012-01-23
... monitors with missing data. Maximum recorded values are substituted for the missing data. The resulting... which the incomplete site is missing data. The linear regression relationship is based on time periods... between the monitors is used to fill in missing data for the incomplete monitor, so that the normal data...
Normal Theory Two-Stage ML Estimator When Data Are Missing at the Item Level
ERIC Educational Resources Information Center
Savalei, Victoria; Rhemtulla, Mijke
2017-01-01
In many modeling contexts, the variables in the model are linear composites of the raw items measured for each participant; for instance, regression and path analysis models rely on scale scores, and structural equation models often use parcels as indicators of latent constructs. Currently, no analytic estimation method exists to appropriately…
Helping Students Assess the Relative Importance of Different Intermolecular Interactions
ERIC Educational Resources Information Center
Jasien, Paul G.
2008-01-01
A semi-quantitative model has been developed to estimate the relative effects of dispersion, dipole-dipole interactions, and H-bonding on the normal boiling points ("T[subscript b]") for a subset of simple organic systems. The model is based upon a statistical analysis using multiple linear regression on a series of straight-chain organic…
Jones, Andrew M; Lomas, James; Moore, Peter T; Rice, Nigel
2016-10-01
We conduct a quasi-Monte-Carlo comparison of the recent developments in parametric and semiparametric regression methods for healthcare costs, both against each other and against standard practice. The population of English National Health Service hospital in-patient episodes for the financial year 2007-2008 (summed for each patient) is randomly divided into two equally sized subpopulations to form an estimation set and a validation set. Evaluating out-of-sample using the validation set, a conditional density approximation estimator shows considerable promise in forecasting conditional means, performing best for accuracy of forecasting and among the best four for bias and goodness of fit. The best performing model for bias is linear regression with square-root-transformed dependent variables, whereas a generalized linear model with square-root link function and Poisson distribution performs best in terms of goodness of fit. Commonly used models utilizing a log-link are shown to perform badly relative to other models considered in our comparison.
A systematic evaluation of normalization methods in quantitative label-free proteomics.
Välikangas, Tommi; Suomi, Tomi; Elo, Laura L
2018-01-01
To date, mass spectrometry (MS) data remain inherently biased as a result of reasons ranging from sample handling to differences caused by the instrumentation. Normalization is the process that aims to account for the bias and make samples more comparable. The selection of a proper normalization method is a pivotal task for the reliability of the downstream analysis and results. Many normalization methods commonly used in proteomics have been adapted from the DNA microarray techniques. Previous studies comparing normalization methods in proteomics have focused mainly on intragroup variation. In this study, several popular and widely used normalization methods representing different strategies in normalization are evaluated using three spike-in and one experimental mouse label-free proteomic data sets. The normalization methods are evaluated in terms of their ability to reduce variation between technical replicates, their effect on differential expression analysis and their effect on the estimation of logarithmic fold changes. Additionally, we examined whether normalizing the whole data globally or in segments for the differential expression analysis has an effect on the performance of the normalization methods. We found that variance stabilization normalization (Vsn) reduced variation the most between technical replicates in all examined data sets. Vsn also performed consistently well in the differential expression analysis. Linear regression normalization and local regression normalization performed also systematically well. Finally, we discuss the choice of a normalization method and some qualities of a suitable normalization method in the light of the results of our evaluation. © The Author 2016. Published by Oxford University Press.
Wilke, Marko
2018-02-01
This dataset contains the regression parameters derived by analyzing segmented brain MRI images (gray matter and white matter) from a large population of healthy subjects, using a multivariate adaptive regression splines approach. A total of 1919 MRI datasets ranging in age from 1-75 years from four publicly available datasets (NIH, C-MIND, fCONN, and IXI) were segmented using the CAT12 segmentation framework, writing out gray matter and white matter images normalized using an affine-only spatial normalization approach. These images were then subjected to a six-step DARTEL procedure, employing an iterative non-linear registration approach and yielding increasingly crisp intermediate images. The resulting six datasets per tissue class were then analyzed using multivariate adaptive regression splines, using the CerebroMatic toolbox. This approach allows for flexibly modelling smoothly varying trajectories while taking into account demographic (age, gender) as well as technical (field strength, data quality) predictors. The resulting regression parameters described here can be used to generate matched DARTEL or SHOOT templates for a given population under study, from infancy to old age. The dataset and the algorithm used to generate it are publicly available at https://irc.cchmc.org/software/cerebromatic.php.
Missing-value estimation using linear and non-linear regression with Bayesian gene selection.
Zhou, Xiaobo; Wang, Xiaodong; Dougherty, Edward R
2003-11-22
Data from microarray experiments are usually in the form of large matrices of expression levels of genes under different experimental conditions. Owing to various reasons, there are frequently missing values. Estimating these missing values is important because they affect downstream analysis, such as clustering, classification and network design. Several methods of missing-value estimation are in use. The problem has two parts: (1) selection of genes for estimation and (2) design of an estimation rule. We propose Bayesian variable selection to obtain genes to be used for estimation, and employ both linear and nonlinear regression for the estimation rule itself. Fast implementation issues for these methods are discussed, including the use of QR decomposition for parameter estimation. The proposed methods are tested on data sets arising from hereditary breast cancer and small round blue-cell tumors. The results compare very favorably with currently used methods based on the normalized root-mean-square error. The appendix is available from http://gspsnap.tamu.edu/gspweb/zxb/missing_zxb/ (user: gspweb; passwd: gsplab).
Bone mineral density and correlation factor analysis in normal Taiwanese children.
Shu, San-Ging
2007-01-01
Our aim was to establish reference data and linear regression equations for lumbar bone mineral density (BMD) in normal Taiwanese children. Several influencing factors of lumbar BMD were investigated. Two hundred fifty-seven healthy children were recruited from schools, 136 boys and 121 girls, aged 4-18 years were enrolled on a voluntary basis with written consent. Their height, weight, blood pressure, puberty stage, bone age and lumbar BMD (L2-4) by dual energy x-ray absorptiometry (DEXA) were measured. Data were analyzed using Pearson correlation and stepwise regression tests. All measurements increased with age. Prior to age 8, there was no gender difference. Parameters such as height, weight, and bone age (BA) in girls surpassed boys between ages 8-13 without statistical significance (p> or =0.05). This was reversed subsequently after age 14 in height (p<0.05). BMD difference had the same trend but was not statistically significant either. The influencing power of puberty stage and bone age over BMD was almost equal to or higher than that of height and weight. All the other factors correlated with BMD to variable powers. Multiple linear regression equations for boys and girls were formulated. BMD reference data is provided and can be used to monitor childhood pathological conditions. However, BMD in those with abnormal bone age or pubertal development could need modifications to ensure accuracy.
Handling nonnormality and variance heterogeneity for quantitative sublethal toxicity tests.
Ritz, Christian; Van der Vliet, Leana
2009-09-01
The advantages of using regression-based techniques to derive endpoints from environmental toxicity data are clear, and slowly, this superior analytical technique is gaining acceptance. As use of regression-based analysis becomes more widespread, some of the associated nuances and potential problems come into sharper focus. Looking at data sets that cover a broad spectrum of standard test species, we noticed that some model fits to data failed to meet two key assumptions-variance homogeneity and normality-that are necessary for correct statistical analysis via regression-based techniques. Failure to meet these assumptions often is caused by reduced variance at the concentrations showing severe adverse effects. Although commonly used with linear regression analysis, transformation of the response variable only is not appropriate when fitting data using nonlinear regression techniques. Through analysis of sample data sets, including Lemna minor, Eisenia andrei (terrestrial earthworm), and algae, we show that both the so-called Box-Cox transformation and use of the Poisson distribution can help to correct variance heterogeneity and nonnormality and so allow nonlinear regression analysis to be implemented. Both the Box-Cox transformation and the Poisson distribution can be readily implemented into existing protocols for statistical analysis. By correcting for nonnormality and variance heterogeneity, these two statistical tools can be used to encourage the transition to regression-based analysis and the depreciation of less-desirable and less-flexible analytical techniques, such as linear interpolation.
Discrimination of serum Raman spectroscopy between normal and colorectal cancer
NASA Astrophysics Data System (ADS)
Li, Xiaozhou; Yang, Tianyue; Yu, Ting; Li, Siqi
2011-07-01
Raman spectroscopy of tissues has been widely studied for the diagnosis of various cancers, but biofluids were seldom used as the analyte because of the low concentration. Herein, serum of 30 normal people, 46 colon cancer, and 44 rectum cancer patients were measured Raman spectra and analyzed. The information of Raman peaks (intensity and width) and that of the fluorescence background (baseline function coefficients) were selected as parameters for statistical analysis. Principal component regression (PCR) and partial least square regression (PLSR) were used on the selected parameters separately to see the performance of the parameters. PCR performed better than PLSR in our spectral data. Then linear discriminant analysis (LDA) was used on the principal components (PCs) of the two regression method on the selected parameters, and a diagnostic accuracy of 88% and 83% were obtained. The conclusion is that the selected features can maintain the information of original spectra well and Raman spectroscopy of serum has the potential for the diagnosis of colorectal cancer.
Yousefi, Siamak; Balasubramanian, Madhusudhanan; Goldbaum, Michael H; Medeiros, Felipe A; Zangwill, Linda M; Weinreb, Robert N; Liebmann, Jeffrey M; Girkin, Christopher A; Bowd, Christopher
2016-05-01
To validate Gaussian mixture-model with expectation maximization (GEM) and variational Bayesian independent component analysis mixture-models (VIM) for detecting glaucomatous progression along visual field (VF) defect patterns (GEM-progression of patterns (POP) and VIM-POP). To compare GEM-POP and VIM-POP with other methods. GEM and VIM models separated cross-sectional abnormal VFs from 859 eyes and normal VFs from 1117 eyes into abnormal and normal clusters. Clusters were decomposed into independent axes. The confidence limit (CL) of stability was established for each axis with a set of 84 stable eyes. Sensitivity for detecting progression was assessed in a sample of 83 eyes with known progressive glaucomatous optic neuropathy (PGON). Eyes were classified as progressed if any defect pattern progressed beyond the CL of stability. Performance of GEM-POP and VIM-POP was compared to point-wise linear regression (PLR), permutation analysis of PLR (PoPLR), and linear regression (LR) of mean deviation (MD), and visual field index (VFI). Sensitivity and specificity for detecting glaucomatous VFs were 89.9% and 93.8%, respectively, for GEM and 93.0% and 97.0%, respectively, for VIM. Receiver operating characteristic (ROC) curve areas for classifying progressed eyes were 0.82 for VIM-POP, 0.86 for GEM-POP, 0.81 for PoPLR, 0.69 for LR of MD, and 0.76 for LR of VFI. GEM-POP was significantly more sensitive to PGON than PoPLR and linear regression of MD and VFI in our sample, while providing localized progression information. Detection of glaucomatous progression can be improved by assessing longitudinal changes in localized patterns of glaucomatous defect identified by unsupervised machine learning.
NASA Astrophysics Data System (ADS)
Fernández-Manso, O.; Fernández-Manso, A.; Quintano, C.
2014-09-01
Aboveground biomass (AGB) estimation from optical satellite data is usually based on regression models of original or synthetic bands. To overcome the poor relation between AGB and spectral bands due to mixed-pixels when a medium spatial resolution sensor is considered, we propose to base the AGB estimation on fraction images from Linear Spectral Mixture Analysis (LSMA). Our study area is a managed Mediterranean pine woodland (Pinus pinaster Ait.) in central Spain. A total of 1033 circular field plots were used to estimate AGB from Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) optical data. We applied Pearson correlation statistics and stepwise multiple regression to identify suitable predictors from the set of variables of original bands, fraction imagery, Normalized Difference Vegetation Index and Tasselled Cap components. Four linear models and one nonlinear model were tested. A linear combination of ASTER band 2 (red, 0.630-0.690 μm), band 8 (short wave infrared 5, 2.295-2.365 μm) and green vegetation fraction (from LSMA) was the best AGB predictor (Radj2=0.632, the root-mean-squared error of estimated AGB was 13.3 Mg ha-1 (or 37.7%), resulting from cross-validation), rather than other combinations of the above cited independent variables. Results indicated that using ASTER fraction images in regression models improves the AGB estimation in Mediterranean pine forests. The spatial distribution of the estimated AGB, based on a multiple linear regression model, may be used as baseline information for forest managers in future studies, such as quantifying the regional carbon budget, fuel accumulation or monitoring of management practices.
Karabatsos, George
2017-02-01
Most of applied statistics involves regression analysis of data. In practice, it is important to specify a regression model that has minimal assumptions which are not violated by data, to ensure that statistical inferences from the model are informative and not misleading. This paper presents a stand-alone and menu-driven software package, Bayesian Regression: Nonparametric and Parametric Models, constructed from MATLAB Compiler. Currently, this package gives the user a choice from 83 Bayesian models for data analysis. They include 47 Bayesian nonparametric (BNP) infinite-mixture regression models; 5 BNP infinite-mixture models for density estimation; and 31 normal random effects models (HLMs), including normal linear models. Each of the 78 regression models handles either a continuous, binary, or ordinal dependent variable, and can handle multi-level (grouped) data. All 83 Bayesian models can handle the analysis of weighted observations (e.g., for meta-analysis), and the analysis of left-censored, right-censored, and/or interval-censored data. Each BNP infinite-mixture model has a mixture distribution assigned one of various BNP prior distributions, including priors defined by either the Dirichlet process, Pitman-Yor process (including the normalized stable process), beta (two-parameter) process, normalized inverse-Gaussian process, geometric weights prior, dependent Dirichlet process, or the dependent infinite-probits prior. The software user can mouse-click to select a Bayesian model and perform data analysis via Markov chain Monte Carlo (MCMC) sampling. After the sampling completes, the software automatically opens text output that reports MCMC-based estimates of the model's posterior distribution and model predictive fit to the data. Additional text and/or graphical output can be generated by mouse-clicking other menu options. This includes output of MCMC convergence analyses, and estimates of the model's posterior predictive distribution, for selected functionals and values of covariates. The software is illustrated through the BNP regression analysis of real data.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-07-23
... missing data. The linear regression relationship is based on time periods in which both monitors were... fill in missing data for the incomplete monitor, so that the normal data completeness requirement of 75 percent of data in each quarter of the three years is met. After the missing data for the site is filled...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Siewicki, T.C.; Chandler, G.T.
1995-12-31
Eastern oysters (Crassostrea virginica) were continuously exposed to suspended {sup 14}C-fluoranthene spiked-sediment for either: (1) five days followed by 24 days deputation, or (2) 28 days exposure. Sediment less than 63 um contained fluoranthene concentrations one or ten times that measured at suburbanized sites in southeastern estuaries (133 or 1,300 ng/g). The data were evaluated both raw and normalized for tissue lipid and sediment organic carbon concentrations. Uptake rate constants were estimated using non-linear regression methods. Depuration rate constants were estimated by linear regression of the deputation phase following five-days exposure and as the second partial derivative of the non-linearmore » regression for the 28-day exposures. Uptake and deputation rate constants, bioconcentration factors and half-lives were similar regardless of exposure time, sediment fluoranthene concentration or use of data normalization. Uptake and deputation rate constants, bioconcentration factors and half-lives (days) were similar and low for all experiments, ranging from 0.02 to 0.10, 0.14 to 0.30, 0.09 to 0.46, and 2.4 to 5.0, respectively. Degradation by the mixed function oxidase system is not expected in oysters allowing the use of radiotracers for measuring very low concentrations of fluoranthene. The results suggest that short-term exposures followed by deputation are effective for estimating kinetic rate constants and that normalization provides little benefit in these controlled studies. The results further show that bioconcentration of sediment-associated fluoranthene, and possibly other polycyclic aromatic hydrocarbons, is very low compared to either dissolved forms or levels commonly used in regulatory actions.« less
Dey, Jacob K; Ishii, Masaru; Boahene, Kofi D O; Byrne, Patrick J; Ishii, Lisa E
2014-01-01
Determine the effect of facial reanimation surgery on observer-graded attractiveness and negative facial perception of patients with facial paralysis. Randomized controlled experiment. Ninety observers viewed images of paralyzed faces, smiling and in repose, before and after reanimation surgery, as well as normal comparison faces. Observers rated the attractiveness of each face and characterized the paralyzed faces by rating severity, disfigured/bothersome, and importance to repair. Iterated factor analysis indicated these highly correlated variables measure a common domain, so they were combined to create the disfigured, important to repair, bothersome, severity (DIBS) factor score. Mixed effects linear regression determined the effect of facial reanimation surgery on attractiveness and DIBS score. Facial paralysis induces an attractiveness penalty of 2.51 on a 10-point scale for faces in repose and 3.38 for smiling faces. Mixed effects linear regression showed that reanimation surgery improved attractiveness for faces both in repose and smiling by 0.84 (95% confidence interval [CI]: 0.67, 1.01) and 1.24 (95% CI: 1.07, 1.42) respectively. Planned hypothesis tests confirmed statistically significant differences in attractiveness ratings between postoperative and normal faces, indicating attractiveness was not completely normalized. Regression analysis also showed that reanimation surgery decreased DIBS by 0.807 (95% CI: 0.704, 0.911) for faces in repose and 0.989 (95% CI: 0.886, 1.093), an entire standard deviation, for smiling faces. Facial reanimation surgery increases attractiveness and decreases negative facial perception of patients with facial paralysis. These data emphasize the need to optimize reanimation surgery to restore not only function, but also symmetry and cosmesis to improve facial perception and patient quality of life. © 2013 The American Laryngological, Rhinological and Otological Society, Inc.
NASA Astrophysics Data System (ADS)
Lockwood, M.; Owens, M. J.; Barnard, L.; Usoskin, I. G.
2016-11-01
We use sunspot-group observations from the Royal Greenwich Observatory (RGO) to investigate the effects of intercalibrating data from observers with different visual acuities. The tests are made by counting the number of groups [RB] above a variable cut-off threshold of observed total whole spot area (uncorrected for foreshortening) to simulate what a lower-acuity observer would have seen. The synthesised annual means of RB are then re-scaled to the full observed RGO group number [RA] using a variety of regression techniques. It is found that a very high correlation between RA and RB (r_{AB} > 0.98) does not prevent large errors in the intercalibration (for example sunspot-maximum values can be over 30 % too large even for such levels of r_{AB}). In generating the backbone sunspot number [R_{BB}], Svalgaard and Schatten ( Solar Phys., 2016) force regression fits to pass through the scatter-plot origin, which generates unreliable fits (the residuals do not form a normal distribution) and causes sunspot-cycle amplitudes to be exaggerated in the intercalibrated data. It is demonstrated that the use of Quantile-Quantile ("Q-Q") plots to test for a normal distribution is a useful indicator of erroneous and misleading regression fits. Ordinary least-squares linear fits, not forced to pass through the origin, are sometimes reliable (although the optimum method used is shown to be different when matching peak and average sunspot-group numbers). However, other fits are only reliable if non-linear regression is used. From these results it is entirely possible that the inflation of solar-cycle amplitudes in the backbone group sunspot number as one goes back in time, relative to related solar-terrestrial parameters, is entirely caused by the use of inappropriate and non-robust regression techniques to calibrate the sunspot data.
Prenatal Lead Exposure and Fetal Growth: Smaller Infants Have Heightened Susceptibility
Rodosthenous, Rodosthenis S.; Burris, Heather H.; Svensson, Katherine; Amarasiriwardena, Chitra J.; Cantoral, Alejandra; Schnaas, Lourdes; Mercado-García, Adriana; Coull, Brent A.; Wright, Robert O.; Téllez-Rojo, Martha M.; Baccarelli, Andrea A.
2016-01-01
Background As population lead levels decrease, the toxic effects of lead may be distributed to more sensitive populations, such as infants with poor fetal growth. Objectives To determine the association of prenatal lead exposure and fetal growth; and to evaluate whether infants with poor fetal growth are more susceptible to lead toxicity than those with normal fetal growth. Methods We examined the association of second trimester maternal blood lead levels (BLL) with birthweight-for-gestational age (BWGA) z-score in 944 mother-infant participants of the PROGRESS cohort. We determined the association between maternal BLL and BWGA z-score by using both linear and quantile regression. We estimated odds ratios for small-for-gestational age (SGA) infants between maternal BLL quartiles using logistic regression. Maternal age, body mass index, socioeconomic status, parity, household smoking exposure, hemoglobin levels, and infant sex were included as confounders. Results While linear regression showed a negative association between maternal BLL and BWGA z-score (β=−0.06 z-score units per log2 BLL increase; 95% CI: −0.13, 0.003; P=0.06), quantile regression revealed larger magnitudes of this association in the <30th percentiles of BWGA z-score (β range [−0.08, −0.13] z-score units per log2 BLL increase; all P values <0.05). Mothers in the highest BLL quartile had an odds ratio of 1.62 (95% CI: 0.99–2.65) for having a SGA infant compared to the lowest BLL quartile. Conclusions While both linear and quantile regression showed a negative association between prenatal lead exposure and birthweight, quantile regression revealed that smaller infants may represent a more susceptible subpopulation. PMID:27923585
Prenatal lead exposure and fetal growth: Smaller infants have heightened susceptibility.
Rodosthenous, Rodosthenis S; Burris, Heather H; Svensson, Katherine; Amarasiriwardena, Chitra J; Cantoral, Alejandra; Schnaas, Lourdes; Mercado-García, Adriana; Coull, Brent A; Wright, Robert O; Téllez-Rojo, Martha M; Baccarelli, Andrea A
2017-02-01
As population lead levels decrease, the toxic effects of lead may be distributed to more sensitive populations, such as infants with poor fetal growth. To determine the association of prenatal lead exposure and fetal growth; and to evaluate whether infants with poor fetal growth are more susceptible to lead toxicity than those with normal fetal growth. We examined the association of second trimester maternal blood lead levels (BLL) with birthweight-for-gestational age (BWGA) z-score in 944 mother-infant participants of the PROGRESS cohort. We determined the association between maternal BLL and BWGA z-score by using both linear and quantile regression. We estimated odds ratios for small-for-gestational age (SGA) infants between maternal BLL quartiles using logistic regression. Maternal age, body mass index, socioeconomic status, parity, household smoking exposure, hemoglobin levels, and infant sex were included as confounders. While linear regression showed a negative association between maternal BLL and BWGA z-score (β=-0.06 z-score units per log 2 BLL increase; 95% CI: -0.13, 0.003; P=0.06), quantile regression revealed larger magnitudes of this association in the <30th percentiles of BWGA z-score (β range [-0.08, -0.13] z-score units per log 2 BLL increase; all P values<0.05). Mothers in the highest BLL quartile had an odds ratio of 1.62 (95% CI: 0.99-2.65) for having a SGA infant compared to the lowest BLL quartile. While both linear and quantile regression showed a negative association between prenatal lead exposure and birthweight, quantile regression revealed that smaller infants may represent a more susceptible subpopulation. Copyright © 2016 Elsevier Ltd. All rights reserved.
Temperament affects sympathetic nervous function in a normal population.
Kim, Bora; Lee, Jae-Hon; Kang, Eun-Ho; Yu, Bum-Hee
2012-09-01
Although specific temperaments have been known to be related to autonomic nervous function in some psychiatric disorders, there are few studies that have examined the relationship between temperaments and autonomic nervous function in a normal population. In this study, we examined the effect of temperament on the sympathetic nervous function in a normal population. Sixty eight healthy subjects participated in the present study. Temperament was assessed using the Korean version of the Cloninger Temperament and Character Inventory (TCI). Autonomic nervous function was determined by measuring skin temperature in a resting state, which was recorded for 5 minutes from the palmar surface of the left 5th digit using a thermistor secured with a Velcro® band. Pearson's correlation analysis and multiple linear regression were used to examine the relationship between temperament and skin temperature. A higher harm avoidance score was correlated with a lower skin temperature (i.e. an increased sympathetic tone; r=-0.343, p=0.004) whereas a higher persistence score was correlated with a higher skin temperature (r=0.433, p=0.001). Hierarchical linear regression analysis revealed that harm avoidance was able to predict the variance of skin temperature independently, with a variance of 7.1% after controlling for sex, blood pressure and state anxiety and persistence was the factor predicting the variance of skin temperature with a variance of 5.0%. These results suggest that high harm avoidance is related to an increased sympathetic nervous function whereas high persistence is related to decreased sympathetic nervous function in a normal population.
Temperament Affects Sympathetic Nervous Function in a Normal Population
Kim, Bora; Lee, Jae-Hon; Kang, Eun-Ho
2012-01-01
Objective Although specific temperaments have been known to be related to autonomic nervous function in some psychiatric disorders, there are few studies that have examined the relationship between temperaments and autonomic nervous function in a normal population. In this study, we examined the effect of temperament on the sympathetic nervous function in a normal population. Methods Sixty eight healthy subjects participated in the present study. Temperament was assessed using the Korean version of the Cloninger Temperament and Character Inventory (TCI). Autonomic nervous function was determined by measuring skin temperature in a resting state, which was recorded for 5 minutes from the palmar surface of the left 5th digit using a thermistor secured with a Velcro® band. Pearson's correlation analysis and multiple linear regression were used to examine the relationship between temperament and skin temperature. Results A higher harm avoidance score was correlated with a lower skin temperature (i.e. an increased sympathetic tone; r=-0.343, p=0.004) whereas a higher persistence score was correlated with a higher skin temperature (r=0.433, p=0.001). Hierarchical linear regression analysis revealed that harm avoidance was able to predict the variance of skin temperature independently, with a variance of 7.1% after controlling for sex, blood pressure and state anxiety and persistence was the factor predicting the variance of skin temperature with a variance of 5.0%. Conclusion These results suggest that high harm avoidance is related to an increased sympathetic nervous function whereas high persistence is related to decreased sympathetic nervous function in a normal population. PMID:22993530
Islam, Md. Rafiqul; Arslan, Iqbal; Attia, John; McEvoy, Mark; McElduff, Patrick; Basher, Ariful; Rahman, Waliur; Peel, Roseanne; Akhter, Ayesha; Akter, Shahnaz; Vashum, Khanrin P.; Milton, Abul Hasnat
2013-01-01
Aims To determine serum zinc level and other relevant biological markers in normal, prediabetic and diabetic individuals and their association with Homeostasis Model Assessment (HOMA) parameters. Methods This cross-sectional study was conducted between March and December 2009. Any patient aged ≥30 years attending the medicine outpatient department of a medical university hospital in Dhaka, Bangladesh and who had a blood glucose level ordered by a physician was eligible to participate. Results A total of 280 participants were analysed. On fasting blood sugar results, 51% were normal, 13% had prediabetes and 36% had diabetes. Mean serum zinc level was lowest in prediabetic compared to normal and diabetic participants (mean differences were approximately 65 ppb/L and 33 ppb/L, respectively). In multiple linear regression, serum zinc level was found to be significantly lower in prediabetes than in those with normoglycemia. Beta cell function was significantly lower in prediabetes than normal participants. Adjusted linear regression for HOMA parameters did not show a statistically significant association between serum zinc level, beta cell function (P = 0.07) and insulin resistance (P = 0.08). Low serum zinc accentuated the increase in insulin resistance seen with increasing BMI. Conclusion Participants with prediabetes have lower zinc levels than controls and zinc is significantly associated with beta cell function and insulin resistance. Further longitudinal population based studies are warranted and controlled trials would be valuable for establishing whether zinc supplementation in prediabetes could be a useful strategy in preventing progression to Type 2 diabetes. PMID:23613929
Efficient Robust Regression via Two-Stage Generalized Empirical Likelihood
Bondell, Howard D.; Stefanski, Leonard A.
2013-01-01
Large- and finite-sample efficiency and resistance to outliers are the key goals of robust statistics. Although often not simultaneously attainable, we develop and study a linear regression estimator that comes close. Efficiency obtains from the estimator’s close connection to generalized empirical likelihood, and its favorable robustness properties are obtained by constraining the associated sum of (weighted) squared residuals. We prove maximum attainable finite-sample replacement breakdown point, and full asymptotic efficiency for normal errors. Simulation evidence shows that compared to existing robust regression estimators, the new estimator has relatively high efficiency for small sample sizes, and comparable outlier resistance. The estimator is further illustrated and compared to existing methods via application to a real data set with purported outliers. PMID:23976805
Relationships between use of television during meals and children's food consumption patterns.
Coon, K A; Goldberg, J; Rogers, B L; Tucker, K L
2001-01-01
We examined relationships between the presence of television during meals and children's food consumption patterns to test whether children's overall food consumption patterns, including foods not normally advertised, vary systematically with the extent to which television is part of normal mealtime routines. Ninety-one parent-child pairs from suburbs adjacent to Washington, DC, recruited via advertisements and word of mouth, participated. Children were in the fourth, fifth, or sixth grades. Socioeconomic data and information on television use were collected during survey interviews. Three nonconsecutive 24-hour dietary recalls, conducted with each child, were used to construct nutrient and food intake outcome variables. Independent sample t tests were used to compare mean food and nutrient intakes of children from families in which the television was usually on during 2 or more meals (n = 41) to those of children from families in which the television was either never on or only on during one meal (n = 50). Multiple linear regression models, controlling for socioeconomic factors and other covariates, were used to test strength of associations between television and children's consumption of food groups and nutrients. Children from families with high television use derived, on average, 6% more of their total daily energy intake from meats; 5% more from pizza, salty snacks, and soda; and nearly 5% less of their energy intake from fruits, vegetables, and juices than did children from families with low television use. Associations between television and children's consumption of food groups remained statistically significant in multiple linear regression models that controlled for socioeconomic factors and other covariates. Children from high television families derived less of their total energy from carbohydrate and consumed twice as much caffeine as children from low television families. There continued to be a significant association between television and children's consumption of caffeine when these relationships were tested in multiple linear regression models. The dietary patterns of children from families in which television viewing is a normal part of meal routines may include fewer fruits and vegetables and more pizzas, snack foods, and sodas than the dietary patterns of children from families in which television viewing and eating are separate activities.
Mameli, Chiara; Krakauer, Nir Y; Krakauer, Jesse C; Bosetti, Alessandra; Ferrari, Chiara Matilde; Moiana, Norma; Schneider, Laura; Borsani, Barbara; Genoni, Teresa; Zuccotti, Gianvincenzo
2018-01-01
A Body Shape Index (ABSI) and normalized hip circumference (Hip Index, HI) have been recently shown to be strong risk factors for mortality and for cardiovascular disease in adults. We conducted an observational cross-sectional study to evaluate the relationship between ABSI, HI and cardiometabolic risk factors and obesity-related comorbidities in overweight and obese children and adolescents aged 2-18 years. We performed multivariate linear and logistic regression analyses with BMI, ABSI, and HI age and sex normalized z scores as predictors to examine the association with cardiometabolic risk markers (systolic and diastolic blood pressure, fasting glucose and insulin, total cholesterol and its components, transaminases, fat mass % detected by bioelectrical impedance analysis) and obesity-related conditions (including hepatic steatosis and metabolic syndrome). We recruited 217 patients (114 males), mean age 11.3 years. Multivariate linear regression showed a significant association of ABSI z score with 10 out of 15 risk markers expressed as continuous variables, while BMI z score showed a significant correlation with 9 and HI only with 1. In multivariate logistic regression to predict occurrence of obesity-related conditions and above-threshold values of risk factors, BMI z score was significantly correlated to 7 out of 12, ABSI to 5, and HI to 1. Overall, ABSI is an independent anthropometric index that was significantly associated with cardiometabolic risk markers in a pediatric population affected by overweight and obesity.
Advanced statistics: linear regression, part I: simple linear regression.
Marill, Keith A
2004-01-01
Simple linear regression is a mathematical technique used to model the relationship between a single independent predictor variable and a single dependent outcome variable. In this, the first of a two-part series exploring concepts in linear regression analysis, the four fundamental assumptions and the mechanics of simple linear regression are reviewed. The most common technique used to derive the regression line, the method of least squares, is described. The reader will be acquainted with other important concepts in simple linear regression, including: variable transformations, dummy variables, relationship to inference testing, and leverage. Simplified clinical examples with small datasets and graphic models are used to illustrate the points. This will provide a foundation for the second article in this series: a discussion of multiple linear regression, in which there are multiple predictor variables.
Simultaneous multiple non-crossing quantile regression estimation using kernel constraints
Liu, Yufeng; Wu, Yichao
2011-01-01
Quantile regression (QR) is a very useful statistical tool for learning the relationship between the response variable and covariates. For many applications, one often needs to estimate multiple conditional quantile functions of the response variable given covariates. Although one can estimate multiple quantiles separately, it is of great interest to estimate them simultaneously. One advantage of simultaneous estimation is that multiple quantiles can share strength among them to gain better estimation accuracy than individually estimated quantile functions. Another important advantage of joint estimation is the feasibility of incorporating simultaneous non-crossing constraints of QR functions. In this paper, we propose a new kernel-based multiple QR estimation technique, namely simultaneous non-crossing quantile regression (SNQR). We use kernel representations for QR functions and apply constraints on the kernel coefficients to avoid crossing. Both unregularised and regularised SNQR techniques are considered. Asymptotic properties such as asymptotic normality of linear SNQR and oracle properties of the sparse linear SNQR are developed. Our numerical results demonstrate the competitive performance of our SNQR over the original individual QR estimation. PMID:22190842
NASA Astrophysics Data System (ADS)
Zhang, L.; Han, X. X.; Ge, J.; Wang, C. H.
2018-01-01
To determine the relationship between compressive strength and flexural strength of pavement geopolymer grouting material, 20 groups of geopolymer grouting materials were prepared, the compressive strength and flexural strength were determined by mechanical properties test. On the basis of excluding the abnormal values through boxplot, the results show that, the compressive strength test results were normal, but there were two mild outliers in 7days flexural strength test. The compressive strength and flexural strength were linearly fitted by SPSS, six regression models were obtained by linear fitting of compressive strength and flexural strength. The linear relationship between compressive strength and flexural strength can be better expressed by the cubic curve model, and the correlation coefficient was 0.842.
Sonographic Measurement of Fetal Ear Length in Turkish Women with a Normal Pregnancy
Özdemir, Mucize Eriç; Uzun, Işıl; Karahasanoğlu, Ayşe; Aygün, Mehmet; Akın, Hale; Yazıcıoğlu, Fehmi
2014-01-01
Background: Abnormal fetal ear length is a feature of chromosomal disorders. Fetal ear length measurement is a simple measurement that can be obtained during ultrasonographic examinations. Aims: To develop a nomogram for fetal ear length measurements in our population and investigate the correlation between fetal ear length, gestational age, and other standard fetal biometric measurements. Study Design: Cohort study. Methods: Ear lengths of the fetuses were measured in normal singleton pregnancies. The relationship between gestational age and fetal ear length in millimetres was analysed by simple linear regression. In addition, the correlation of fetal ear length measurements with biparietal diameter, head circumference, abdominal circumference, and femur length were evaluated.Ear length measurements were obtained from fetuses in 389 normal singleton pregnancies ranging between 16 and 28 weeks of gestation. Results: A nomogram was developed by linear regression analysis of the parameters ear length and gestational age. Fetal ear length (mm) = y = (1.348 X gestational age)−12.265), where gestational ages is in weeks. A high correlation was found between fetal ear length and gestational age, and a significant correlation was also found between fetal ear length and the biparietal diameter (r=0.962; p<0.001). Similar correlations were found between fetal ear length and head circumference, and fetal ear length and femur length. Conclusion: The results of this study provide a nomogram for fetal ear length. The study also demonstrates the relationship between ear length and other biometric measurements. PMID:25667783
Stenset, V; Hofoss, D; Johnsen, L; Skinningsrud, A; Berstad, A E; Negaard, A; Reinvang, I; Gjerstad, L; Fladby, T
2008-12-01
To identify possible associations between white matter lesions (WML) and cognition in patients with memory complaints, stratified in groups with normal and low cerebrospinal fluid (CSF) Abeta42 values. 215 consecutive patients with subjective memory complaints were retrospectively included. Patients were stratified into two groups with normal (n = 127) or low (n = 88) CSF Abeta42 levels (cut-off is 450 ng/l). Cognitive scores from the Mini-Mental State Examination (MMSE) and the Neurobehavioral Cognitive Status Examination (Cognistat) were used as continuous dependent variables in linear regression. WML load was used as a continuous independent variable and was scored with a visual rating scale. The regression model was corrected for possible confounding factors. WML were significantly associated with MMSE and all Cognistat subscores except language (repetition and naming) and attention in patients with normal CSF Abeta42 levels. No significant associations were observed in patients with low CSF Abeta42. WML were associated with affection of multiple cognitive domains, including delayed recall and executive functions, in patients with normal CSF Abeta42 levels. The lack of such associations for patients with low CSF Abeta42 (i.e. with evidence for amyloid deposition), suggests that amyloid pathology may obscure cognitive effects of WML.
Catalog of Air Force Weather Technical Documents, 1941-2006
2006-05-19
radiosondes in current use in USA. Elementary discussion of statistical terms and concepts used for expressing accuracy or error is discussed. AWS TR 105...Techniques, Appendix B: Vorticity—An Elementary Discussion of the Concept, August 1956, 27pp. Formerly AWSM 105– 50/1A. Provides the necessary back...steps involved in ordinary multiple linear regression. Conditional probability is calculated using transnormalized variables in the multivariate normal
Pang, Haowen; Sun, Xiaoyang; Yang, Bo; Wu, Jingbo
2018-05-01
To ensure good quality intensity-modulated radiation therapy (IMRT) planning, we proposed the use of a quality control method based on generalized equivalent uniform dose (gEUD) that predicts absorbed radiation doses in organs at risk (OAR). We conducted a retrospective analysis of patients who underwent IMRT for the treatment of cervical carcinoma, nasopharyngeal carcinoma (NPC), or non-small cell lung cancer (NSCLC). IMRT plans were randomly divided into data acquisition and data verification groups. OAR in the data acquisition group for cervical carcinoma and NPC were further classified as sub-organs at risk (sOAR). The normalized volume of sOAR and normalized gEUD (a = 1) were analyzed using multiple linear regression to establish a fitting formula. For NSCLC, the normalized intersection volume of the planning target volume (PTV) and lung, the maximum diameter of the PTV (left-right, anterior-posterior, and superior-inferior), and the normalized gEUD (a = 1) were analyzed using multiple linear regression to establish a fitting formula for the lung gEUD (a = 1). The r-squared and P values indicated that the fitting formula was a good fit. In the data verification group, IMRT plans verified the accuracy of the fitting formula, and compared the gEUD (a = 1) for each OAR between the subjective method and the gEUD-based method. In conclusion, the gEUD-based method can be used effectively for quality control and can reduce the influence of subjective factors on IMRT planning optimization. © 2018 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.
Müller, Christian; Schillert, Arne; Röthemeier, Caroline; Trégouët, David-Alexandre; Proust, Carole; Binder, Harald; Pfeiffer, Norbert; Beutel, Manfred; Lackner, Karl J.; Schnabel, Renate B.; Tiret, Laurence; Wild, Philipp S.; Blankenberg, Stefan
2016-01-01
Technical variation plays an important role in microarray-based gene expression studies, and batch effects explain a large proportion of this noise. It is therefore mandatory to eliminate technical variation while maintaining biological variability. Several strategies have been proposed for the removal of batch effects, although they have not been evaluated in large-scale longitudinal gene expression data. In this study, we aimed at identifying a suitable method for batch effect removal in a large study of microarray-based longitudinal gene expression. Monocytic gene expression was measured in 1092 participants of the Gutenberg Health Study at baseline and 5-year follow up. Replicates of selected samples were measured at both time points to identify technical variability. Deming regression, Passing-Bablok regression, linear mixed models, non-linear models as well as ReplicateRUV and ComBat were applied to eliminate batch effects between replicates. In a second step, quantile normalization prior to batch effect correction was performed for each method. Technical variation between batches was evaluated by principal component analysis. Associations between body mass index and transcriptomes were calculated before and after batch removal. Results from association analyses were compared to evaluate maintenance of biological variability. Quantile normalization, separately performed in each batch, combined with ComBat successfully reduced batch effects and maintained biological variability. ReplicateRUV performed perfectly in the replicate data subset of the study, but failed when applied to all samples. All other methods did not substantially reduce batch effects in the replicate data subset. Quantile normalization plus ComBat appears to be a valuable approach for batch correction in longitudinal gene expression data. PMID:27272489
Log-normal frailty models fitted as Poisson generalized linear mixed models.
Hirsch, Katharina; Wienke, Andreas; Kuss, Oliver
2016-12-01
The equivalence of a survival model with a piecewise constant baseline hazard function and a Poisson regression model has been known since decades. As shown in recent studies, this equivalence carries over to clustered survival data: A frailty model with a log-normal frailty term can be interpreted and estimated as a generalized linear mixed model with a binary response, a Poisson likelihood, and a specific offset. Proceeding this way, statistical theory and software for generalized linear mixed models are readily available for fitting frailty models. This gain in flexibility comes at the small price of (1) having to fix the number of pieces for the baseline hazard in advance and (2) having to "explode" the data set by the number of pieces. In this paper we extend the simulations of former studies by using a more realistic baseline hazard (Gompertz) and by comparing the model under consideration with competing models. Furthermore, the SAS macro %PCFrailty is introduced to apply the Poisson generalized linear mixed approach to frailty models. The simulations show good results for the shared frailty model. Our new %PCFrailty macro provides proper estimates, especially in case of 4 events per piece. The suggested Poisson generalized linear mixed approach for log-normal frailty models based on the %PCFrailty macro provides several advantages in the analysis of clustered survival data with respect to more flexible modelling of fixed and random effects, exact (in the sense of non-approximate) maximum likelihood estimation, and standard errors and different types of confidence intervals for all variance parameters. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Lin, Lei; Wang, Qian; Sadek, Adel W
2016-06-01
The duration of freeway traffic accidents duration is an important factor, which affects traffic congestion, environmental pollution, and secondary accidents. Among previous studies, the M5P algorithm has been shown to be an effective tool for predicting incident duration. M5P builds a tree-based model, like the traditional classification and regression tree (CART) method, but with multiple linear regression models as its leaves. The problem with M5P for accident duration prediction, however, is that whereas linear regression assumes that the conditional distribution of accident durations is normally distributed, the distribution for a "time-to-an-event" is almost certainly nonsymmetrical. A hazard-based duration model (HBDM) is a better choice for this kind of a "time-to-event" modeling scenario, and given this, HBDMs have been previously applied to analyze and predict traffic accidents duration. Previous research, however, has not yet applied HBDMs for accident duration prediction, in association with clustering or classification of the dataset to minimize data heterogeneity. The current paper proposes a novel approach for accident duration prediction, which improves on the original M5P tree algorithm through the construction of a M5P-HBDM model, in which the leaves of the M5P tree model are HBDMs instead of linear regression models. Such a model offers the advantage of minimizing data heterogeneity through dataset classification, and avoids the need for the incorrect assumption of normality for traffic accident durations. The proposed model was then tested on two freeway accident datasets. For each dataset, the first 500 records were used to train the following three models: (1) an M5P tree; (2) a HBDM; and (3) the proposed M5P-HBDM, and the remainder of data were used for testing. The results show that the proposed M5P-HBDM managed to identify more significant and meaningful variables than either M5P or HBDMs. Moreover, the M5P-HBDM had the lowest overall mean absolute percentage error (MAPE). Copyright © 2016 Elsevier Ltd. All rights reserved.
Analysis of regression methods for solar activity forecasting
NASA Technical Reports Server (NTRS)
Lundquist, C. A.; Vaughan, W. W.
1979-01-01
The paper deals with the potential use of the most recent solar data to project trends in the next few years. Assuming that a mode of solar influence on weather can be identified, advantageous use of that knowledge presumably depends on estimating future solar activity. A frequently used technique for solar cycle predictions is a linear regression procedure along the lines formulated by McNish and Lincoln (1949). The paper presents a sensitivity analysis of the behavior of such regression methods relative to the following aspects: cycle minimum, time into cycle, composition of historical data base, and unnormalized vs. normalized solar cycle data. Comparative solar cycle forecasts for several past cycles are presented as to these aspects of the input data. Implications for the current cycle, No. 21, are also given.
Correlation of Respirator Fit Measured on Human Subjects and a Static Advanced Headform
Bergman, Michael S.; He, Xinjian; Joseph, Michael E.; Zhuang, Ziqing; Heimbuch, Brian K.; Shaffer, Ronald E.; Choe, Melanie; Wander, Joseph D.
2015-01-01
This study assessed the correlation of N95 filtering face-piece respirator (FFR) fit between a Static Advanced Headform (StAH) and 10 human test subjects. Quantitative fit evaluations were performed on test subjects who made three visits to the laboratory. On each visit, one fit evaluation was performed on eight different FFRs of various model/size variations. Additionally, subject breathing patterns were recorded. Each fit evaluation comprised three two-minute exercises: “Normal Breathing,” “Deep Breathing,” and again “Normal Breathing.” The overall test fit factors (FF) for human tests were recorded. The same respirator samples were later mounted on the StAH and the overall test manikin fit factors (MFF) were assessed utilizing the recorded human breathing patterns. Linear regression was performed on the mean log10-transformed FF and MFF values to assess the relationship between the values obtained from humans and the StAH. This is the first study to report a positive correlation of respirator fit between a headform and test subjects. The linear regression by respirator resulted in R2 = 0.95, indicating a strong linear correlation between FF and MFF. For all respirators the geometric mean (GM) FF values were consistently higher than those of the GM MFF. For 50% of respirators, GM FF and GM MFF values were significantly different between humans and the StAH. For data grouped by subject/respirator combinations, the linear regression resulted in R2 = 0.49. A weaker correlation (R2 = 0.11) was found using only data paired by subject/respirator combination where both the test subject and StAH had passed a real-time leak check before performing the fit evaluation. For six respirators, the difference in passing rates between the StAH and humans was < 20%, while two respirators showed a difference of 29% and 43%. For data by test subject, GM FF and GM MFF values were significantly different for 40% of the subjects. Overall, the advanced headform system has potential for assessing fit for some N95 FFR model/sizes. PMID:25265037
Correlation and simple linear regression.
Eberly, Lynn E
2007-01-01
This chapter highlights important steps in using correlation and simple linear regression to address scientific questions about the association of two continuous variables with each other. These steps include estimation and inference, assessing model fit, the connection between regression and ANOVA, and study design. Examples in microbiology are used throughout. This chapter provides a framework that is helpful in understanding more complex statistical techniques, such as multiple linear regression, linear mixed effects models, logistic regression, and proportional hazards regression.
Sieve estimation of Cox models with latent structures.
Cao, Yongxiu; Huang, Jian; Liu, Yanyan; Zhao, Xingqiu
2016-12-01
This article considers sieve estimation in the Cox model with an unknown regression structure based on right-censored data. We propose a semiparametric pursuit method to simultaneously identify and estimate linear and nonparametric covariate effects based on B-spline expansions through a penalized group selection method with concave penalties. We show that the estimators of the linear effects and the nonparametric component are consistent. Furthermore, we establish the asymptotic normality of the estimator of the linear effects. To compute the proposed estimators, we develop a modified blockwise majorization descent algorithm that is efficient and easy to implement. Simulation studies demonstrate that the proposed method performs well in finite sample situations. We also use the primary biliary cirrhosis data to illustrate its application. © 2016, The International Biometric Society.
Estimating Blade Section Airloads from Blade Leading-Edge Pressure Measurements
NASA Technical Reports Server (NTRS)
vanAken, Johannes M.
2003-01-01
The Tilt-Rotor Aeroacoustic Model (TRAM) test in the Duitse-Nederlandse Wind (DNW) Tunnel acquired blade pressure data for forward flight test conditions of a tiltrotor in helicopter mode. Chordwise pressure data at seven radial locations were integrated to obtain the blade section normal force. The present investigation evaluates the use of linear regression analysis and of neural networks in estimating the blade section normal force coefficient from a limited number of blade leading-edge pressure measurements and representative operating conditions. These network models are subsequently used to estimate the airloads at intermediate radial locations where only blade pressure measurements at the 3.5% chordwise stations are available.
Crytzer, Theresa M; Keramati, Mariam; Anthony, Steven J; Cheng, Yu-Ting; Robertson, Robert J; Dicianno, Brad E
2018-02-03
People with spina bifida (SB) face personal and environmental barriers to exercise that contribute to physical inactivity, obesity, risk of cardiovascular disease, and poor aerobic fitness. The WHEEL rating of perceived exertion (RPE) Scale was validated in people with SB to monitor exercise intensity. However, the psycho-physiological link between RPE and ventilatory breakpoint (Vpt), the group-normalized perceptual response, has not been determined and would provide a starting point for aerobic exercise in this cohort. The primary objectives were to determine the group-normalized RPE equivalent to Vpt based on WHEEL and Borg Scale ratings and to develop a regression model to predict Borg Scale (conditional metric) from WHEEL Scale (criterion metric). The secondary objective was to create a table of interchangeable values between WHEEL and Borg Scale RPE for people with SB performing a load incremental stress test. Cross-sectional observational. University laboratory. Twenty-nine participants with SB. Participants completed a load incremented arm ergometer exercise stress test. WHEEL and Borg Scale ratings were recorded the last 15 seconds of each 1-minute test phase. WHEEL and Borg Scale ratings, metabolic measures (eg, oxygen consumption, carbon dioxide production). Determined Vpt via plots of oxygen consumption and carbon dioxide production against time. Nineteen of 29 participants achieved Vpt (Group A). The mean ± standard deviation peak oxygen consumption at Vpt for Group A was 61.76 ± 16.26. The WHEEL and Borg Scale RPE at Vpt were 5.74 ± 2.58 (range 0-10) and 13.95 ± 3.50 (range 6-19), respectively. A significant linear regression model was developed (Borg Scale rating = 1.22 × WHEEL Scale rating + 7.14) and used to create a WHEEL-to-Borg Scale RPE conversion table. A significant linear regression model and table of interchangeable values was developed for participants with SB. The group-normalized RPE (WHEEL, 5.74; Borg, 13.95) can be used to prescribe and self-regulate arm ergometer exercise intensity approximating the Vpt. II. Copyright © 2018. Published by Elsevier Inc.
Pérez-Rodríguez, Paulino; Gianola, Daniel; González-Camacho, Juan Manuel; Crossa, José; Manès, Yann; Dreisigacker, Susanne
2012-01-01
In genome-enabled prediction, parametric, semi-parametric, and non-parametric regression models have been used. This study assessed the predictive ability of linear and non-linear models using dense molecular markers. The linear models were linear on marker effects and included the Bayesian LASSO, Bayesian ridge regression, Bayes A, and Bayes B. The non-linear models (this refers to non-linearity on markers) were reproducing kernel Hilbert space (RKHS) regression, Bayesian regularized neural networks (BRNN), and radial basis function neural networks (RBFNN). These statistical models were compared using 306 elite wheat lines from CIMMYT genotyped with 1717 diversity array technology (DArT) markers and two traits, days to heading (DTH) and grain yield (GY), measured in each of 12 environments. It was found that the three non-linear models had better overall prediction accuracy than the linear regression specification. Results showed a consistent superiority of RKHS and RBFNN over the Bayesian LASSO, Bayesian ridge regression, Bayes A, and Bayes B models. PMID:23275882
Pérez-Rodríguez, Paulino; Gianola, Daniel; González-Camacho, Juan Manuel; Crossa, José; Manès, Yann; Dreisigacker, Susanne
2012-12-01
In genome-enabled prediction, parametric, semi-parametric, and non-parametric regression models have been used. This study assessed the predictive ability of linear and non-linear models using dense molecular markers. The linear models were linear on marker effects and included the Bayesian LASSO, Bayesian ridge regression, Bayes A, and Bayes B. The non-linear models (this refers to non-linearity on markers) were reproducing kernel Hilbert space (RKHS) regression, Bayesian regularized neural networks (BRNN), and radial basis function neural networks (RBFNN). These statistical models were compared using 306 elite wheat lines from CIMMYT genotyped with 1717 diversity array technology (DArT) markers and two traits, days to heading (DTH) and grain yield (GY), measured in each of 12 environments. It was found that the three non-linear models had better overall prediction accuracy than the linear regression specification. Results showed a consistent superiority of RKHS and RBFNN over the Bayesian LASSO, Bayesian ridge regression, Bayes A, and Bayes B models.
El Beltagi, Tarek A; Bowd, Christopher; Boden, Catherine; Amini, Payam; Sample, Pamela A; Zangwill, Linda M; Weinreb, Robert N
2003-11-01
To determine the relationship between areas of glaucomatous retinal nerve fiber layer thinning identified by optical coherence tomography and areas of decreased visual field sensitivity identified by standard automated perimetry in glaucomatous eyes. Retrospective observational case series. Forty-three patients with glaucomatous optic neuropathy identified by optic disc stereo photographs and standard automated perimetry mean deviations >-8 dB were included. Participants were imaged with optical coherence tomography within 6 months of reliable standard automated perimetry testing. The location and number of optical coherence tomography clock hour retinal nerve fiber layer thickness measures outside normal limits were compared with the location and number of standard automated perimetry visual field zones outside normal limits. Further, the relationship between the deviation from normal optical coherence tomography-measured retinal nerve fiber layer thickness at each clock hour and the average pattern deviation in each visual field zone was examined by using linear regression (R(2)). The retinal nerve fiber layer areas most frequently outside normal limits were the inferior and inferior temporal regions. The least sensitive visual field zones were in the superior hemifield. Linear regression results (R(2)) showed that deviation from the normal retinal nerve fiber layer thickness at optical coherence tomography clock hour positions 6 o'clock, 7 o'clock, and 8 o'clock (inferior and inferior temporal) was best correlated with standard automated perimetry pattern deviation in visual field zones corresponding to the superior arcuate and nasal step regions (R(2) range, 0.34-0.57). These associations were much stronger than those between clock hour position 6 o'clock and the visual field zone corresponding to the inferior nasal step region (R(2) = 0.01). Localized retinal nerve fiber layer thinning, measured by optical coherence tomography, is topographically related to decreased localized standard automated perimetry sensitivity in glaucoma patients.
Noninvasive and fast measurement of blood glucose in vivo by near infrared (NIR) spectroscopy
NASA Astrophysics Data System (ADS)
Jintao, Xue; Liming, Ye; Yufei, Liu; Chunyan, Li; Han, Chen
2017-05-01
This research was to develop a method for noninvasive and fast blood glucose assay in vivo. Near-infrared (NIR) spectroscopy, a more promising technique compared to other methods, was investigated in rats with diabetes and normal rats. Calibration models are generated by two different multivariate strategies: partial least squares (PLS) as linear regression method and artificial neural networks (ANN) as non-linear regression method. The PLS model was optimized individually by considering spectral range, spectral pretreatment methods and number of model factors, while the ANN model was studied individually by selecting spectral pretreatment methods, parameters of network topology, number of hidden neurons, and times of epoch. The results of the validation showed the two models were robust, accurate and repeatable. Compared to the ANN model, the performance of the PLS model was much better, with lower root mean square error of validation (RMSEP) of 0.419 and higher correlation coefficients (R) of 96.22%.
Wang, Hai-Mei; Li, Zheng-Hai; Wang, Zhen
2013-01-01
Based on the monthly temperature and precipitation data of 15 meteorological stations and the statistical data of livestock density in Xilinguole League in 1981-2007, and by using ArcGIS, this paper analyzed the spatial distribution of the climate aridity and livestock density in the League, and in combining with the ten-day data of the normalized difference vegetation index (NDVI) in 1981-2007, the driving factors of the vegetation cover change in the League were discussed. In the study period, there was a satisfactory linear regression relationship between the climate aridity and the vegetation coverage. The NDVI and the livestock density had a favorable binomial regression relationship. With the increase of NDVI, the livestock density increased first and decreased then. The vegetation coverage had a complex linear relationship with livestock density and climate aridity. The NDVI had a positive correlation with climate aridity, but a negative correlation with livestock density. Compared with livestock density, climate aridity had far greater effects on the NDVI.
Johnston, J L; Leong, M S; Checkland, E G; Zuberbuhler, P C; Conger, P R; Quinney, H A
1988-12-01
Body density and skinfold thickness at four sites were measured in 140 normal boys, 168 normal girls, and 6 boys and 7 girls with cystic fibrosis, all aged 8-14 y. Prediction equations for the normal boys and girls for the estimation of body-fat content from skinfold measurements were derived from linear regression of body density vs the log of the sum of the skinfold thickness. The relationship between body density and the log of the sum of the skinfold measurements differed from normal for the boys and girls with cystic fibrosis because of their high body density even though their large residual volume was corrected for. However the sum of skinfold measurements in the children with cystic fibrosis did not differ from normal. Thus body fat percent of these children with cystic fibrosis was underestimated when calculated from body density and invalid when calculated from skinfold thickness.
Lee, Seung-Mi; Choi, In-Sun; Han, Euna; Suh, David; Shin, Eun-Kyung; Je, Seyunghe; Lee, Sung Su; Suh, Dong-Churl
2018-01-01
This study aimed to estimate treatment costs attributable to overweight and obesity in patients with diabetes who were less than 65 years of age in the United States. This study used data from the Medical Expenditure Panel Survey from 2001 to 2013. Patients with diabetes were identified by using the International Classification of Diseases, Ninth Revision, Clinical Modification code (250), clinical classification codes (049 and 050), or self-reported physician diagnoses. Total treatment costs attributable to overweight and obesity were calculated as the differences in the adjusted costs compared with individuals with diabetes and normal weight. Adjusted costs were estimated by using generalized linear models or unconditional quantile regression models. The mean annual treatment costs attributable to obesity were $1,852 higher than those attributable to normal weight, while costs attributable to overweight were $133 higher. The unconditional quantile regression results indicated that the impact of obesity on total treatment costs gradually became more significant as treatment costs approached the upper quantile. Among patients with diabetes who were less than 65 years of age, patients with diabetes and obesity have significantly higher treatment costs than patients with diabetes and normal weight. The economic burden of diabetes to society will continue to increase unless more proactive preventive measures are taken to effectively treat patients with overweight or obesity. © 2017 The Obesity Society.
Raizes, Meytal; Elkana, Odelia; Franko, Motty; Ravona Springer, Ramit; Segev, Shlomo; Beeri, Michal Schnaider
2016-01-01
We explored the association of plasma glucose levels within the normal range with processing speed in high functioning young elderly, free of type 2 diabetes mellitus (T2DM). A sample of 41 participants (mean age = 64.7, SD = 10; glucose 94.5 mg/dL, SD = 9.3), were examined with a computerized cognitive battery. Hierarchical linear regression analysis showed that higher plasma glucose levels, albeit within the normal range (<110 mg/dL), were associated with longer reaction times (p < 0.01). These findings suggest that even in the subclinical range and in the absence of T2DM, monitoring plasma glucose levels may have an impact on cognitive function.
Marginal regression analysis of recurrent events with coarsened censoring times.
Hu, X Joan; Rosychuk, Rhonda J
2016-12-01
Motivated by an ongoing pediatric mental health care (PMHC) study, this article presents weakly structured methods for analyzing doubly censored recurrent event data where only coarsened information on censoring is available. The study extracted administrative records of emergency department visits from provincial health administrative databases. The available information of each individual subject is limited to a subject-specific time window determined up to concealed data. To evaluate time-dependent effect of exposures, we adapt the local linear estimation with right censored survival times under the Cox regression model with time-varying coefficients (cf. Cai and Sun, Scandinavian Journal of Statistics 2003, 30, 93-111). We establish the pointwise consistency and asymptotic normality of the regression parameter estimator, and examine its performance by simulation. The PMHC study illustrates the proposed approach throughout the article. © 2016, The International Biometric Society.
Linear Regression with a Randomly Censored Covariate: Application to an Alzheimer's Study.
Atem, Folefac D; Qian, Jing; Maye, Jacqueline E; Johnson, Keith A; Betensky, Rebecca A
2017-01-01
The association between maternal age of onset of dementia and amyloid deposition (measured by in vivo positron emission tomography (PET) imaging) in cognitively normal older offspring is of interest. In a regression model for amyloid, special methods are required due to the random right censoring of the covariate of maternal age of onset of dementia. Prior literature has proposed methods to address the problem of censoring due to assay limit of detection, but not random censoring. We propose imputation methods and a survival regression method that do not require parametric assumptions about the distribution of the censored covariate. Existing imputation methods address missing covariates, but not right censored covariates. In simulation studies, we compare these methods to the simple, but inefficient complete case analysis, and to thresholding approaches. We apply the methods to the Alzheimer's study.
Daily magnesium intake and serum magnesium concentration among Japanese people.
Akizawa, Yoriko; Koizumi, Sadayuki; Itokawa, Yoshinori; Ojima, Toshiyuki; Nakamura, Yosikazu; Tamura, Tarou; Kusaka, Yukinori
2008-01-01
The vitamins and minerals that are deficient in the daily diet of a normal adult remain unknown. To answer this question, we conducted a population survey focusing on the relationship between dietary magnesium intake and serum magnesium level. The subjects were 62 individuals from Fukui Prefecture who participated in the 1998 National Nutrition Survey. The survey investigated the physical status, nutritional status, and dietary data of the subjects. Holidays and special occasions were avoided, and a day when people are most likely to be on an ordinary diet was selected as the survey date. The mean (+/-standard deviation) daily magnesium intake was 322 (+/-132), 323 (+/-163), and 322 (+/-147) mg/day for men, women, and the entire group, respectively. The mean (+/-standard deviation) serum magnesium concentration was 20.69 (+/-2.83), 20.69 (+/-2.88), and 20.69 (+/-2.83) ppm for men, women, and the entire group, respectively. The distribution of serum magnesium concentration was normal. Dietary magnesium intake showed a log-normal distribution, which was then transformed by logarithmic conversion for examining the regression coefficients. The slope of the regression line between the serum magnesium concentration (Y ppm) and daily magnesium intake (X mg) was determined using the formula Y = 4.93 (log(10)X) + 8.49. The coefficient of correlation (r) was 0.29. A regression line (Y = 14.65X + 19.31) was observed between the daily intake of magnesium (Y mg) and serum magnesium concentration (X ppm). The coefficient of correlation was 0.28. The daily magnesium intake correlated with serum magnesium concentration, and a linear regression model between them was proposed.
On estimation of linear transformation models with nested case–control sampling
Liu, Mengling
2011-01-01
Nested case–control (NCC) sampling is widely used in large epidemiological cohort studies for its cost effectiveness, but its data analysis primarily relies on the Cox proportional hazards model. In this paper, we consider a family of linear transformation models for analyzing NCC data and propose an inverse selection probability weighted estimating equation method for inference. Consistency and asymptotic normality of our estimators for regression coefficients are established. We show that the asymptotic variance has a closed analytic form and can be easily estimated. Numerical studies are conducted to support the theory and an application to the Wilms’ Tumor Study is also given to illustrate the methodology. PMID:21912975
Liu, Fei; Ye, Lanhan; Peng, Jiyu; Song, Kunlin; Shen, Tingting; Zhang, Chu; He, Yong
2018-02-27
Fast detection of heavy metals is very important for ensuring the quality and safety of crops. Laser-induced breakdown spectroscopy (LIBS), coupled with uni- and multivariate analysis, was applied for quantitative analysis of copper in three kinds of rice (Jiangsu rice, regular rice, and Simiao rice). For univariate analysis, three pre-processing methods were applied to reduce fluctuations, including background normalization, the internal standard method, and the standard normal variate (SNV). Linear regression models showed a strong correlation between spectral intensity and Cu content, with an R 2 more than 0.97. The limit of detection (LOD) was around 5 ppm, lower than the tolerance limit of copper in foods. For multivariate analysis, partial least squares regression (PLSR) showed its advantage in extracting effective information for prediction, and its sensitivity reached 1.95 ppm, while support vector machine regression (SVMR) performed better in both calibration and prediction sets, where R c 2 and R p 2 reached 0.9979 and 0.9879, respectively. This study showed that LIBS could be considered as a constructive tool for the quantification of copper contamination in rice.
Ye, Lanhan; Song, Kunlin; Shen, Tingting
2018-01-01
Fast detection of heavy metals is very important for ensuring the quality and safety of crops. Laser-induced breakdown spectroscopy (LIBS), coupled with uni- and multivariate analysis, was applied for quantitative analysis of copper in three kinds of rice (Jiangsu rice, regular rice, and Simiao rice). For univariate analysis, three pre-processing methods were applied to reduce fluctuations, including background normalization, the internal standard method, and the standard normal variate (SNV). Linear regression models showed a strong correlation between spectral intensity and Cu content, with an R2 more than 0.97. The limit of detection (LOD) was around 5 ppm, lower than the tolerance limit of copper in foods. For multivariate analysis, partial least squares regression (PLSR) showed its advantage in extracting effective information for prediction, and its sensitivity reached 1.95 ppm, while support vector machine regression (SVMR) performed better in both calibration and prediction sets, where Rc2 and Rp2 reached 0.9979 and 0.9879, respectively. This study showed that LIBS could be considered as a constructive tool for the quantification of copper contamination in rice. PMID:29495445
Kumar, K Vasanth
2007-04-02
Kinetic experiments were carried out for the sorption of safranin onto activated carbon particles. The kinetic data were fitted to pseudo-second order model of Ho, Sobkowsk and Czerwinski, Blanchard et al. and Ritchie by linear and non-linear regression methods. Non-linear method was found to be a better way of obtaining the parameters involved in the second order rate kinetic expressions. Both linear and non-linear regression showed that the Sobkowsk and Czerwinski and Ritchie's pseudo-second order models were the same. Non-linear regression analysis showed that both Blanchard et al. and Ho have similar ideas on the pseudo-second order model but with different assumptions. The best fit of experimental data in Ho's pseudo-second order expression by linear and non-linear regression method showed that Ho pseudo-second order model was a better kinetic expression when compared to other pseudo-second order kinetic expressions.
Gómez Navarro, Rafael
2009-01-01
To study the renal function (FR) of the hypertensive patients by means of estimating equations and serum creatinine (Crp). To calculate the percentage of patients with chronic kidney disease (ERC) that present normal values of Crp. To analyze which factors collaborate in the deterioration of the FR. Descriptive cross-sectional study of patients with HTA. Crp and arterial tension (TA) were determined. The glomerular filtration rate was calculated by means of Cockroft-Gault and MDRD's formula. The years of evolution of the HTA were registered. A descriptive study of the variables and the possible dependence among them was completed, using several times linear multiple regression. 52 patients were studied (57,7% women). Average age 72,4 +/- 10,8. 32,6% (Cockcroft-Gault) or 21,5% (MDRD) were fulfilling ERC criterion. The ERC was mainly diagnosed in females. 21,4% (Cockcroft-Gault) and 9,5 % patients (MDRD) with ERC had normal Crp values. We do not find linear dependence between the numbers of TA and the FR. The TA check-up objectives do not suppose less development of ERC. In males we find linear dependence within the FR (MDRD) and the years of evolution of the HTA. The ERC is a frequent pathology in the hypertense persons. The systematical utilization of estimating equations facilitates the detection of hidden ERC in patients with normal Crp.
The problem of natural funnel asymmetries: a simulation analysis of meta-analysis in macroeconomics.
Callot, Laurent; Paldam, Martin
2011-06-01
Effect sizes in macroeconomic are estimated by regressions on data published by statistical agencies. Funnel plots are a representation of the distribution of the resulting regression coefficients. They are normally much wider than predicted by the t-ratio of the coefficients and often asymmetric. The standard method of meta-analysts in economics assumes that the asymmetries are because of publication bias causing censoring and adjusts the average accordingly. The paper shows that some funnel asymmetries may be 'natural' so that they occur without censoring. We investigate such asymmetries by simulating funnels by pairs of data generating processes (DGPs) and estimating models (EMs), in which the EM has the problem that it disregards a property of the DGP. The problems are data dependency, structural breaks, non-normal residuals, non-linearity, and omitted variables. We show that some of these problems generate funnel asymmetries. When they do, the standard method often fails. Copyright © 2011 John Wiley & Sons, Ltd. Copyright © 2011 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Kargoll, Boris; Omidalizarandi, Mohammad; Loth, Ina; Paffenholz, Jens-André; Alkhatib, Hamza
2018-03-01
In this paper, we investigate a linear regression time series model of possibly outlier-afflicted observations and autocorrelated random deviations. This colored noise is represented by a covariance-stationary autoregressive (AR) process, in which the independent error components follow a scaled (Student's) t-distribution. This error model allows for the stochastic modeling of multiple outliers and for an adaptive robust maximum likelihood (ML) estimation of the unknown regression and AR coefficients, the scale parameter, and the degree of freedom of the t-distribution. This approach is meant to be an extension of known estimators, which tend to focus only on the regression model, or on the AR error model, or on normally distributed errors. For the purpose of ML estimation, we derive an expectation conditional maximization either algorithm, which leads to an easy-to-implement version of iteratively reweighted least squares. The estimation performance of the algorithm is evaluated via Monte Carlo simulations for a Fourier as well as a spline model in connection with AR colored noise models of different orders and with three different sampling distributions generating the white noise components. We apply the algorithm to a vibration dataset recorded by a high-accuracy, single-axis accelerometer, focusing on the evaluation of the estimated AR colored noise model.
Woo, John H; Wang, Sumei; Melhem, Elias R; Gee, James C; Cucchiara, Andrew; McCluskey, Leo; Elman, Lauren
2014-01-01
To assess the relationship between clinically assessed Upper Motor Neuron (UMN) disease in Amyotrophic Lateral Sclerosis (ALS) and local diffusion alterations measured in the brain corticospinal tract (CST) by a tractography-driven template-space region-of-interest (ROI) analysis of Diffusion Tensor Imaging (DTI). This cross-sectional study included 34 patients with ALS, on whom DTI was performed. Clinical measures were separately obtained including the Penn UMN Score, a summary metric based upon standard clinical methods. After normalizing all DTI data to a population-specific template, tractography was performed to determine a region-of-interest (ROI) outlining the CST, in which average Mean Diffusivity (MD) and Fractional Anisotropy (FA) were estimated. Linear regression analyses were used to investigate associations of DTI metrics (MD, FA) with clinical measures (Penn UMN Score, ALSFRS-R, duration-of-disease), along with age, sex, handedness, and El Escorial category as covariates. For MD, the regression model was significant (p = 0.02), and the only significant predictors were the Penn UMN Score (p = 0.005) and age (p = 0.03). The FA regression model was also significant (p = 0.02); the only significant predictor was the Penn UMN Score (p = 0.003). Measured by the template-space ROI method, both MD and FA were linearly associated with the Penn UMN Score, supporting the hypothesis that DTI alterations reflect UMN pathology as assessed by the clinical examination.
Comparison between Linear and Nonlinear Regression in a Laboratory Heat Transfer Experiment
ERIC Educational Resources Information Center
Gonçalves, Carine Messias; Schwaab, Marcio; Pinto, José Carlos
2013-01-01
In order to interpret laboratory experimental data, undergraduate students are used to perform linear regression through linearized versions of nonlinear models. However, the use of linearized models can lead to statistically biased parameter estimates. Even so, it is not an easy task to introduce nonlinear regression and show for the students…
Wright, Melecia; Sotres-Alvarez, Daniela; Mendez, Michelle A; Adair, Linda
2017-03-01
No study has analysed how protein intake from early childhood to young adulthood relate to adult BMI in a single cohort. To estimate the association of protein intake at 2, 11, 15, 19 and 22 years with age- and sex-standardised BMI at 22 years (early adulthood), we used linear regression models with dietary and anthropometric data from a Filipino birth cohort (1985-2005, n 2586). We used latent growth curve analysis to identify trajectories of protein intake relative to age-specific recommended daily allowance (intake in g/kg body weight) from 2 to 22 years, then related trajectory membership to early adulthood BMI using linear regression models. Lean mass and fat mass were secondary outcomes. Regression models included socioeconomic, dietary and anthropometric confounders from early life and adulthood. Protein intake relative to needs at age 2 years was positively associated with BMI and lean mass at age 22 years, but intakes at ages 11, 15 and 22 years were inversely associated with early adulthood BMI. Individuals were classified into four mutually exclusive trajectories: (i) normal consumers (referent trajectory, 58 % of cohort), (ii) high protein consumers in infancy (20 %), (iii) usually high consumers (18 %) and (iv) always high consumers (5 %). Compared with the normal consumers, 'usually high' consumption was inversely associated with BMI, lean mass and fat mass at age 22 years whereas 'always high' consumption was inversely associated with male lean mass in males. Proximal protein intakes were more important contributors to early adult BMI relative to early-childhood protein intake; protein intake history was differentially associated with adulthood body size.
Box-Cox transformation of firm size data in statistical analysis
NASA Astrophysics Data System (ADS)
Chen, Ting Ting; Takaishi, Tetsuya
2014-03-01
Firm size data usually do not show the normality that is often assumed in statistical analysis such as regression analysis. In this study we focus on two firm size data: the number of employees and sale. Those data deviate considerably from a normal distribution. To improve the normality of those data we transform them by the Box-Cox transformation with appropriate parameters. The Box-Cox transformation parameters are determined so that the transformed data best show the kurtosis of a normal distribution. It is found that the two firm size data transformed by the Box-Cox transformation show strong linearity. This indicates that the number of employees and sale have the similar property as a firm size indicator. The Box-Cox parameters obtained for the firm size data are found to be very close to zero. In this case the Box-Cox transformations are approximately a log-transformation. This suggests that the firm size data we used are approximately log-normal distributions.
A quantitative description of normal AV nodal conduction curve in man.
Teague, S; Collins, S; Wu, D; Denes, P; Rosen, K; Arzbaecher, R
1976-01-01
The AV nodal conduction curve generated by the atrial extrastimulus technique has been described only qualitatively in man, making clinical comparison of known normal curves with those of suspected AV nodal dysfunction difficult. Also, the effects of physiological and pharmacological interventions have not been quantifiable. In 50 patients with normal AV conduction as defined by normal AH (less than 130 ms), normal AV nodal effective and functional refractory periods (less than 380 and less than 500 ms), and absence of demonstrable dual AV nodal pathways, we found that conduction curves (at sinus rhythm or longest paced cycle length) can be described by an exponential equation of the form delta = Ae-Bx. In this equation, delta is the increase in AV nodal conduction time of an extrastimulus compared to that of a regular beat and x is extrastimulus interval. The natural logarithm of this equation is linear in the semilogarithmic plane, thus permitting the constants A and B to be easily determined by a least-squares regression analysis with a hand calculator.
Flow-covariate prediction of stream pesticide concentrations.
Mosquin, Paul L; Aldworth, Jeremy; Chen, Wenlin
2018-01-01
Potential peak functions (e.g., maximum rolling averages over a given duration) of annual pesticide concentrations in the aquatic environment are important exposure parameters (or target quantities) for ecological risk assessments. These target quantities require accurate concentration estimates on nonsampled days in a monitoring program. We examined stream flow as a covariate via universal kriging to improve predictions of maximum m-day (m = 1, 7, 14, 30, 60) rolling averages and the 95th percentiles of atrazine concentration in streams where data were collected every 7 or 14 d. The universal kriging predictions were evaluated against the target quantities calculated directly from the daily (or near daily) measured atrazine concentration at 32 sites (89 site-yr) as part of the Atrazine Ecological Monitoring Program in the US corn belt region (2008-2013) and 4 sites (62 site-yr) in Ohio by the National Center for Water Quality Research (1993-2008). Because stream flow data are strongly skewed to the right, 3 transformations of the flow covariate were considered: log transformation, short-term flow anomaly, and normalized Box-Cox transformation. The normalized Box-Cox transformation resulted in predictions of the target quantities that were comparable to those obtained from log-linear interpolation (i.e., linear interpolation on the log scale) for 7-d sampling. However, the predictions appeared to be negatively affected by variability in regression coefficient estimates across different sample realizations of the concentration time series. Therefore, revised models incorporating seasonal covariates and partially or fully constrained regression parameters were investigated, and they were found to provide much improved predictions in comparison with those from log-linear interpolation for all rolling average measures. Environ Toxicol Chem 2018;37:260-273. © 2017 SETAC. © 2017 SETAC.
Liu, Shu; Yu, Marco; Weinreb, Robert N; Lai, Gilda; Lam, Dennis Shun-Chiu; Leung, Christopher Kai-Shun
2014-05-02
We compared the detection of visual field progression and its rate of change between standard automated perimetry (SAP) and Matrix frequency doubling technology perimetry (FDTP) in glaucoma. We followed prospectively 217 eyes (179 glaucoma and 38 normal eyes) for SAP and FDTP testing at 4-month intervals for ≥36 months. Pointwise linear regression analysis was performed. A test location was considered progressing when the rate of change of visual sensitivity was ≤-1 dB/y for nonedge and ≤-2 dB/y for edge locations. Three criteria were used to define progression in an eye: ≥3 adjacent nonedge test locations (conservative), any three locations (moderate), and any two locations (liberal) progressed. The rate of change of visual sensitivity was calculated with linear mixed models. Of the 217 eyes, 6.1% and 3.9% progressed with the conservative criteria, 14.5% and 5.6% of eyes progressed with the moderate criteria, and 20.1% and 11.7% of eyes progressed with the liberal criteria by FDTP and SAP, respectively. Taking all test locations into consideration (total, 54 × 179 locations), FDTP detected more progressing locations (176) than SAP (103, P < 0.001). The rate of change of visual field mean deviation (MD) was significantly faster for FDTP (all with P < 0.001). No eyes showed progression in the normal group using the conservative and the moderate criteria. With a faster rate of change of visual sensitivity, FDTP detected more progressing eyes than SAP at a comparable level of specificity. Frequency doubling technology perimetry can provide a useful alternative to monitor glaucoma progression.
NASA Astrophysics Data System (ADS)
Rock, N. M. S.; Duffy, T. R.
REGRES allows a range of regression equations to be calculated for paired sets of data values in which both variables are subject to error (i.e. neither is the "independent" variable). Nonparametric regressions, based on medians of all possible pairwise slopes and intercepts, are treated in detail. Estimated slopes and intercepts are output, along with confidence limits, Spearman and Kendall rank correlation coefficients. Outliers can be rejected with user-determined stringency. Parametric regressions can be calculated for any value of λ (the ratio of the variances of the random errors for y and x)—including: (1) major axis ( λ = 1); (2) reduced major axis ( λ = variance of y/variance of x); (3) Y on Xλ = infinity; or (4) X on Y ( λ = 0) solutions. Pearson linear correlation coefficients also are output. REGRES provides an alternative to conventional isochron assessment techniques where bivariate normal errors cannot be assumed, or weighting methods are inappropriate.
The Application of the Cumulative Logistic Regression Model to Automated Essay Scoring
ERIC Educational Resources Information Center
Haberman, Shelby J.; Sinharay, Sandip
2010-01-01
Most automated essay scoring programs use a linear regression model to predict an essay score from several essay features. This article applied a cumulative logit model instead of the linear regression model to automated essay scoring. Comparison of the performances of the linear regression model and the cumulative logit model was performed on a…
NASA Astrophysics Data System (ADS)
Xie, Yanan; Zhou, Mingliang; Pan, Dengke
2017-10-01
The forward-scattering model is introduced to describe the response of normalized radar cross section (NRCS) of precipitation with synthetic aperture radar (SAR). Since the distribution of near-surface rainfall is related to the rate of near-surface rainfall and horizontal distribution factor, a retrieval algorithm called modified regression empirical and model-oriented statistical (M-M) based on the volterra integration theory is proposed. Compared with the model-oriented statistical and volterra integration (MOSVI) algorithm, the biggest difference is that the M-M algorithm is based on the modified regression empirical algorithm rather than the linear regression formula to retrieve the value of near-surface rainfall rate. Half of the empirical parameters are reduced in the weighted integral work and a smaller average relative error is received while the rainfall rate is less than 100 mm/h. Therefore, the algorithm proposed in this paper can obtain high-precision rainfall information.
Galluzzi, Paolo; de Jong, Marcus C; Sirin, Selma; Maeder, Philippe; Piu, Pietro; Cerase, Alfonso; Monti, Lucia; Brisse, Hervé J; Castelijns, Jonas A; de Graaf, Pim; Goericke, Sophia L
2016-07-01
Differentiation between normal solid (non-cystic) pineal glands and pineal pathologies on brain MRI is difficult. The aim of this study was to assess the size of the solid pineal gland in children (0-5 years) and compare the findings with published pineoblastoma cases. We retrospectively analyzed the size (width, height, planimetric area) of solid pineal glands in 184 non-retinoblastoma patients (73 female, 111 male) aged 0-5 years on MRI. The effect of age and gender on gland size was evaluated. Linear regression analysis was performed to analyze the relation between size and age. Ninety-nine percent prediction intervals around the mean were added to construct a normal size range per age, with the upper bound of the predictive interval as the parameter of interest as a cutoff for normalcy. There was no significant interaction of gender and age for all the three pineal gland parameters (width, height, and area). Linear regression analysis gave 99 % upper prediction bounds of 7.9, 4.8, and 25.4 mm(2), respectively, for width, height, and area. The slopes (size increase per month) of each parameter were 0.046, 0.023, and 0.202, respectively. Ninety-three percent (95 % CI 66-100 %) of asymptomatic solid pineoblastomas were larger in size than the 99 % upper bound. This study establishes norms for solid pineal gland size in non-retinoblastoma children aged 0-5 years. Knowledge of the size of the normal pineal gland is helpful for detection of pineal gland abnormalities, particularly pineoblastoma.
Sirin, Selma; de Jong, Marcus C; Galluzzi, Paolo; Maeder, Philippe; Brisse, Hervé J; Castelijns, Jonas A; de Graaf, Pim; Goericke, Sophia L
2016-07-01
Pineal cysts are a common incidental finding on brain MRI with resulting difficulties in differentiation between normal glands and pineal pathologies. The aim of this study was to assess the size and morphology of the cystic pineal gland in children (0-5 years) and compare the findings with published pineoblastoma cases. In this retrospective multicenter study, 257 MR examinations (232 children, 0-5 years) were evaluated regarding pineal gland size (width, height, planimetric area, maximal cyst(s) size) and morphology. We performed linear regression analysis with 99 % prediction intervals of gland size versus age for the size parameters. Results were compared with a recent meta-analysis of pineoblastoma by de Jong et al. Follow-up was available in 25 children showing stable cystic findings in 48 %, cyst size increase in 36 %, and decrease in 16 %. Linear regression analysis gave 99 % upper prediction bounds of 10.8 mm, 10.9 mm, 7.7 mm and 66.9 mm(2), respectively, for cyst size, width, height, and area. The slopes (size increase per month) of each parameter were 0.030, 0.046, 0.021, and 0.25, respectively. Most of the pineoblastomas showed a size larger than the 99 % upper prediction margin, but with considerable overlap between the groups. We presented age-adapted normal values for size and morphology of the cystic pineal gland in children aged 0 to 5 years. Analysis of size is helpful in discriminating normal glands from cystic pineal pathologies such as pineoblastoma. We also presented guidelines for the approach of a solid or cystic pineal gland in hereditary retinoblastoma patients.
Excess adiposity, inflammation, and iron-deficiency in female adolescents.
Tussing-Humphreys, Lisa M; Liang, Huifang; Nemeth, Elizabeta; Freels, Sally; Braunschweig, Carol A
2009-02-01
Iron deficiency is more prevalent in overweight children and adolescents but the mechanisms that underlie this condition remain unclear. The purpose of this cross-sectional study was to assess the relationship between iron status and excess adiposity, inflammation, menarche, diet, physical activity, and poverty status in female adolescents included in the National Health and Nutrition Examination Survey 2003-2004 dataset. Descriptive and simple comparative statistics (t test, chi(2)) were used to assess differences between normal-weight (5th < or = body mass index [BMI] percentile <85th) and heavier-weight girls (< or = 85th percentile for BMI) for demographic, biochemical, dietary, and physical activity variables. In addition, logistic regression analyses predicting iron deficiency and linear regression predicting serum iron levels were performed. Heavier-weight girls had an increased prevalence of iron deficiency compared to those with normal weight. Dietary iron, age of and time since first menarche, poverty status, and physical activity were similar between the two groups and were not independent predictors of iron deficiency or log serum iron levels. Logistic modeling predicting iron deficiency revealed having a BMI > or = 85th percentile and for each 1 mg/dL increase in C-reactive protein the odds ratio for iron deficiency more than doubled. The best-fit linear model to predict serum iron levels included both serum transferrin receptor and C-reactive protein following log-transformation for normalization of these variables. Findings indicate that heavier-weight female adolescents are at greater risk for iron deficiency and that inflammation stemming from excess adipose tissue contributes to this phenomenon. Food and nutrition professionals should consider elevated BMI as an additional risk factor for iron deficiency in female adolescents.
Yan, Qun; Sun, Dongmei; Li, Xu; Chen, Guoliang; Zheng, Qinghu; Li, Lun; Gu, Chenhong; Feng, Bo
2016-07-13
There is a scarcity of epidemiological researches examining the relationship between blood pressure (BP) and glucose level among older adults. The objective of the current study was to investigate the association of high BP and glucose level in elderly Chinese. A cross-sectional study of a population of 2092 Chinese individuals aged over 65 years was conducted. Multiple logistic analysis was used to explore the association between hypertension and hyperglycemia. Independent risk factors for systolic and diastolic BP were analyzed using stepwise linear regression. Subjects in impaired fasting glucose group (IFG) (n = 144) and diabetes (n = 346), as compared with normal fasting glucose (NFG) (n = 1277), had a significant higher risk for hypertension, with odds ratios (ORs) of 1.81 (95 % CI, 1.39-2.35) (P = 0.000) and 1.40 (95 % CI, 1.09-1.80) (P = 0.009), respectively. Higher fasting plasma glucose (FPG) levels in the normal range were still significantly associated with a higher prevalence of hypertension in both genders, with ORs of 1.24 (95 % CI, 0.85-1.80), R (2) = 0.114, P = 0.023 in men and 1.61 (95 % CI, 1.12-2.30), R (2) = 0.082, P = 0.010 in women, respectively, when compared with lower FPG. Linear regression analysis revealed FPG was an independent factor of systolic and diastolic BP. Our findings suggest that hyperglycemia as well as higher FPG within the normal range is associated with a higher prevalence of hypertension independent of other cardiovascular risk factors in elderly Chinese. Further studies are needed to explore the relationship between hyperglycemia and hypertension in a longitudinal setting.
NASA Astrophysics Data System (ADS)
Gao, Xiangyun; An, Haizhong; Fang, Wei; Huang, Xuan; Li, Huajiao; Zhong, Weiqiong; Ding, Yinghui
2014-07-01
The linear regression parameters between two time series can be different under different lengths of observation period. If we study the whole period by the sliding window of a short period, the change of the linear regression parameters is a process of dynamic transmission over time. We tackle fundamental research that presents a simple and efficient computational scheme: a linear regression patterns transmission algorithm, which transforms linear regression patterns into directed and weighted networks. The linear regression patterns (nodes) are defined by the combination of intervals of the linear regression parameters and the results of the significance testing under different sizes of the sliding window. The transmissions between adjacent patterns are defined as edges, and the weights of the edges are the frequency of the transmissions. The major patterns, the distance, and the medium in the process of the transmission can be captured. The statistical results of weighted out-degree and betweenness centrality are mapped on timelines, which shows the features of the distribution of the results. Many measurements in different areas that involve two related time series variables could take advantage of this algorithm to characterize the dynamic relationships between the time series from a new perspective.
Gao, Xiangyun; An, Haizhong; Fang, Wei; Huang, Xuan; Li, Huajiao; Zhong, Weiqiong; Ding, Yinghui
2014-07-01
The linear regression parameters between two time series can be different under different lengths of observation period. If we study the whole period by the sliding window of a short period, the change of the linear regression parameters is a process of dynamic transmission over time. We tackle fundamental research that presents a simple and efficient computational scheme: a linear regression patterns transmission algorithm, which transforms linear regression patterns into directed and weighted networks. The linear regression patterns (nodes) are defined by the combination of intervals of the linear regression parameters and the results of the significance testing under different sizes of the sliding window. The transmissions between adjacent patterns are defined as edges, and the weights of the edges are the frequency of the transmissions. The major patterns, the distance, and the medium in the process of the transmission can be captured. The statistical results of weighted out-degree and betweenness centrality are mapped on timelines, which shows the features of the distribution of the results. Many measurements in different areas that involve two related time series variables could take advantage of this algorithm to characterize the dynamic relationships between the time series from a new perspective.
Gierlinger, Notburga; Luss, Saskia; König, Christian; Konnerth, Johannes; Eder, Michaela; Fratzl, Peter
2010-01-01
The functional characteristics of plant cell walls depend on the composition of the cell wall polymers, as well as on their highly ordered architecture at scales from a few nanometres to several microns. Raman spectra of wood acquired with linear polarized laser light include information about polymer composition as well as the alignment of cellulose microfibrils with respect to the fibre axis (microfibril angle). By changing the laser polarization direction in 3 degrees steps, the dependency between cellulose and laser orientation direction was investigated. Orientation-dependent changes of band height ratios and spectra were described by quadratic linear regression and partial least square regressions, respectively. Using the models and regressions with high coefficients of determination (R(2) > 0.99) microfibril orientation was predicted in the S1 and S2 layers distinguished by the Raman imaging approach in cross-sections of spruce normal, opposite, and compression wood. The determined microfibril angle (MFA) in the different S2 layers ranged from 0 degrees to 49.9 degrees and was in coincidence with X-ray diffraction determination. With the prerequisite of geometric sample and laser alignment, exact MFA prediction can complete the picture of the chemical cell wall design gained by the Raman imaging approach at the micron level in all plant tissues.
Chen, Xuexia; Vogelmann, James E.; Chander, Gyanesh; Ji, Lei; Tolk, Brian; Huang, Chengquan; Rollins, Matthew
2013-01-01
Routine acquisition of Landsat 5 Thematic Mapper (TM) data was discontinued recently and Landsat 7 Enhanced Thematic Mapper Plus (ETM+) has an ongoing problem with the scan line corrector (SLC), thereby creating spatial gaps when covering images obtained during the process. Since temporal and spatial discontinuities of Landsat data are now imminent, it is therefore important to investigate other potential satellite data that can be used to replace Landsat data. We thus cross-compared two near-simultaneous images obtained from Landsat 5 TM and the Indian Remote Sensing (IRS)-P6 Advanced Wide Field Sensor (AWiFS), both captured on 29 May 2007 over Los Angeles, CA. TM and AWiFS reflectances were compared for the green, red, near-infrared (NIR), and shortwave infrared (SWIR) bands, as well as the normalized difference vegetation index (NDVI) based on manually selected polygons in homogeneous areas. All R2 values of linear regressions were found to be higher than 0.99. The temporally invariant cluster (TIC) method was used to calculate the NDVI correlation between the TM and AWiFS images. The NDVI regression line derived from selected polygons passed through several invariant cluster centres of the TIC density maps and demonstrated that both the scene-dependent polygon regression method and TIC method can generate accurate radiometric normalization. A scene-independent normalization method was also used to normalize the AWiFS data. Image agreement assessment demonstrated that the scene-dependent normalization using homogeneous polygons provided slightly higher accuracy values than those obtained by the scene-independent method. Finally, the non-normalized and relatively normalized ‘Landsat-like’ AWiFS 2007 images were integrated into 1984 to 2010 Landsat time-series stacks (LTSS) for disturbance detection using the Vegetation Change Tracker (VCT) model. Both scene-dependent and scene-independent normalized AWiFS data sets could generate disturbance maps similar to what were generated using the LTSS data set, and their kappa coefficients were higher than 0.97. These results indicate that AWiFS can be used instead of Landsat data to detect multitemporal disturbance in the event of Landsat data discontinuity.
Daily Magnesium Intake and Serum Magnesium Concentration among Japanese People
Akizawa, Yoriko; Koizumi, Sadayuki; Itokawa, Yoshinori; Ojima, Toshiyuki; Nakamura, Yosikazu; Tamura, Tarou; Kusaka, Yukinori
2008-01-01
Background The vitamins and minerals that are deficient in the daily diet of a normal adult remain unknown. To answer this question, we conducted a population survey focusing on the relationship between dietary magnesium intake and serum magnesium level. Methods The subjects were 62 individuals from Fukui Prefecture who participated in the 1998 National Nutrition Survey. The survey investigated the physical status, nutritional status, and dietary data of the subjects. Holidays and special occasions were avoided, and a day when people are most likely to be on an ordinary diet was selected as the survey date. Results The mean (±standard deviation) daily magnesium intake was 322 (±132), 323 (±163), and 322 (±147) mg/day for men, women, and the entire group, respectively. The mean (±standard deviation) serum magnesium concentration was 20.69 (±2.83), 20.69 (±2.88), and 20.69 (±2.83) ppm for men, women, and the entire group, respectively. The distribution of serum magnesium concentration was normal. Dietary magnesium intake showed a log-normal distribution, which was then transformed by logarithmic conversion for examining the regression coefficients. The slope of the regression line between the serum magnesium concentration (Y ppm) and daily magnesium intake (X mg) was determined using the formula Y = 4.93 (log10X) + 8.49. The coefficient of correlation (r) was 0.29. A regression line (Y = 14.65X + 19.31) was observed between the daily intake of magnesium (Y mg) and serum magnesium concentration (X ppm). The coefficient of correlation was 0.28. Conclusion The daily magnesium intake correlated with serum magnesium concentration, and a linear regression model between them was proposed. PMID:18635902
Korany, Mohamed A; Gazy, Azza A; Khamis, Essam F; Ragab, Marwa A A; Kamal, Miranda F
2018-06-01
This study outlines two robust regression approaches, namely least median of squares (LMS) and iteratively re-weighted least squares (IRLS) to investigate their application in instrument analysis of nutraceuticals (that is, fluorescence quenching of merbromin reagent upon lipoic acid addition). These robust regression methods were used to calculate calibration data from the fluorescence quenching reaction (∆F and F-ratio) under ideal or non-ideal linearity conditions. For each condition, data were treated using three regression fittings: Ordinary Least Squares (OLS), LMS and IRLS. Assessment of linearity, limits of detection (LOD) and quantitation (LOQ), accuracy and precision were carefully studied for each condition. LMS and IRLS regression line fittings showed significant improvement in correlation coefficients and all regression parameters for both methods and both conditions. In the ideal linearity condition, the intercept and slope changed insignificantly, but a dramatic change was observed for the non-ideal condition and linearity intercept. Under both linearity conditions, LOD and LOQ values after the robust regression line fitting of data were lower than those obtained before data treatment. The results obtained after statistical treatment indicated that the linearity ranges for drug determination could be expanded to lower limits of quantitation by enhancing the regression equation parameters after data treatment. Analysis results for lipoic acid in capsules, using both fluorimetric methods, treated by parametric OLS and after treatment by robust LMS and IRLS were compared for both linearity conditions. Copyright © 2018 John Wiley & Sons, Ltd.
Yiming, Gulinuer; Zhou, Xianhui; Lv, Wenkui; Peng, Yi; Zhang, Wenhui; Cheng, Xinchun; Li, Yaodong; Xing, Qiang; Zhang, Jianghua; Zhou, Qina; Zhang, Ling; Lu, Yanmei; Wang, Hongli; Tang, Baopeng
2017-01-01
Brachial-ankle pulse wave velocity (baPWV), a direct measure of aortic stiffness, has increasingly become an important assessment for cardiovascular risk. The present study established the reference and normal values of baPWV in a Central Asia population in Xinjiang, China. We recruited participants from a central Asia population in Xinjiang, China. We performed multiple regression analysis to investigate the determinants of baPWV. The median and 10th-90th percentiles were calculated to establish the reference and normal values based on these categories. In total, 5,757 Han participants aged 15-88 years were included in the present study. Spearman correlation analysis showed that age (r = 0.587, p < 0.001) and mean blood pressure (MBP, r = 0.599, p <0.001) were the major factors influencing the values of baPWV in the reference population. Furthermore, in the multiple linear regression analysis, the standardized regression coefficients of age (0.445) and MBP (0.460) were much higher than those of body mass index, triglyceride, and glycemia (-0.054, 0.035, and 0.033, respectively). In the covariance analysis, after adjustment for age and MBP, only diabetes was the significant independent determinant of baPWV (p = 0.009). Thus, participants with diabetes were excluded from the reference value population. The reference values ranged from 14.3 to 25.2 m/s, and the normal values ranged from 13.9 to 21.2 m/s. This is the first study that has established the reference and normal values for baPWV according to age and blood pressure in a Central Asia population.
1974-01-01
REGRESSION MODEL - THE UNCONSTRAINED, LINEAR EQUALITY AND INEQUALITY CONSTRAINED APPROACHES January 1974 Nelson Delfino d’Avila Mascarenha;? Image...Report 520 DIGITAL IMAGE RESTORATION UNDER A REGRESSION MODEL THE UNCONSTRAINED, LINEAR EQUALITY AND INEQUALITY CONSTRAINED APPROACHES January...a two- dimensional form adequately describes the linear model . A dis- cretization is performed by using quadrature methods. By trans
NASA Astrophysics Data System (ADS)
Donroman, T.; Chesoh, S.; Lim, A.
2018-04-01
This study aimed to investigate the variation patterns of fish fingerling abundance based on month, year and sampling site. Monthly collecting data set of the Na Thap tidal river of southern Thailand, were obtained from June 2005 to October 2015. The square root transformation was employed for maintaining the fingerling data normality. Factor analysis was applied for clustering number of fingerling species and multiple linear regression was used to examine the association between fingerling density and year, month and site. Results from factor analysis classified fingerling into 3 factors based on saline preference; saline water, freshwater and ubiquitous species. The results showed a statistically high significant relation between fingerling density, month, year and site. Abundance of saline water and ubiquitous fingerling density showed similar pattern. Downstream site presented highest fingerling density whereas almost of freshwater fingerling occurred in upstream. This finding confirmed that factor analysis and the general linear regression method can be used as an effective tool for predicting and monitoring wild fingerling density in order to sustain fish stock management.
Repeated Kicking Actions in Karate: Effect on Technical Execution in Elite Practitioners.
Quinzi, Federico; Camomilla, Valentina; Di Mario, Alberto; Felici, Francesco; Sbriccoli, Paola
2016-04-01
Training in martial arts is commonly performed by repeating a technical action continuously for a given number of times. This study aimed to investigate if the repetition of the task alters the proper technical execution, limiting the training efficacy for the technical evaluation during competition. This aim was pursued analyzing lower-limb kinematics and muscle activation during repeated roundhouse kicks. Six junior karate practitioners performed continuously 20 repetitions of the kick. Hip and knee kinematics and sEMG of vastus lateralis, biceps (BF), and rectus femoris were recorded. For each repetition, hip abduction-adduction and flexion-extension and knee flexion-extension peak angular displacements and velocities, agonist and antagonist muscle activation were computed. Moreover, to monitor for the presence of myoelectric fatigue, if any, the median frequency of the sEMG was computed. All variables were normalized with respect to their individual maximum observed during the sequence of kicks. Linear regressions were fitted to each normalized parameter to test its relationship with the repetition number. Linear-regression analysis showed that, during the sequence, the athletes modified their technique: Knee flexion, BF median frequency, hip abduction, knee-extension angular velocity, and BF antagonist activation significantly decreased. Conversely, hip flexion increased significantly. Since karate combat competitions require proper technical execution, training protocols combining severe fatigue and technical actions should be carefully proposed because of technique adaptations. Moreover, trainers and karate masters should consider including specific strength exercises for the BF and more generally for knee flexors.
Statistics for nuclear engineers and scientists. Part 1. Basic statistical inference
DOE Office of Scientific and Technical Information (OSTI.GOV)
Beggs, W.J.
1981-02-01
This report is intended for the use of engineers and scientists working in the nuclear industry, especially at the Bettis Atomic Power Laboratory. It serves as the basis for several Bettis in-house statistics courses. The objectives of the report are to introduce the reader to the language and concepts of statistics and to provide a basic set of techniques to apply to problems of the collection and analysis of data. Part 1 covers subjects of basic inference. The subjects include: descriptive statistics; probability; simple inference for normally distributed populations, and for non-normal populations as well; comparison of two populations; themore » analysis of variance; quality control procedures; and linear regression analysis.« less
Barlough, J E; Jacobson, R H; Downing, D R; Lynch, T J; Scott, F W
1987-01-01
The computer-assisted, kinetics-based enzyme-linked immunosorbent assay for coronavirus antibodies in cats was calibrated to the conventional indirect immunofluorescence assay by linear regression analysis and computerized interpolation (generation of "immunofluorescence assay-equivalent" titers). Procedures were developed for normalization and standardization of kinetics-based enzyme-linked immunosorbent assay results through incorporation of five different control sera of predetermined ("expected") titer in daily runs. When used with such sera and with computer assistance, the kinetics-based enzyme-linked immunosorbent assay minimized both within-run and between-run variability while allowing also for efficient data reduction and statistical analysis and reporting of results. PMID:3032390
Barlough, J E; Jacobson, R H; Downing, D R; Lynch, T J; Scott, F W
1987-01-01
The computer-assisted, kinetics-based enzyme-linked immunosorbent assay for coronavirus antibodies in cats was calibrated to the conventional indirect immunofluorescence assay by linear regression analysis and computerized interpolation (generation of "immunofluorescence assay-equivalent" titers). Procedures were developed for normalization and standardization of kinetics-based enzyme-linked immunosorbent assay results through incorporation of five different control sera of predetermined ("expected") titer in daily runs. When used with such sera and with computer assistance, the kinetics-based enzyme-linked immunosorbent assay minimized both within-run and between-run variability while allowing also for efficient data reduction and statistical analysis and reporting of results.
The microcomputer scientific software series 2: general linear model--regression.
Harold M. Rauscher
1983-01-01
The general linear model regression (GLMR) program provides the microcomputer user with a sophisticated regression analysis capability. The output provides a regression ANOVA table, estimators of the regression model coefficients, their confidence intervals, confidence intervals around the predicted Y-values, residuals for plotting, a check for multicollinearity, a...
Emmert, J L; Baker, D H
1997-05-01
Our objectives were to use a soy protein isolate (SPI) diet containing 2-amino-2-methyl-1-propanol, an inhibitor of choline biosynthesis, to determine the bioavailable choline content of normal and overheated soybean meal (SBM), canola meal (CM) and peanut meal (PM). In the first four experiments, it was determined that weight gain of chicks fed the basal diet would respond linearly (P < 0.05) to graded levels of crystalline choline and would not respond to betaine, and that when fortified with adequate choline, no weight gain or feed intake response would occur upon addition of 100 g/kg SBM, CM or PM to the basal diet. Furthermore, addition of crystalline amino acids simulating the amino acid composition of 100 g/kg SBM did not alter the utilization of crystalline choline. In Experiment 5, feeding graded doses of choline, SBM, CM or PM resulted in linear (P < 0.05) increases in weight gain. Multiple linear regression analysis indicated bioavailable choline concentrations of 1708, 1545 and 1203 mg/kg for SBM, CM and PM, respectively. In Experiment 6, no differences (P > 0.05) in bioavailable choline concentrations occurred between normal and overheated SBM, CM or PM, and the bioavailable choline concentration of normal SBM, CM and PM was 2002, 1464 and 1320 mg/kg, respectively. Average bioavailable choline levels were 83, 24 and 76% of analytically determined choline levels in SBM, CM and PM, respectively. Canola meal, although three times as rich in total choline as SBM, has less bioavailable choline than SBM. A substantial portion of choline in SBM, CM and PM is unavailable, and overheating does not appear to decrease the bioavailability of choline in these products.
NASA Astrophysics Data System (ADS)
Szilagyi, Jozsef
2015-11-01
Thirty year normal (1981-2010) monthly latent heat fluxes (ET) over the conterminous United States were estimated by a modified Advection-Aridity model from North American Regional Reanalysis (NARR) radiation and wind as well as Parameter-Elevation Regressions on Independent Slopes Model (PRISM) air and dew-point temperature data. Mean annual ET values were calibrated with PRISM precipitation (P) and validated against United States Geological Survey runoff (Q) data. At the six-digit Hydrologic Unit Code level (sample size of 334) the estimated 30 year normal runoff (P - ET) had a bias of 18 mm yr-1, a root-mean-square error of 96 mm yr-1, and a linear correlation coefficient value of 0.95, making the estimates on par with the latest Land Surface Model results but without the need for soil and vegetation information or any soil moisture budgeting.
Wang, D Z; Wang, C; Shen, C F; Zhang, Y; Zhang, H; Song, G D; Xue, X D; Xu, Z L; Zhang, S; Jiang, G H
2017-05-10
We described the time trend of acute myocardial infarction (AMI) from 1999 to 2013 in Tianjin incidence rate with Cochran-Armitage trend (CAT) test and linear regression analysis, and the results were compared. Based on actual population, CAT test had much stronger statistical power than linear regression analysis for both overall incidence trend and age specific incidence trend (Cochran-Armitage trend P value
Yang, Ruiqi; Wang, Fei; Zhang, Jialing; Zhu, Chonglei; Fan, Limei
2015-05-19
To establish the reference values of thalamus, caudate nucleus and lenticular nucleus diameters through fetal thalamic transverse section. A total of 265 fetuses at our hospital were randomly selected from November 2012 to August 2014. And the transverse and length diameters of thalamus, caudate nucleus and lenticular nucleus were measured. SPSS 19.0 statistical software was used to calculate the regression curve of fetal diameter changes and gestational weeks of pregnancy. P < 0.05 was considered as having statistical significance. The linear regression equation of fetal thalamic length diameter and gestational week was: Y = 0.051X+0.201, R = 0.876, linear regression equation of thalamic transverse diameter and fetal gestational week was: Y = 0.031X+0.229, R = 0.817, linear regression equation of fetal head of caudate nucleus length diameter and gestational age was: Y = 0.033X+0.101, R = 0.722, linear regression equation of fetal head of caudate nucleus transverse diameter and gestational week was: R = 0.025 - 0.046, R = 0.711, linear regression equation of fetal lentiform nucleus length diameter and gestational week was: Y = 0.046+0.229, R = 0.765, linear regression equation of fetal lentiform nucleus diameter and gestational week was: Y = 0.025 - 0.05, R = 0.772. Ultrasonic measurement of diameter of fetal thalamus caudate nucleus, and lenticular nucleus through thalamic transverse section is simple and convenient. And measurements increase with fetal gestational weeks and there is linear regression relationship between them.
Local Linear Regression for Data with AR Errors.
Li, Runze; Li, Yan
2009-07-01
In many statistical applications, data are collected over time, and they are likely correlated. In this paper, we investigate how to incorporate the correlation information into the local linear regression. Under the assumption that the error process is an auto-regressive process, a new estimation procedure is proposed for the nonparametric regression by using local linear regression method and the profile least squares techniques. We further propose the SCAD penalized profile least squares method to determine the order of auto-regressive process. Extensive Monte Carlo simulation studies are conducted to examine the finite sample performance of the proposed procedure, and to compare the performance of the proposed procedures with the existing one. From our empirical studies, the newly proposed procedures can dramatically improve the accuracy of naive local linear regression with working-independent error structure. We illustrate the proposed methodology by an analysis of real data set.
Orthogonal Regression: A Teaching Perspective
ERIC Educational Resources Information Center
Carr, James R.
2012-01-01
A well-known approach to linear least squares regression is that which involves minimizing the sum of squared orthogonal projections of data points onto the best fit line. This form of regression is known as orthogonal regression, and the linear model that it yields is known as the major axis. A similar method, reduced major axis regression, is…
Practical Session: Simple Linear Regression
NASA Astrophysics Data System (ADS)
Clausel, M.; Grégoire, G.
2014-12-01
Two exercises are proposed to illustrate the simple linear regression. The first one is based on the famous Galton's data set on heredity. We use the lm R command and get coefficients estimates, standard error of the error, R2, residuals …In the second example, devoted to data related to the vapor tension of mercury, we fit a simple linear regression, predict values, and anticipate on multiple linear regression. This pratical session is an excerpt from practical exercises proposed by A. Dalalyan at EPNC (see Exercises 1 and 2 of http://certis.enpc.fr/~dalalyan/Download/TP_ENPC_4.pdf).
Poulin, Patrick; Hop, Cornelis Eca; Salphati, Laurent; Liederer, Bianca M
2013-04-01
Understanding drug distribution and accumulation in tumors would be informative in the assessment of efficacy in targeted therapy; however, existing methods for predicting tissue drug distribution focus on normal tissues and do not incorporate tumors. The main objective of this study was to describe the relationships between tissue-plasma concentration ratios (Kp ) of normal tissues and those of subcutaneous xenograft tumors under nonsteady-state conditions, and establish regression equations that could potentially be used for the prediction of drug levels in several human tumor xenografts in mouse, based solely on a Kp value determined in a normal tissue (e.g., muscle). A dataset of 17 compounds was collected from the literature and from Genentech. Tissue and plasma concentration data in mouse were obtained following oral gavage or intraperitoneal administration. Linear regression analyses were performed between Kp values in several normal tissues (muscle, lung, liver, or brain) and those in human tumor xenografts (CL6, EBC-1, HT-29, PC3, U-87, MCF-7-neo-Her2, or BT474M1.1). The tissue-plasma ratios in normal tissues reasonably correlated with the tumor-plasma ratios in CL6, EBC-1, HT-29, U-87, BT474M1.1, and MCF-7-neo-Her2 xenografts (r(2) in the range 0.62-1) but not with the PC3 xenograft. In general, muscle and lung exhibited the strongest correlation with tumor xenografts, followed by liver. Regression coefficients from brain were low, except between brain and the glioblastoma U-87 xenograft (r(2) in the range 0.62-0.94). Furthermore, reasonably strong correlations were observed between muscle and lung and between muscle and liver (r(2) in the range 0.67-0.96). The slopes of the regressions differed depending on the class of drug (strong vs. weak base) and type of tissue (brain vs. other tissues and tumors). Overall, this study will contribute to our understanding of tissue-plasma partition coefficients for tumors and facilitate the use of physiologically based pharmacokinetics (PBPK) modeling for chemotherapy in oncology studies. © 2013 Wiley Periodicals, Inc. and the American Pharmacists Association J Pharm Sci 102:1355-1369, 2013. Copyright © 2013 Wiley Periodicals, Inc.
Evaluating acoustic speaker normalization algorithms: evidence from longitudinal child data.
Kohn, Mary Elizabeth; Farrington, Charlie
2012-03-01
Speaker vowel formant normalization, a technique that controls for variation introduced by physical differences between speakers, is necessary in variationist studies to compare speakers of different ages, genders, and physiological makeup in order to understand non-physiological variation patterns within populations. Many algorithms have been established to reduce variation introduced into vocalic data from physiological sources. The lack of real-time studies tracking the effectiveness of these normalization algorithms from childhood through adolescence inhibits exploration of child participation in vowel shifts. This analysis compares normalization techniques applied to data collected from ten African American children across five time points. Linear regressions compare the reduction in variation attributable to age and gender for each speaker for the vowels BEET, BAT, BOT, BUT, and BOAR. A normalization technique is successful if it maintains variation attributable to a reference sociolinguistic variable, while reducing variation attributable to age. Results indicate that normalization techniques which rely on both a measure of central tendency and range of the vowel space perform best at reducing variation attributable to age, although some variation attributable to age persists after normalization for some sections of the vowel space. © 2012 Acoustical Society of America
Bivariate categorical data analysis using normal linear conditional multinomial probability model.
Sun, Bingrui; Sutradhar, Brajendra
2015-02-10
Bivariate multinomial data such as the left and right eyes retinopathy status data are analyzed either by using a joint bivariate probability model or by exploiting certain odds ratio-based association models. However, the joint bivariate probability model yields marginal probabilities, which are complicated functions of marginal and association parameters for both variables, and the odds ratio-based association model treats the odds ratios involved in the joint probabilities as 'working' parameters, which are consequently estimated through certain arbitrary 'working' regression models. Also, this later odds ratio-based model does not provide any easy interpretations of the correlations between two categorical variables. On the basis of pre-specified marginal probabilities, in this paper, we develop a bivariate normal type linear conditional multinomial probability model to understand the correlations between two categorical variables. The parameters involved in the model are consistently estimated using the optimal likelihood and generalized quasi-likelihood approaches. The proposed model and the inferences are illustrated through an intensive simulation study as well as an analysis of the well-known Wisconsin Diabetic Retinopathy status data. Copyright © 2014 John Wiley & Sons, Ltd.
Morse Code, Scrabble, and the Alphabet
ERIC Educational Resources Information Center
Richardson, Mary; Gabrosek, John; Reischman, Diann; Curtiss, Phyliss
2004-01-01
In this paper we describe an interactive activity that illustrates simple linear regression. Students collect data and analyze it using simple linear regression techniques taught in an introductory applied statistics course. The activity is extended to illustrate checks for regression assumptions and regression diagnostics taught in an…
Cer, Regina Z; Herrera-Galeano, J Enrique; Anderson, Joseph J; Bishop-Lilly, Kimberly A; Mokashi, Vishwesh P
2014-01-01
Understanding the biological roles of microRNAs (miRNAs) is a an active area of research that has produced a surge of publications in PubMed, particularly in cancer research. Along with this increasing interest, many open-source bioinformatics tools to identify existing and/or discover novel miRNAs in next-generation sequencing (NGS) reads become available. While miRNA identification and discovery tools are significantly improved, the development of miRNA differential expression analysis tools, especially in temporal studies, remains substantially challenging. Further, the installation of currently available software is non-trivial and steps of testing with example datasets, trying with one's own dataset, and interpreting the results require notable expertise and time. Subsequently, there is a strong need for a tool that allows scientists to normalize raw data, perform statistical analyses, and provide intuitive results without having to invest significant efforts. We have developed miRNA Temporal Analyzer (mirnaTA), a bioinformatics package to identify differentially expressed miRNAs in temporal studies. mirnaTA is written in Perl and R (Version 2.13.0 or later) and can be run across multiple platforms, such as Linux, Mac and Windows. In the current version, mirnaTA requires users to provide a simple, tab-delimited, matrix file containing miRNA name and count data from a minimum of two to a maximum of 20 time points and three replicates. To recalibrate data and remove technical variability, raw data is normalized using Normal Quantile Transformation (NQT), and linear regression model is used to locate any miRNAs which are differentially expressed in a linear pattern. Subsequently, remaining miRNAs which do not fit a linear model are further analyzed in two different non-linear methods 1) cumulative distribution function (CDF) or 2) analysis of variances (ANOVA). After both linear and non-linear analyses are completed, statistically significant miRNAs (P < 0.05) are plotted as heat maps using hierarchical cluster analysis and Euclidean distance matrix computation methods. mirnaTA is an open-source, bioinformatics tool to aid scientists in identifying differentially expressed miRNAs which could be further mined for biological significance. It is expected to provide researchers with a means of interpreting raw data to statistical summaries in a fast and intuitive manner.
Advanced statistics: linear regression, part II: multiple linear regression.
Marill, Keith A
2004-01-01
The applications of simple linear regression in medical research are limited, because in most situations, there are multiple relevant predictor variables. Univariate statistical techniques such as simple linear regression use a single predictor variable, and they often may be mathematically correct but clinically misleading. Multiple linear regression is a mathematical technique used to model the relationship between multiple independent predictor variables and a single dependent outcome variable. It is used in medical research to model observational data, as well as in diagnostic and therapeutic studies in which the outcome is dependent on more than one factor. Although the technique generally is limited to data that can be expressed with a linear function, it benefits from a well-developed mathematical framework that yields unique solutions and exact confidence intervals for regression coefficients. Building on Part I of this series, this article acquaints the reader with some of the important concepts in multiple regression analysis. These include multicollinearity, interaction effects, and an expansion of the discussion of inference testing, leverage, and variable transformations to multivariate models. Examples from the first article in this series are expanded on using a primarily graphic, rather than mathematical, approach. The importance of the relationships among the predictor variables and the dependence of the multivariate model coefficients on the choice of these variables are stressed. Finally, concepts in regression model building are discussed.
NASA Astrophysics Data System (ADS)
Kang, Pilsang; Koo, Changhoi; Roh, Hokyu
2017-11-01
Since simple linear regression theory was established at the beginning of the 1900s, it has been used in a variety of fields. Unfortunately, it cannot be used directly for calibration. In practical calibrations, the observed measurements (the inputs) are subject to errors, and hence they vary, thus violating the assumption that the inputs are fixed. Therefore, in the case of calibration, the regression line fitted using the method of least squares is not consistent with the statistical properties of simple linear regression as already established based on this assumption. To resolve this problem, "classical regression" and "inverse regression" have been proposed. However, they do not completely resolve the problem. As a fundamental solution, we introduce "reversed inverse regression" along with a new methodology for deriving its statistical properties. In this study, the statistical properties of this regression are derived using the "error propagation rule" and the "method of simultaneous error equations" and are compared with those of the existing regression approaches. The accuracy of the statistical properties thus derived is investigated in a simulation study. We conclude that the newly proposed regression and methodology constitute the complete regression approach for univariate linear calibrations.
Spacecraft platform cost estimating relationships
NASA Technical Reports Server (NTRS)
Gruhl, W. M.
1972-01-01
The three main cost areas of unmanned satellite development are discussed. The areas are identified as: (1) the spacecraft platform (SCP), (2) the payload or experiments, and (3) the postlaunch ground equipment and operations. The SCP normally accounts for over half of the total project cost and accurate estimates of SCP costs are required early in project planning as a basis for determining total project budget requirements. The development of single formula SCP cost estimating relationships (CER) from readily available data by statistical linear regression analysis is described. The advantages of single formula CER are presented.
Savjani, Ricky R; Taylor, Brian A; Acion, Laura; Wilde, Elisabeth A; Jorge, Ricardo E
2017-11-15
Finding objective and quantifiable imaging markers of mild traumatic brain injury (TBI) has proven challenging, especially in the military population. Changes in cortical thickness after injury have been reported in animals and in humans, but it is unclear how these alterations manifest in the chronic phase, and it is difficult to characterize accurately with imaging. We used cortical thickness measures derived from Advanced Normalization Tools (ANTs) to predict a continuous demographic variable: age. We trained four different regression models (linear regression, support vector regression, Gaussian process regression, and random forests) to predict age from healthy control brains from publicly available datasets (n = 762). We then used these models to predict brain age in military Service Members with TBI (n = 92) and military Service Members without TBI (n = 34). Our results show that all four models overpredicted age in Service Members with TBI, and the predicted age difference was significantly greater compared with military controls. These data extend previous civilian findings and show that cortical thickness measures may reveal an association of accelerated changes over time with military TBI.
Vajargah, Kianoush Fathi; Sadeghi-Bazargani, Homayoun; Mehdizadeh-Esfanjani, Robab; Savadi-Oskouei, Daryoush; Farhoudi, Mehdi
2012-01-01
The objective of the present study was to assess the comparable applicability of orthogonal projections to latent structures (OPLS) statistical model vs traditional linear regression in order to investigate the role of trans cranial doppler (TCD) sonography in predicting ischemic stroke prognosis. The study was conducted on 116 ischemic stroke patients admitted to a specialty neurology ward. The Unified Neurological Stroke Scale was used once for clinical evaluation on the first week of admission and again six months later. All data was primarily analyzed using simple linear regression and later considered for multivariate analysis using PLS/OPLS models through the SIMCA P+12 statistical software package. The linear regression analysis results used for the identification of TCD predictors of stroke prognosis were confirmed through the OPLS modeling technique. Moreover, in comparison to linear regression, the OPLS model appeared to have higher sensitivity in detecting the predictors of ischemic stroke prognosis and detected several more predictors. Applying the OPLS model made it possible to use both single TCD measures/indicators and arbitrarily dichotomized measures of TCD single vessel involvement as well as the overall TCD result. In conclusion, the authors recommend PLS/OPLS methods as complementary rather than alternative to the available classical regression models such as linear regression.
Interpretation of commonly used statistical regression models.
Kasza, Jessica; Wolfe, Rory
2014-01-01
A review of some regression models commonly used in respiratory health applications is provided in this article. Simple linear regression, multiple linear regression, logistic regression and ordinal logistic regression are considered. The focus of this article is on the interpretation of the regression coefficients of each model, which are illustrated through the application of these models to a respiratory health research study. © 2013 The Authors. Respirology © 2013 Asian Pacific Society of Respirology.
Determinants of spirometric abnormalities among silicotic patients in Hong Kong.
Leung, Chi C; Chang, Kwok C; Law, Wing S; Yew, Wing W; Tam, Cheuk M; Chan, Chi K; Wong, Man Y
2005-09-01
Silicosis is the second commonest notified occupational disease in Hong Kong. To characterize the determinants of spirometric abnormalities in silicosis. The spirometric patterns of consecutive silicotic patients on confirmation by the Pneumoconiosis Medical Board from 1991 to 2002 were correlated with demographic characteristics, occupational history, smoking history, tuberculosis (TB) history and radiographic features by univariate and multiple regression analyses. Of 1576 silicotic patients included, 55.6% showed normal spirometry, 28.5% normal forced vital capacity (FVC>or=80% predicted) but reduced forced expiratory ratio (FER<70%), 7.6% reduced FVC but normal FER, and 8.4% reduced both FVC and FER. Age, ever-smoking, cigarette pack-years, industry, job type, history of TB, size of lung nodules and progressive massive fibrosis (PMF) were all significantly associated with airflow limitation on univariate analysis (all P<0.05), while sex and profusion of nodules were not. Only age, cigarette pack-years, history of TB, size of lung nodules and PMF remained as significant independent predictors of airflow obstruction in multiple logistic regression analysis. After controlling for airflow obstruction, only shorter exposure duration, history of TB and profusion of nodules were significant independent predictors of reduced FVC. As well as age, history of TB, cigarette pack-years, PMF and nodule size contributed comparable effects to airflow obstruction in multiple linear regression analyses, while profusion of nodules was the strongest factor for reduced vital capacity. In an occupational compensation setting, disease indices and history of tuberculosis are independent predictors of both airflow obstruction and reduced vital capacity for silicotic patients.
Multiple imputation for cure rate quantile regression with censored data.
Wu, Yuanshan; Yin, Guosheng
2017-03-01
The main challenge in the context of cure rate analysis is that one never knows whether censored subjects are cured or uncured, or whether they are susceptible or insusceptible to the event of interest. Considering the susceptible indicator as missing data, we propose a multiple imputation approach to cure rate quantile regression for censored data with a survival fraction. We develop an iterative algorithm to estimate the conditionally uncured probability for each subject. By utilizing this estimated probability and Bernoulli sample imputation, we can classify each subject as cured or uncured, and then employ the locally weighted method to estimate the quantile regression coefficients with only the uncured subjects. Repeating the imputation procedure multiple times and taking an average over the resultant estimators, we obtain consistent estimators for the quantile regression coefficients. Our approach relaxes the usual global linearity assumption, so that we can apply quantile regression to any particular quantile of interest. We establish asymptotic properties for the proposed estimators, including both consistency and asymptotic normality. We conduct simulation studies to assess the finite-sample performance of the proposed multiple imputation method and apply it to a lung cancer study as an illustration. © 2016, The International Biometric Society.
Use of probabilistic weights to enhance linear regression myoelectric control
NASA Astrophysics Data System (ADS)
Smith, Lauren H.; Kuiken, Todd A.; Hargrove, Levi J.
2015-12-01
Objective. Clinically available prostheses for transradial amputees do not allow simultaneous myoelectric control of degrees of freedom (DOFs). Linear regression methods can provide simultaneous myoelectric control, but frequently also result in difficulty with isolating individual DOFs when desired. This study evaluated the potential of using probabilistic estimates of categories of gross prosthesis movement, which are commonly used in classification-based myoelectric control, to enhance linear regression myoelectric control. Approach. Gaussian models were fit to electromyogram (EMG) feature distributions for three movement classes at each DOF (no movement, or movement in either direction) and used to weight the output of linear regression models by the probability that the user intended the movement. Eight able-bodied and two transradial amputee subjects worked in a virtual Fitts’ law task to evaluate differences in controllability between linear regression and probability-weighted regression for an intramuscular EMG-based three-DOF wrist and hand system. Main results. Real-time and offline analyses in able-bodied subjects demonstrated that probability weighting improved performance during single-DOF tasks (p < 0.05) by preventing extraneous movement at additional DOFs. Similar results were seen in experiments with two transradial amputees. Though goodness-of-fit evaluations suggested that the EMG feature distributions showed some deviations from the Gaussian, equal-covariance assumptions used in this experiment, the assumptions were sufficiently met to provide improved performance compared to linear regression control. Significance. Use of probability weights can improve the ability to isolate individual during linear regression myoelectric control, while maintaining the ability to simultaneously control multiple DOFs.
Caravaggi, Paolo; Leardini, Alberto; Giacomozzi, Claudia
2016-10-03
Plantar load can be considered as a measure of the foot ability to transmit forces at the foot/ground, or foot/footwear interface during ambulatory activities via the lower limb kinematic chain. While morphological and functional measures have been shown to be correlated with plantar load, no exhaustive data are currently available on the possible relationships between range of motion of foot joints and plantar load regional parameters. Joints' kinematics from a validated multi-segmental foot model were recorded together with plantar pressure parameters in 21 normal-arched healthy subjects during three barefoot walking trials. Plantar pressure maps were divided into six anatomically-based regions of interest associated to corresponding foot segments. A stepwise multiple regression analysis was performed to determine the relationships between pressure-based parameters, joints range of motion and normalized walking speed (speed/subject height). Sagittal- and frontal-plane joint motion were those most correlated to plantar load. Foot joints' range of motion and normalized walking speed explained between 6% and 43% of the model variance (adjusted R 2 ) for pressure-based parameters. In general, those joints' presenting lower mobility during stance were associated to lower vertical force at forefoot and to larger mean and peak pressure at hindfoot and forefoot. Normalized walking speed was always positively correlated to mean and peak pressure at hindfoot and forefoot. While a large variance in plantar pressure data is still not accounted for by the present models, this study provides statistical corroboration of the close relationship between joint mobility and plantar pressure during stance in the normal healthy foot. Copyright © 2016 Elsevier Ltd. All rights reserved.
Gierlinger, Notburga; Luss, Saskia; König, Christian; Konnerth, Johannes; Eder, Michaela; Fratzl, Peter
2010-01-01
The functional characteristics of plant cell walls depend on the composition of the cell wall polymers, as well as on their highly ordered architecture at scales from a few nanometres to several microns. Raman spectra of wood acquired with linear polarized laser light include information about polymer composition as well as the alignment of cellulose microfibrils with respect to the fibre axis (microfibril angle). By changing the laser polarization direction in 3° steps, the dependency between cellulose and laser orientation direction was investigated. Orientation-dependent changes of band height ratios and spectra were described by quadratic linear regression and partial least square regressions, respectively. Using the models and regressions with high coefficients of determination (R2 > 0.99) microfibril orientation was predicted in the S1 and S2 layers distinguished by the Raman imaging approach in cross-sections of spruce normal, opposite, and compression wood. The determined microfibril angle (MFA) in the different S2 layers ranged from 0° to 49.9° and was in coincidence with X-ray diffraction determination. With the prerequisite of geometric sample and laser alignment, exact MFA prediction can complete the picture of the chemical cell wall design gained by the Raman imaging approach at the micron level in all plant tissues. PMID:20007198
Simplified large African carnivore density estimators from track indices.
Winterbach, Christiaan W; Ferreira, Sam M; Funston, Paul J; Somers, Michael J
2016-01-01
The range, population size and trend of large carnivores are important parameters to assess their status globally and to plan conservation strategies. One can use linear models to assess population size and trends of large carnivores from track-based surveys on suitable substrates. The conventional approach of a linear model with intercept may not intercept at zero, but may fit the data better than linear model through the origin. We assess whether a linear regression through the origin is more appropriate than a linear regression with intercept to model large African carnivore densities and track indices. We did simple linear regression with intercept analysis and simple linear regression through the origin and used the confidence interval for ß in the linear model y = αx + ß, Standard Error of Estimate, Mean Squares Residual and Akaike Information Criteria to evaluate the models. The Lion on Clay and Low Density on Sand models with intercept were not significant ( P > 0.05). The other four models with intercept and the six models thorough origin were all significant ( P < 0.05). The models using linear regression with intercept all included zero in the confidence interval for ß and the null hypothesis that ß = 0 could not be rejected. All models showed that the linear model through the origin provided a better fit than the linear model with intercept, as indicated by the Standard Error of Estimate and Mean Square Residuals. Akaike Information Criteria showed that linear models through the origin were better and that none of the linear models with intercept had substantial support. Our results showed that linear regression through the origin is justified over the more typical linear regression with intercept for all models we tested. A general model can be used to estimate large carnivore densities from track densities across species and study areas. The formula observed track density = 3.26 × carnivore density can be used to estimate densities of large African carnivores using track counts on sandy substrates in areas where carnivore densities are 0.27 carnivores/100 km 2 or higher. To improve the current models, we need independent data to validate the models and data to test for non-linear relationship between track indices and true density at low densities.
Modeling Laterality of the Globus Pallidus Internus in Patients With Parkinson's Disease.
Sharim, Justin; Yazdi, Daniel; Baohan, Amy; Behnke, Eric; Pouratian, Nader
2017-04-01
Neurosurgical interventions such as deep brain stimulation surgery of the globus pallidus internus (GPi) play an important role in the treatment of medically refractory Parkinson's disease (PD), and require high targeting accuracy. Variability in the laterality of the GPi across patients with PD has not been well characterized. The aim of this report is to identify factors that may contribute to differences in position of the motor region of GPi. The charts and operative reports of 101 PD patients following deep brain stimulation surgery (70 males, aged 11-78 years) representing 201 GPi were retrospectively reviewed. Data extracted for each subject include age, gender, anterior and posterior commissures (AC-PC) distance, and third ventricular width. Multiple linear regression, stepwise regression, and relative importance of regressors analysis were performed to assess the predictive ability of these variables on GPi laterality. Multiple linear regression for target vs. third ventricular width, gender, AC-PC distance, and age were significant for normalized linear regression coefficients of 0.333 (p < 0.0001), 0.206 (p = 0.00219), 0.168 (p = 0.0119), and 0.159 (p = 0.0136), respectively. Third ventricular width, gender, AC-PC distance, and age each account for 44.06% (21.38-65.69%, 95% CI), 20.82% (10.51-35.88%), 21.46% (8.28-37.05%), and 13.66% (2.62-28.64%) of the R 2 value, respectively. Effect size calculation was significant for a change in the GPi laterality of 0.19 mm per mm of ventricular width, 0.11 mm per mm of AC-PC distance, 0.017 mm per year in age, and 0.54 mm increase for male gender. This variability highlights the limitations of indirect targeting alone, and argues for the continued use of MRI as well as intraoperative physiological testing to account for such factors that contribute to patient-specific variability in GPi localization. © 2016 International Neuromodulation Society.
Geszke-Moritz, Małgorzata; Moritz, Michał
2016-12-01
The present study deals with the adsorption of boldine onto pure and propyl-sulfonic acid-functionalized SBA-15, SBA-16 and mesocellular foam (MCF) materials. Siliceous adsorbents were characterized by nitrogen sorption analysis, transmission electron microscopy (TEM), scanning electron microscopy (SEM), Fourier-transform infrared (FT-IR) spectroscopy and thermogravimetric analysis. The equilibrium adsorption data were analyzed using the Langmuir, Freundlich, Redlich-Peterson, and Temkin isotherms. Moreover, the Dubinin-Radushkevich and Dubinin-Astakhov isotherm models based on the Polanyi adsorption potential were employed. The latter was calculated using two alternative formulas including solubility-normalized (S-model) and empirical C-model. In order to find the best-fit isotherm, both linear regression and nonlinear fitting analysis were carried out. The Dubinin-Astakhov (S-model) isotherm revealed the best fit to the experimental points for adsorption of boldine onto pure mesoporous materials using both linear and nonlinear fitting analysis. Meanwhile, the process of boldine sorption onto modified silicas was described the best by the Langmuir and Temkin isotherms using linear regression and nonlinear fitting analysis, respectively. The values of adsorption energy (below 8kJ/mol) indicate the physical nature of boldine adsorption onto unmodified silicas whereas the ionic interactions seem to be the main force of alkaloid adsorption onto functionalized sorbents (energy of adsorption above 8kJ/mol). Copyright © 2016 Elsevier B.V. All rights reserved.
García-Diego, Fernando-Juan; Sánchez-Quinche, Angel; Merello, Paloma; Beltrán, Pedro; Peris, Cristófol
2013-01-01
In this study we propose an electronic system for linear positioning of a magnet independent of its modulus, which could vary because of aging, different fabrication process, etc. The system comprises a linear array of 24 Hall Effect sensors of proportional response. The data from all sensors are subject to a pretreatment (normalization) by row (position) making them independent on the temporary variation of its magnetic field strength. We analyze the particular case of the individual flow in milking of goats. The multiple regression analysis allowed us to calibrate the electronic system with a percentage of explanation R2 = 99.96%. In our case, the uncertainty in the linear position of the magnet is 0.51 mm that represents 0.019 L of goat milk. The test in farm compared the results obtained by direct reading of the volume with those obtained by the proposed electronic calibrated system, achieving a percentage of explanation of 99.05%. PMID:23793020
On the use of log-transformation vs. nonlinear regression for analyzing biological power laws.
Xiao, Xiao; White, Ethan P; Hooten, Mevin B; Durham, Susan L
2011-10-01
Power-law relationships are among the most well-studied functional relationships in biology. Recently the common practice of fitting power laws using linear regression (LR) on log-transformed data has been criticized, calling into question the conclusions of hundreds of studies. It has been suggested that nonlinear regression (NLR) is preferable, but no rigorous comparison of these two methods has been conducted. Using Monte Carlo simulations, we demonstrate that the error distribution determines which method performs better, with NLR better characterizing data with additive, homoscedastic, normal error and LR better characterizing data with multiplicative, heteroscedastic, lognormal error. Analysis of 471 biological power laws shows that both forms of error occur in nature. While previous analyses based on log-transformation appear to be generally valid, future analyses should choose methods based on a combination of biological plausibility and analysis of the error distribution. We provide detailed guidelines and associated computer code for doing so, including a model averaging approach for cases where the error structure is uncertain.
Miller, Nathan; Prevatt, Frances
2017-10-01
The purpose of this study was to reexamine the latent structure of ADHD and sluggish cognitive tempo (SCT) due to issues with construct validity. Two proposed changes to the construct include viewing hyperactivity and sluggishness (hypoactivity) as a single continuum of activity level, and viewing inattention as a separate dimension from activity level. Data were collected from 1,398 adults using Amazon's MTurk. A new scale measuring activity level was developed, and scores of Inattention were regressed onto scores of Activity Level using curvilinear regression. The Activity Level scale showed acceptable levels of internal consistency, normality, and unimodality. Curvilinear regression indicates that a quadratic (curvilinear) model accurately explains a small but significant portion of the variance in levels of inattention. Hyperactivity and hypoactivity may be viewed as a continuum, rather than separate disorders. Inattention may have a U-shaped relationship with activity level. Linear analyses may be insufficient and inaccurate for studying ADHD.
Sonographic study of the development of fetal corpus callosum in a Chinese population.
Zhang, Hai-chun; Yang, Jie; Chen, Zhong-ping; Ma, Xiao-yan
2009-02-01
The observation of fetal corpus callosum (CC) is important for the prenatal sonographic assessment of fetal central nervous system development. The aim of this study was to investigate the development of normal Chinese fetal CC. CC measurements were performed using high-resolution transabdominal sonography on 622 Chinese fetuses between 16 and 39 weeks' gestation. The correlation between CC size and gestational age was investigated. The fetal CC length increased in a linear fashion during pregnancy. The length of the CC as a function of gestational age was expressed by the following regression equation: length (mm) = -9.567 + 1.495 x gestational age (weeks) (r = 0.932, p < 0.001). Knowledge of normal CC appearance may help identify developmental anomalies and enable accurate prenatal counseling. (c) 2008 Wiley Periodicals, Inc.
Hemmila, April; McGill, Jim; Ritter, David
2008-03-01
To determine if changes in fingerprint infrared spectra linear with age can be found, partial least squares (PLS1) regression of 155 fingerprint infrared spectra against the person's age was constructed. The regression produced a linear model of age as a function of spectrum with a root mean square error of calibration of less than 4 years, showing an inflection at about 25 years of age. The spectral ranges emphasized by the regression do not correspond to the highest concentration constituents of the fingerprints. Separate linear regression models for old and young people can be constructed with even more statistical rigor. The success of the regression demonstrates that a combination of constituents can be found that changes linearly with age, with a significant shift around puberty.
Gimelfarb, A.; Willis, J. H.
1994-01-01
An experiment was conducted to investigate the offspring-parent regression for three quantitative traits (weight, abdominal bristles and wing length) in Drosophila melanogaster. Linear and polynomial models were fitted for the regressions of a character in offspring on both parents. It is demonstrated that responses by the characters to selection predicted by the nonlinear regressions may differ substantially from those predicted by the linear regressions. This is true even, and especially, if selection is weak. The realized heritability for a character under selection is shown to be determined not only by the offspring-parent regression but also by the distribution of the character and by the form and strength of selection. PMID:7828818
Linear and nonlinear regression techniques for simultaneous and proportional myoelectric control.
Hahne, J M; Biessmann, F; Jiang, N; Rehbaum, H; Farina, D; Meinecke, F C; Muller, K-R; Parra, L C
2014-03-01
In recent years the number of active controllable joints in electrically powered hand-prostheses has increased significantly. However, the control strategies for these devices in current clinical use are inadequate as they require separate and sequential control of each degree-of-freedom (DoF). In this study we systematically compare linear and nonlinear regression techniques for an independent, simultaneous and proportional myoelectric control of wrist movements with two DoF. These techniques include linear regression, mixture of linear experts (ME), multilayer-perceptron, and kernel ridge regression (KRR). They are investigated offline with electro-myographic signals acquired from ten able-bodied subjects and one person with congenital upper limb deficiency. The control accuracy is reported as a function of the number of electrodes and the amount and diversity of training data providing guidance for the requirements in clinical practice. The results showed that KRR, a nonparametric statistical learning method, outperformed the other methods. However, simple transformations in the feature space could linearize the problem, so that linear models could achieve similar performance as KRR at much lower computational costs. Especially ME, a physiologically inspired extension of linear regression represents a promising candidate for the next generation of prosthetic devices.
Unitary Response Regression Models
ERIC Educational Resources Information Center
Lipovetsky, S.
2007-01-01
The dependent variable in a regular linear regression is a numerical variable, and in a logistic regression it is a binary or categorical variable. In these models the dependent variable has varying values. However, there are problems yielding an identity output of a constant value which can also be modelled in a linear or logistic regression with…
NASA Astrophysics Data System (ADS)
Hasan, Haliza; Ahmad, Sanizah; Osman, Balkish Mohd; Sapri, Shamsiah; Othman, Nadirah
2017-08-01
In regression analysis, missing covariate data has been a common problem. Many researchers use ad hoc methods to overcome this problem due to the ease of implementation. However, these methods require assumptions about the data that rarely hold in practice. Model-based methods such as Maximum Likelihood (ML) using the expectation maximization (EM) algorithm and Multiple Imputation (MI) are more promising when dealing with difficulties caused by missing data. Then again, inappropriate methods of missing value imputation can lead to serious bias that severely affects the parameter estimates. The main objective of this study is to provide a better understanding regarding missing data concept that can assist the researcher to select the appropriate missing data imputation methods. A simulation study was performed to assess the effects of different missing data techniques on the performance of a regression model. The covariate data were generated using an underlying multivariate normal distribution and the dependent variable was generated as a combination of explanatory variables. Missing values in covariate were simulated using a mechanism called missing at random (MAR). Four levels of missingness (10%, 20%, 30% and 40%) were imposed. ML and MI techniques available within SAS software were investigated. A linear regression analysis was fitted and the model performance measures; MSE, and R-Squared were obtained. Results of the analysis showed that MI is superior in handling missing data with highest R-Squared and lowest MSE when percent of missingness is less than 30%. Both methods are unable to handle larger than 30% level of missingness.
Yin, Tai-lang; Zhang, Yi; Li, Sai-jiao; Zhao, Meng; Ding, Jin-li; Xu, Wang-ming; Yang, Jing
2015-12-01
Whether the type of culture media utilized in assisted reproductive technology has impacts on laboratory outcomes and birth weight of newborns in in-vitro fertilization (IVF)/intracytoplasmic sperm injection (ICSI) was investigated. A total of 673 patients undergoing IVF/ICSI and giving birth to live singletons after fresh embryo transfer on day 3 from Jan. 1, 2010 to Dec. 31, 2012 were included. Three types of culture media were used during this period: Quinn's Advantage (QA), Single Step Medium (SSM), and Continuous Single Culture medium (CSC). Fertilization rate (FR), normal fertilization rate (NFR), cleavage rate (CR), normal cleavage rate (NCR), good-quality embryo rate (GQER) and neonatal birth weight were compared using one-way ANOVA and χ (2) tests. Multiple linear regression analysis was performed to determine the impact of culture media on laboratory outcomes and birth weight. In IVF cycles, GQER was significantly decreased in SSM medium group as compared with QA or CSC media groups (63.6% vs. 69.0% in QA; vs. 71.3% in CSC, P=0.011). In ICSI cycles, FR, NFR and CR were significantly lower in CSC medium group than in other two media groups. No significant difference was observed in neonatal birthweight among the three groups (P=0.759). Multiple linear regression analyses confirmed that the type of culture medium was correlated with FR, NFR, CR and GQER, but not with neonatal birth weight. The type of culture media had potential influences on laboratory outcomes but did not exhibit an impact on the birth weight of singletons in ART.
Competing regression models for longitudinal data.
Alencar, Airlane P; Singer, Julio M; Rocha, Francisco Marcelo M
2012-03-01
The choice of an appropriate family of linear models for the analysis of longitudinal data is often a matter of concern for practitioners. To attenuate such difficulties, we discuss some issues that emerge when analyzing this type of data via a practical example involving pretest-posttest longitudinal data. In particular, we consider log-normal linear mixed models (LNLMM), generalized linear mixed models (GLMM), and models based on generalized estimating equations (GEE). We show how some special features of the data, like a nonconstant coefficient of variation, may be handled in the three approaches and evaluate their performance with respect to the magnitude of standard errors of interpretable and comparable parameters. We also show how different diagnostic tools may be employed to identify outliers and comment on available software. We conclude by noting that the results are similar, but that GEE-based models may be preferable when the goal is to compare the marginal expected responses. © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Identification of internal properties of fibers and micro-swimmers
NASA Astrophysics Data System (ADS)
Plouraboue, Franck; Thiam, Ibrahima; Delmotte, Blaise; Climent, Eric; PSC Collaboration
2016-11-01
In this presentation we discuss the identifiability of constitutive parameters of passive or active micro-swimmers. We first present a general framework for describing fibers or micro-swimmers using a bead-model description. Using a kinematic constraint formulation to describe fibers, flagellum or cilia, we find explicit linear relationship between elastic constitutive parameters and generalised velocities from computing contact forces. This linear formulation then permits to address explicitly identifiability conditions and solve for parameter identification. We show that both active forcing and passive parameters are both identifiable independently but not simultaneously. We also provide unbiased estimators for elastic parameters as well as active ones in the presence of Langevin-like forcing with Gaussian noise using normal linear regression models and maximum likelihood method. These theoretical results are illustrated in various configurations of relaxed or actuated passives fibers, and active filament of known passive properties, showing the efficiency of the proposed approach for direct parameter identification. The convergence of the proposed estimators is successfully tested numerically.
Multiple Regression Analysis of mRNA-miRNA Associations in Colorectal Cancer Pathway
Wang, Fengfeng; Wong, S. C. Cesar; Chan, Lawrence W. C.; Cho, William C. S.; Yip, S. P.; Yung, Benjamin Y. M.
2014-01-01
Background. MicroRNA (miRNA) is a short and endogenous RNA molecule that regulates posttranscriptional gene expression. It is an important factor for tumorigenesis of colorectal cancer (CRC), and a potential biomarker for diagnosis, prognosis, and therapy of CRC. Our objective is to identify the related miRNAs and their associations with genes frequently involved in CRC microsatellite instability (MSI) and chromosomal instability (CIN) signaling pathways. Results. A regression model was adopted to identify the significantly associated miRNAs targeting a set of candidate genes frequently involved in colorectal cancer MSI and CIN pathways. Multiple linear regression analysis was used to construct the model and find the significant mRNA-miRNA associations. We identified three significantly associated mRNA-miRNA pairs: BCL2 was positively associated with miR-16 and SMAD4 was positively associated with miR-567 in the CRC tissue, while MSH6 was positively associated with miR-142-5p in the normal tissue. As for the whole model, BCL2 and SMAD4 models were not significant, and MSH6 model was significant. The significant associations were different in the normal and the CRC tissues. Conclusion. Our results have laid down a solid foundation in exploration of novel CRC mechanisms, and identification of miRNA roles as oncomirs or tumor suppressor mirs in CRC. PMID:24895601
Zinc Levels in Left Ventricular Hypertrophy.
Huang, Lei; Teng, Tianming; Bian, Bo; Yao, Wei; Yu, Xuefang; Wang, Zhuoqun; Xu, Zhelong; Sun, Yuemin
2017-03-01
Zinc is one of the most important trace elements in the body and zinc homeostasis plays a critical role in maintaining cellular structure and function. Zinc dyshomeostasis can lead to many diseases, such as cardiovascular disease. Our aim was to investigate whether there is a relationship between zinc and left ventricular hypertrophy (LVH). A total of 519 patients was enrolled and their serum zinc levels were measured in this study. We performed analyses on the relationship between zinc levels and LVH and the four LV geometry pattern patients: normal LV geometry, concentric remodeling, eccentric LVH, and concentric LVH. We performed further linear and multiple regression analyses to confirm the relationship between zinc and left ventricular mass (LVM), left ventricular mass index (LVMI), and relative wall thickness (RWT). Our data showed that zinc levels were 710.2 ± 243.0 μg/L in the control group and were 641.9 ± 215.2 μg/L in LVH patients. We observed that zinc levels were 715 ± 243.5 μg/L, 694.2 ± 242.7 μg/L, 643.7 ± 225.0 μg/L, and 638.7 ± 197.0 μg/L in normal LV geometry, concentric remodeling, eccentric LVH, and concentric LVH patients, respectively. We further found that there was a significant inverse linear relationship between zinc and LVM (p = 0.001) and LVMI (p = 0.000) but did not show a significant relationship with RWT (p = 0.561). Multiple regression analyses confirmed that the linear relationship between zinc and LVM and LVMI remained inversely significant. The present study revealed that serum zinc levels were significantly decreased in the LVH patients, especially in the eccentric LVH and concentric LVH patients. Furthermore, zinc levels were significantly inversely correlated with LVM and LVMI.
Compound Identification Using Penalized Linear Regression on Metabolomics
Liu, Ruiqi; Wu, Dongfeng; Zhang, Xiang; Kim, Seongho
2014-01-01
Compound identification is often achieved by matching the experimental mass spectra to the mass spectra stored in a reference library based on mass spectral similarity. Because the number of compounds in the reference library is much larger than the range of mass-to-charge ratio (m/z) values so that the data become high dimensional data suffering from singularity. For this reason, penalized linear regressions such as ridge regression and the lasso are used instead of the ordinary least squares regression. Furthermore, two-step approaches using the dot product and Pearson’s correlation along with the penalized linear regression are proposed in this study. PMID:27212894
Jones, C Jessie; Rutledge, Dana N; Aquino, Jordan
2010-07-01
The purposes of this study were to determine whether people with and without fibromyalgia (FM) age 50 yr and above showed differences in physical performance and perceived functional ability and to determine whether age, gender, depression, and physical activity level altered the impact of FM status on these factors. Dependent variables included perceived function and 6 performance measures (multidimensional balance, aerobic endurance, overall functional mobility, lower body strength, and gait velocity-normal or fast). Independent (predictor) variables were FM status, age, gender, depression, and physical activity level. Results indicated significant differences between adults with and without FM on all physical-performance measures and perceived function. Linear-regression models showed that the contribution of significant predictors was in expected directions. All regression models were significant, accounting for 16-65% of variance in the dependent variables.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kwon, Deukwoo; Little, Mark P.; Miller, Donald L.
Purpose: To determine more accurate regression formulas for estimating peak skin dose (PSD) from reference air kerma (RAK) or kerma-area product (KAP). Methods: After grouping of the data from 21 procedures into 13 clinically similar groups, assessments were made of optimal clustering using the Bayesian information criterion to obtain the optimal linear regressions of (log-transformed) PSD vs RAK, PSD vs KAP, and PSD vs RAK and KAP. Results: Three clusters of clinical groups were optimal in regression of PSD vs RAK, seven clusters of clinical groups were optimal in regression of PSD vs KAP, and six clusters of clinical groupsmore » were optimal in regression of PSD vs RAK and KAP. Prediction of PSD using both RAK and KAP is significantly better than prediction of PSD with either RAK or KAP alone. The regression of PSD vs RAK provided better predictions of PSD than the regression of PSD vs KAP. The partial-pooling (clustered) method yields smaller mean squared errors compared with the complete-pooling method.Conclusion: PSD distributions for interventional radiology procedures are log-normal. Estimates of PSD derived from RAK and KAP jointly are most accurate, followed closely by estimates derived from RAK alone. Estimates of PSD derived from KAP alone are the least accurate. Using a stochastic search approach, it is possible to cluster together certain dissimilar types of procedures to minimize the total error sum of squares.« less
Control Variate Selection for Multiresponse Simulation.
1987-05-01
M. H. Knuter, Applied Linear Regression Mfodels, Richard D. Erwin, Inc., Homewood, Illinois, 1983. Neuts, Marcel F., Probability, Allyn and Bacon...1982. Neter, J., V. Wasserman, and M. H. Knuter, Applied Linear Regression .fodels, Richard D. Erwin, Inc., Homewood, Illinois, 1983. Neuts, Marcel F...Aspects of J%,ultivariate Statistical Theory, John Wiley and Sons, New York, New York, 1982. dY Neter, J., W. Wasserman, and M. H. Knuter, Applied Linear Regression Mfodels
ERIC Educational Resources Information Center
Kobrin, Jennifer L.; Sinharay, Sandip; Haberman, Shelby J.; Chajewski, Michael
2011-01-01
This study examined the adequacy of a multiple linear regression model for predicting first-year college grade point average (FYGPA) using SAT[R] scores and high school grade point average (HSGPA). A variety of techniques, both graphical and statistical, were used to examine if it is possible to improve on the linear regression model. The results…
Quantile Regression in the Study of Developmental Sciences
Petscher, Yaacov; Logan, Jessica A. R.
2014-01-01
Linear regression analysis is one of the most common techniques applied in developmental research, but only allows for an estimate of the average relations between the predictor(s) and the outcome. This study describes quantile regression, which provides estimates of the relations between the predictor(s) and outcome, but across multiple points of the outcome’s distribution. Using data from the High School and Beyond and U.S. Sustained Effects Study databases, quantile regression is demonstrated and contrasted with linear regression when considering models with: (a) one continuous predictor, (b) one dichotomous predictor, (c) a continuous and a dichotomous predictor, and (d) a longitudinal application. Results from each example exhibited the differential inferences which may be drawn using linear or quantile regression. PMID:24329596
Misyura, Maksym; Sukhai, Mahadeo A; Kulasignam, Vathany; Zhang, Tong; Kamel-Reid, Suzanne; Stockley, Tracy L
2018-01-01
Aims A standard approach in test evaluation is to compare results of the assay in validation to results from previously validated methods. For quantitative molecular diagnostic assays, comparison of test values is often performed using simple linear regression and the coefficient of determination (R2), using R2 as the primary metric of assay agreement. However, the use of R2 alone does not adequately quantify constant or proportional errors required for optimal test evaluation. More extensive statistical approaches, such as Bland-Altman and expanded interpretation of linear regression methods, can be used to more thoroughly compare data from quantitative molecular assays. Methods We present the application of Bland-Altman and linear regression statistical methods to evaluate quantitative outputs from next-generation sequencing assays (NGS). NGS-derived data sets from assay validation experiments were used to demonstrate the utility of the statistical methods. Results Both Bland-Altman and linear regression were able to detect the presence and magnitude of constant and proportional error in quantitative values of NGS data. Deming linear regression was used in the context of assay comparison studies, while simple linear regression was used to analyse serial dilution data. Bland-Altman statistical approach was also adapted to quantify assay accuracy, including constant and proportional errors, and precision where theoretical and empirical values were known. Conclusions The complementary application of the statistical methods described in this manuscript enables more extensive evaluation of performance characteristics of quantitative molecular assays, prior to implementation in the clinical molecular laboratory. PMID:28747393
A SEMIPARAMETRIC BAYESIAN MODEL FOR CIRCULAR-LINEAR REGRESSION
We present a Bayesian approach to regress a circular variable on a linear predictor. The regression coefficients are assumed to have a nonparametric distribution with a Dirichlet process prior. The semiparametric Bayesian approach gives added flexibility to the model and is usefu...
van Vught, Anneke J A H; Heitmann, Berit L; Nieuwenhuizen, Arie G; Veldhorst, Margriet A B; Andersen, Lars Bo; Hasselstrom, Henriette; Brummer, Robert-Jan M; Westerterp-Plantenga, Margriet S
2010-05-01
Growth hormone (GH) affects linear growth and body composition, by increasing the secretion of insulin-like growth factor-I (IGF-I), muscle protein synthesis and lipolysis. The intake of protein (PROT) as well as the specific amino acids arginine (ARG) and lysine (LYS) stimulates GH/IGF-I secretion. The present paper aimed to investigate associations between PROT intake as well as intake of the specific amino acids ARG and LYS, and subsequent 3-year-change in linear growth and body composition among 6-year-old children. Children's data were collected from Copenhagen (Denmark), during 2001-2002, and again 3 years later. Boys and girls were separated into normal weight and overweight, based on BMI quintiles. Fat-free mass index (FFMI) and fat mass index (FMI) were calculated. Associations between change (Delta) in height, FMI and FFMI, respectively, and habitual PROT intake as well as ARG and LYS were analysed by multiple linear regressions, adjusted for baseline height, FMI or FFMI and energy intake, age, physical activity and socio-economic status. Eighteen schools in two suburban communities in the Copenhagen (Denmark) area participated in the study. In all, 223 children's data were collected for the present study. High ARG intake was associated with linear growth (beta = 1.09 (se 0.54), P = 0.05) among girls. Furthermore, in girls, DeltaFMI had a stronger inverse association with high ARG intake, if it was combined with high LYS intake, instead of low LYS intake (P = 0.03). No associations were found in boys.ConclusionIn prepubertal girls, linear growth may be influenced by habitual ARG intake and body fat gain may be relatively prevented over time by the intake of the amino acids ARG and LYS.
Al-Shorman, Alaa; Al-Domi, Hayder; Al-Atoum, Muatasem
2018-06-01
Background Increased carotid intima-media thickness is one of the predictors of future cardiovascular diseases. However, it is still unknown which body composition parameter or anthropometric measure is the best predictor for carotid intima-media thickness change among children and young adolescents. Objective To investigate the associations of body composition and anthropometric measures with carotid intima-media thickness among a group of obese and normal bodyweight schoolchildren. Methods A total of 125 schoolchildren (10-15 years) were recruited from four public schools in Amman, Jordan. Of them, 60 (29 boys and 31 girls) were normal bodyweight students and 65 (35 boys and 30 girls) were obese students. Anthropometric measures, fat mass and fat-free mass were determined. Carotid intima-media thickness of the common artery was measured using high-resolution B-mode ultrasound. Results Compared to normal bodyweight students, obese participants exhibited greater carotid intima-media thickness (mm) (0.45 ± 0.10 vs. 0.38 ± 0.08, p = 0.002) and fat-free mass (kg) (48.01 ± 11.39 vs. 32.65 ±7.65, p < 0.001). Pearson's correlation coefficient and linear regression analysis revealed significant associations ( p≤0.05) between mean carotid intima-media thickness and body mass index, waist circumference, hip circumference, waist-to-hip ratio, fat mass and fat-free mass. Stepwise linear regression analysis revealed that waist circumference was the only measure that was statistically significant ( p ≤ 0.05) with mean carotid intima-media thickness (r 2 = 0.129, p = 0.002). Conclusions Obesity is related to greater carotid intima-media thickness and other cardiovascular risk factors among schoolchildren. Waist circumference is more sensitive in predicting increased carotid intima-media thickness than other body composition or anthropometric measures. Waist circumference measurement in the analysis of future studies assessing the cardiovascular risk among obese children is warranted.
Kumar, K Vasanth; Sivanesan, S
2006-08-25
Pseudo second order kinetic expressions of Ho, Sobkowsk and Czerwinski, Blanachard et al. and Ritchie were fitted to the experimental kinetic data of malachite green onto activated carbon by non-linear and linear method. Non-linear method was found to be a better way of obtaining the parameters involved in the second order rate kinetic expressions. Both linear and non-linear regression showed that the Sobkowsk and Czerwinski and Ritchie's pseudo second order model were the same. Non-linear regression analysis showed that both Blanachard et al. and Ho have similar ideas on the pseudo second order model but with different assumptions. The best fit of experimental data in Ho's pseudo second order expression by linear and non-linear regression method showed that Ho pseudo second order model was a better kinetic expression when compared to other pseudo second order kinetic expressions. The amount of dye adsorbed at equilibrium, q(e), was predicted from Ho pseudo second order expression and were fitted to the Langmuir, Freundlich and Redlich Peterson expressions by both linear and non-linear method to obtain the pseudo isotherms. The best fitting pseudo isotherm was found to be the Langmuir and Redlich Peterson isotherm. Redlich Peterson is a special case of Langmuir when the constant g equals unity.
2015-07-15
Long-term effects on cancer survivors’ quality of life of physical training versus physical training combined with cognitive-behavioral therapy ...COMPARISON OF NEURAL NETWORK AND LINEAR REGRESSION MODELS IN STATISTICALLY PREDICTING MENTAL AND PHYSICAL HEALTH STATUS OF BREAST...34Comparison of Neural Network and Linear Regression Models in Statistically Predicting Mental and Physical Health Status of Breast Cancer Survivors
Prediction of the Main Engine Power of a New Container Ship at the Preliminary Design Stage
NASA Astrophysics Data System (ADS)
Cepowski, Tomasz
2017-06-01
The paper presents mathematical relationships that allow us to forecast the estimated main engine power of new container ships, based on data concerning vessels built in 2005-2015. The presented approximations allow us to estimate the engine power based on the length between perpendiculars and the number of containers the ship will carry. The approximations were developed using simple linear regression and multivariate linear regression analysis. The presented relations have practical application for estimation of container ship engine power needed in preliminary parametric design of the ship. It follows from the above that the use of multiple linear regression to predict the main engine power of a container ship brings more accurate solutions than simple linear regression.
ERIC Educational Resources Information Center
Li, Deping; Oranje, Andreas
2007-01-01
Two versions of a general method for approximating standard error of regression effect estimates within an IRT-based latent regression model are compared. The general method is based on Binder's (1983) approach, accounting for complex samples and finite populations by Taylor series linearization. In contrast, the current National Assessment of…
Estimating linear temporal trends from aggregated environmental monitoring data
Erickson, Richard A.; Gray, Brian R.; Eager, Eric A.
2017-01-01
Trend estimates are often used as part of environmental monitoring programs. These trends inform managers (e.g., are desired species increasing or undesired species decreasing?). Data collected from environmental monitoring programs is often aggregated (i.e., averaged), which confounds sampling and process variation. State-space models allow sampling variation and process variations to be separated. We used simulated time-series to compare linear trend estimations from three state-space models, a simple linear regression model, and an auto-regressive model. We also compared the performance of these five models to estimate trends from a long term monitoring program. We specifically estimated trends for two species of fish and four species of aquatic vegetation from the Upper Mississippi River system. We found that the simple linear regression had the best performance of all the given models because it was best able to recover parameters and had consistent numerical convergence. Conversely, the simple linear regression did the worst job estimating populations in a given year. The state-space models did not estimate trends well, but estimated population sizes best when the models converged. We found that a simple linear regression performed better than more complex autoregression and state-space models when used to analyze aggregated environmental monitoring data.
Removing inter-subject technical variability in magnetic resonance imaging studies.
Fortin, Jean-Philippe; Sweeney, Elizabeth M; Muschelli, John; Crainiceanu, Ciprian M; Shinohara, Russell T
2016-05-15
Magnetic resonance imaging (MRI) intensities are acquired in arbitrary units, making scans non-comparable across sites and between subjects. Intensity normalization is a first step for the improvement of comparability of the images across subjects. However, we show that unwanted inter-scan variability associated with imaging site, scanner effect, and other technical artifacts is still present after standard intensity normalization in large multi-site neuroimaging studies. We propose RAVEL (Removal of Artificial Voxel Effect by Linear regression), a tool to remove residual technical variability after intensity normalization. As proposed by SVA and RUV [Leek and Storey, 2007, 2008, Gagnon-Bartsch and Speed, 2012], two batch effect correction tools largely used in genomics, we decompose the voxel intensities of images registered to a template into a biological component and an unwanted variation component. The unwanted variation component is estimated from a control region obtained from the cerebrospinal fluid (CSF), where intensities are known to be unassociated with disease status and other clinical covariates. We perform a singular value decomposition (SVD) of the control voxels to estimate factors of unwanted variation. We then estimate the unwanted factors using linear regression for every voxel of the brain and take the residuals as the RAVEL-corrected intensities. We assess the performance of RAVEL using T1-weighted (T1-w) images from more than 900 subjects with Alzheimer's disease (AD) and mild cognitive impairment (MCI), as well as healthy controls from the Alzheimer's Disease Neuroimaging Initiative (ADNI) database. We compare RAVEL to two intensity-normalization-only methods: histogram matching and White Stripe. We show that RAVEL performs best at improving the replicability of the brain regions that are empirically found to be most associated with AD, and that these regions are significantly more present in structures impacted by AD (hippocampus, amygdala, parahippocampal gyrus, enthorinal area, and fornix stria terminals). In addition, we show that the RAVEL-corrected intensities have the best performance in distinguishing between MCI subjects and healthy subjects using the mean hippocampal intensity (AUC=67%), a marked improvement compared to results from intensity normalization alone (AUC=63% and 59% for histogram matching and White Stripe, respectively). RAVEL is promising for many other imaging modalities. Published by Elsevier Inc.
A parameter estimation subroutine package
NASA Technical Reports Server (NTRS)
Bierman, G. J.; Nead, M. W.
1978-01-01
Linear least squares estimation and regression analyses continue to play a major role in orbit determination and related areas. A library of FORTRAN subroutines were developed to facilitate analyses of a variety of estimation problems. An easy to use, multi-purpose set of algorithms that are reasonably efficient and which use a minimal amount of computer storage are presented. Subroutine inputs, outputs, usage and listings are given, along with examples of how these routines can be used. The routines are compact and efficient and are far superior to the normal equation and Kalman filter data processing algorithms that are often used for least squares analyses.
Echocardiographic measurements of left ventricular mass by a non-geometric method
NASA Technical Reports Server (NTRS)
Parra, Beatriz; Buckey, Jay; Degraff, David; Gaffney, F. Andrew; Blomqvist, C. Gunnar
1987-01-01
The accuracy of a new nongeometric method for calculating left ventricular myocardial volumes from two-dimensional echocardiographic images was assessed in vitro using 20 formalin-fixed normal human hearts. Serial oblique short-axis images were acquired from one point at 5-deg intervals, for a total of 10-12 cross sections. Echocardiographic myocardial volumes were calculated as the difference between the volumes defined by the epi- and endocardial surfaces. Actual myocardial volumes were determined by water displacement. Volumes ranged from 80 to 174 ml (mean 130.8 ml). Linear regression analysis demonstrated excellent agreement between the echocardiographic and direct measurements.
Li, Wei; Qiu, Qi; Sun, Lin; Yue, Ling; Wang, Tao; Li, Xia; Xiao, Shifu
2017-01-01
Sex differences in Alzheimer's disease and mild cognitive impairment have been well recognized. However, sex differences in cognitive function and obesity in cognitively normal aging Chinese Han population have not attracted much attention. The aim of this study was to investigate the relationship between sex, obesity, and cognitive function in an elderly Chinese population with normal cognitive function. A total of 228 cognitively normal aging participants (males/females =93/135) entered this study. Their general demographic information (sex, age, and education) was collected by standardized questionnaire. Apolipoprotein E (APOE) genotype and serum lipid levels were measured. The Montreal Cognitive Assessment (MoCA) was used to assess participants' cognitive function. The prevalence of obesity in elderly women (18/133, 13.5%) was significantly higher than that in men (5/92, 5.4%, P =0.009). Regression analyses showed that obesity was associated with drinking alcohol (OR =13.695, P =0.045) and triglyceride (OR =1.436, P =0.048) in women and limited to low-density lipoprotein (OR =11.829, P =0.023) in men. Women performed worse on the naming score for MoCA than men ( P <0.01). Stepwise linear regression analysis showed that education ( t =3.689, P <0.001) and smoking ( t =2.031, P =0.045) were related to the score of naming in female, while high-density lipoprotein ( t =-2.077, P =0.041) was related to the score of naming in male; however, no correlation was found between body mass index and cognitive function in both male and female ( P >0.05). Our finding suggests that there are significant sex differences in obesity and specific cognitive domains in aging Chinese Han population with normal cognitive function.
Comparing The Effectiveness of a90/95 Calculations (Preprint)
2006-09-01
Nachtsheim, John Neter, William Li, Applied Linear Statistical Models , 5th ed., McGraw-Hill/Irwin, 2005 5. Mood, Graybill and Boes, Introduction...curves is based on methods that are only valid for ordinary linear regression. Requirements for a valid Ordinary Least-Squares Regression Model There... linear . For example is a linear model ; is not. 2. Uniform variance (homoscedasticity
Normalization of respiratory sinus arrhythmia by factoring in tidal volume.
Kobayashi, H
1998-09-01
The amplitude of respiratory sinus arrhythmia (RSA) was measured in eight healthy young male students with special reference to the effect of tidal volume (Vt). Under simultaneously controlled respiratory frequency and Vt, the heart rate variability (HRV) of the subjects was measured. While the respiratory frequency was adjusted to either 0.25 or 0.10 Hz, the Vt was controlled at 13 different volumes for each frequency. Linear relationships between RSA amplitude and Vt were observed and close correlations were obtained for 0.25 Hz compared with 0.10 Hz. However, regression equations showed a marked variation among subjects. Furthermore, RSA amplitude was related to vital capacity. Subjects who had lower vital capacity tended to show higher RSA amplitudes at the same Vt. Therefore, the ratio (% Vt) of Vt to vital capacity is a more effective index in normalizing RSA than raw tidal volume. From these results, we have proposed a normalized RSA (RSA amplitude/% Vt) as a new index of autonomic activity that provides a constant value regardless of Vt.
Pinheiro, Rafael S; Cruz, Ruy J; Andraus, Wellington; Ducatti, Liliana; Martino, Rodrigo B; Nacif, Lucas S; Rocha-Santos, Vinicius; Arantes, Rubens M; Lai, Quirino; Ibuki, Felicia S; Rocha, Manoel S; D Albuquerque, Luiz A C
2017-01-01
Computed tomography volumetry (CTV) is a useful tool for predicting graft weights (GW) for living donor liver transplantation (LDLT). Few studies have examined the correlation between CTV and GW in normal liver parenchyma. To analyze the correlation between CTV and GW in an adult LDLT population and provide a systematic review of the existing mathematical models to calculate partial liver graft weight. Between January 2009 and January 2013, 28 consecutive donors undergoing right hepatectomy for LDLT were retrospectively reviewed. All grafts were perfused with HTK solution. Estimated graft volume was estimated by CTV and these values were compared to the actual graft weight, which was measured after liver harvesting and perfusion. Median actual GW was 782.5 g, averaged 791.43±136 g and ranged from 520-1185 g. Median estimated graft volume was 927.5 ml, averaged 944.86±200.74 ml and ranged from 600-1477 ml. Linear regression of estimated graft volume and actual GW was significantly linear (GW=0.82 estimated graft volume, r2=0.98, slope=0.47, standard deviation of 0.024 and p<0.0001). Spearman Linear correlation was 0.65 with 95% CI of 0.45 - 0.99 (p<0.0001). The one-to-one rule did not applied in patients with normal liver parenchyma. A better estimation of graft weight could be reached by multiplying estimated graft volume by 0.82. A volumetria por tomografia computadorizada (VTC) é uma ferramenta útil para a previsão do peso do enxerto (PE) para o transplante hepático com doador vivo (TFDV). Poucos estudos examinaram a correlação entre o VTC e PE no parênquima hepático normal. Analisar a correlação entre VTC e PE em uma população adulta de doadores para o TFDV e realização de revisão sistemática dos modelos matemáticos existentes para calcular o peso de enxertos hepáticos parciais. Foram revisados retrospectivamente 28 doadores consecutivos submetidos à hepatectomia direita para o TFDV entre janeiro de 2009 a janeiro de 2013. Todos os doadores eram adultos saudáveis com VTC pré-operatório. Os enxertos foram perfundidos com solução de preservação HTK. O volume estimado foi obtido por VTC e estes valores foram comparados com o peso real do enxerto, o qual foi aferido depois da hepatectomia e perfusão do enxerto. A mediana do PE real foi de 782,5 g, média de 791,43±136 g, variando de 520-1185 g. A mediana do volume estimado do enxerto foi de 927,5 ml, média de 944,86±200,74 ml e variou de 600-1477 ml. A regressão linear volume estimado do enxerto e PE real foi significativamente linear (PE=0.82 do volume estimado do enxerto, r2=0,98, declive=0,47, desvio-padrão de 0,024 e p<0,0001). Correlação linear de Spearman foi de 0,65, com IC de 95% do 0,45-0,99 (p<0,0001). A regra de "um-para-um" não deve ser empregada em pacientes com parênquima hepático normal. A melhor estimativa do peso do enxerto hepático de doador vivo pode ser alcançado através da multiplicação do VTC por 0,82.
Correlation and simple linear regression.
Zou, Kelly H; Tuncali, Kemal; Silverman, Stuart G
2003-06-01
In this tutorial article, the concepts of correlation and regression are reviewed and demonstrated. The authors review and compare two correlation coefficients, the Pearson correlation coefficient and the Spearman rho, for measuring linear and nonlinear relationships between two continuous variables. In the case of measuring the linear relationship between a predictor and an outcome variable, simple linear regression analysis is conducted. These statistical concepts are illustrated by using a data set from published literature to assess a computed tomography-guided interventional technique. These statistical methods are important for exploring the relationships between variables and can be applied to many radiologic studies.
Misyura, Maksym; Sukhai, Mahadeo A; Kulasignam, Vathany; Zhang, Tong; Kamel-Reid, Suzanne; Stockley, Tracy L
2018-02-01
A standard approach in test evaluation is to compare results of the assay in validation to results from previously validated methods. For quantitative molecular diagnostic assays, comparison of test values is often performed using simple linear regression and the coefficient of determination (R 2 ), using R 2 as the primary metric of assay agreement. However, the use of R 2 alone does not adequately quantify constant or proportional errors required for optimal test evaluation. More extensive statistical approaches, such as Bland-Altman and expanded interpretation of linear regression methods, can be used to more thoroughly compare data from quantitative molecular assays. We present the application of Bland-Altman and linear regression statistical methods to evaluate quantitative outputs from next-generation sequencing assays (NGS). NGS-derived data sets from assay validation experiments were used to demonstrate the utility of the statistical methods. Both Bland-Altman and linear regression were able to detect the presence and magnitude of constant and proportional error in quantitative values of NGS data. Deming linear regression was used in the context of assay comparison studies, while simple linear regression was used to analyse serial dilution data. Bland-Altman statistical approach was also adapted to quantify assay accuracy, including constant and proportional errors, and precision where theoretical and empirical values were known. The complementary application of the statistical methods described in this manuscript enables more extensive evaluation of performance characteristics of quantitative molecular assays, prior to implementation in the clinical molecular laboratory. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Max dD/Dt: A Novel Parameter to Assess Fetal Cardiac Contractility and a Substitute for Max dP/Dt.
Fujita, Yasuyuki; Kiyokoba, Ryo; Yumoto, Yasuo; Kato, Kiyoko
2018-07-01
Aortic pulse waveforms are composed of a forward wave from the heart and a reflection wave from the periphery. We focused on this forward wave and suggested a new parameter, the maximum slope of aortic pulse waveforms (max dD/dt), for fetal cardiac contractility. Max dD/dt was calculated from fetal aortic pulse waveforms recorded with an echo-tracking system. A normal range of max dD/dt was constructed in 105 healthy fetuses using linear regression analysis. Twenty-two fetuses with suspected fetal cardiac dysfunction were divided into normal and decreased max dD/dt groups, and their clinical parameters were compared. Max dD/dt of aortic pulse waveforms increased linearly with advancing gestational age (r = 0.93). The decreased max dD/dt was associated with abnormal cardiotocography findings and short- and long-term prognosis. In conclusion, max dD/dt calculated from the aortic pulse waveforms in fetuses can substitute for max dP/dt, an index of cardiac contractility in adults. Copyright © 2018 World Federation for Ultrasound in Medicine and Biology. Published by Elsevier Inc. All rights reserved.
Detection of chewing from piezoelectric film sensor signals using ensemble classifiers.
Farooq, Muhammad; Sazonov, Edward
2016-08-01
Selection and use of pattern recognition algorithms is application dependent. In this work, we explored the use of several ensembles of weak classifiers to classify signals captured from a wearable sensor system to detect food intake based on chewing. Three sensor signals (Piezoelectric sensor, accelerometer, and hand to mouth gesture) were collected from 12 subjects in free-living conditions for 24 hrs. Sensor signals were divided into 10 seconds epochs and for each epoch combination of time and frequency domain features were computed. In this work, we present a comparison of three different ensemble techniques: boosting (AdaBoost), bootstrap aggregation (bagging) and stacking, each trained with 3 different weak classifiers (Decision Trees, Linear Discriminant Analysis (LDA) and Logistic Regression). Type of feature normalization used can also impact the classification results. For each ensemble method, three feature normalization techniques: (no-normalization, z-score normalization, and minmax normalization) were tested. A 12 fold cross-validation scheme was used to evaluate the performance of each model where the performance was evaluated in terms of precision, recall, and accuracy. Best results achieved here show an improvement of about 4% over our previous algorithms.
2017-10-01
ENGINEERING CENTER GRAIN EVALUATION SOFTWARE TO NUMERICALLY PREDICT LINEAR BURN REGRESSION FOR SOLID PROPELLANT GRAIN GEOMETRIES Brian...author(s) and should not be construed as an official Department of the Army position, policy, or decision, unless so designated by other documentation...U.S. ARMY ARMAMENT RESEARCH, DEVELOPMENT AND ENGINEERING CENTER GRAIN EVALUATION SOFTWARE TO NUMERICALLY PREDICT LINEAR BURN REGRESSION FOR SOLID
Linear regression in astronomy. II
NASA Technical Reports Server (NTRS)
Feigelson, Eric D.; Babu, Gutti J.
1992-01-01
A wide variety of least-squares linear regression procedures used in observational astronomy, particularly investigations of the cosmic distance scale, are presented and discussed. The classes of linear models considered are (1) unweighted regression lines, with bootstrap and jackknife resampling; (2) regression solutions when measurement error, in one or both variables, dominates the scatter; (3) methods to apply a calibration line to new data; (4) truncated regression models, which apply to flux-limited data sets; and (5) censored regression models, which apply when nondetections are present. For the calibration problem we develop two new procedures: a formula for the intercept offset between two parallel data sets, which propagates slope errors from one regression to the other; and a generalization of the Working-Hotelling confidence bands to nonstandard least-squares lines. They can provide improved error analysis for Faber-Jackson, Tully-Fisher, and similar cosmic distance scale relations.
Häggström, J; Andersson, Å O; Falk, T; Nilsfors, L; OIsson, U; Kresken, J G; Höglund, K; Rishniw, M; Tidholm, A; Ljungvall, I
2016-09-01
Echocardiography is a cost-efficient method to screen cats for presence of heart disease. Current reference intervals for feline cardiac dimensions do not account for body weight (BW). To study the effect of BW on heart rate (HR), aortic (Ao), left atrial (LA) and ventricular (LV) linear dimensions in cats, and to calculate 95% prediction intervals for these variables in normal adult pure-bred cats. 19 866 pure-bred cats. Clinical data from heart screens conducted between 1999 and 2014 were included. Associations between BW, HR, and cardiac dimensions were assessed using univariate linear models and allometric scaling, including all cats, and only those considered normal, respectively. Prediction intervals were created using 95% confidence intervals obtained from regression curves. Associations between BW and echocardiographic dimensions were best described by allometric scaling, and all dimensions increased with increasing BW (all P<0.001). Strongest associations were found between BW and Ao, LV end diastolic, LA dimensions, and thickness of LV free wall. Weak linear associations were found between BW and HR and left atrial to aortic ratio (LA:Ao), for which HR decreased with increasing BW (P<0.001), and LA:Ao increased with increasing BW (P<0.001). Marginal differences were found for prediction formulas and prediction intervals when the dataset included all cats versus only those considered normal. BW had a clinically relevant effect on echocardiographic dimensions in cats, and BW based 95% prediction intervals may help in screening cats for heart disease. Copyright © 2016 The Authors. Journal of Veterinary Internal Medicine published by Wiley Periodicals, Inc. on behalf of the American College of Veterinary Internal Medicine.
A Constrained Linear Estimator for Multiple Regression
ERIC Educational Resources Information Center
Davis-Stober, Clintin P.; Dana, Jason; Budescu, David V.
2010-01-01
"Improper linear models" (see Dawes, Am. Psychol. 34:571-582, "1979"), such as equal weighting, have garnered interest as alternatives to standard regression models. We analyze the general circumstances under which these models perform well by recasting a class of "improper" linear models as "proper" statistical models with a single predictor. We…
On the design of classifiers for crop inventories
NASA Technical Reports Server (NTRS)
Heydorn, R. P.; Takacs, H. C.
1986-01-01
Crop proportion estimators that use classifications of satellite data to correct, in an additive way, a given estimate acquired from ground observations are discussed. A linear version of these estimators is optimal, in terms of minimum variance, when the regression of the ground observations onto the satellite observations in linear. When this regression is not linear, but the reverse regression (satellite observations onto ground observations) is linear, the estimator is suboptimal but still has certain appealing variance properties. In this paper expressions are derived for those regressions which relate the intercepts and slopes to conditional classification probabilities. These expressions are then used to discuss the question of classifier designs that can lead to low-variance crop proportion estimates. Variance expressions for these estimates in terms of classifier omission and commission errors are also derived.
Correlation between quantified breast densities from digital mammography and 18F-FDG PET uptake.
Lakhani, Paras; Maidment, Andrew D A; Weinstein, Susan P; Kung, Justin W; Alavi, Abass
2009-01-01
To correlate breast density quantified from digital mammograms with mean and maximum standardized uptake values (SUVs) from positron emission tomography (PET). This was a prospective study that included 56 women with a history of suspicion of breast cancer (mean age 49.2 +/- 9.3 years), who underwent 18F-fluoro-2-deoxyglucose (FDG)-PET imaging of their breasts as well as digital mammography. A computer thresholding algorithm was applied to the contralateral nonmalignant breasts to quantitatively estimate the breast density on digital mammograms. The breasts were also classified into one of four Breast Imaging Reporting and Data System categories for density. Comparisons between SUV and breast density were made using linear regression and the Student's t-test. Linear regression of mean SUV versus average breast density showed a positive relationship with a Pearson's correlation coefficient of R(2) = 0.83. The quantified breast densities and mean SUVs were significantly greater for mammographically dense than nondense breasts (p < 0.0001 for both). The average quantified densities and mean SUVs of the breasts were significantly greater for premenopausal than postmenopausal patients (p < 0.05). 8/51 (16%) of the patients had maximum SUVs that equaled 1.6 or greater. There is a positive linear correlation between quantified breast density on digital mammography and FDG uptake on PET. Menopausal status affects the metabolic activity of normal breast tissue, resulting in higher SUVs in pre- versus postmenopausal patients.
Kokaly, R.F.; Clark, R.N.
1999-01-01
We develop a new method for estimating the biochemistry of plant material using spectroscopy. Normalized band depths calculated from the continuum-removed reflectance spectra of dried and ground leaves were used to estimate their concentrations of nitrogen, lignin, and cellulose. Stepwise multiple linear regression was used to select wavelengths in the broad absorption features centered at 1.73 ??m, 2.10 ??m, and 2.30 ??m that were highly correlated with the chemistry of samples from eastern U.S. forests. Band depths of absorption features at these wavelengths were found to also be highly correlated with the chemistry of four other sites. A subset of data from the eastern U.S. forest sites was used to derive linear equations that were applied to the remaining data to successfully estimate their nitrogen, lignin, and cellulose concentrations. Correlations were highest for nitrogen (R2 from 0.75 to 0.94). The consistent results indicate the possibility of establishing a single equation capable of estimating the chemical concentrations in a wide variety of species from the reflectance spectra of dried leaves. The extension of this method to remote sensing was investigated. The effects of leaf water content, sensor signal-to-noise and bandpass, atmospheric effects, and background soil exposure were examined. Leaf water was found to be the greatest challenge to extending this empirical method to the analysis of fresh whole leaves and complete vegetation canopies. The influence of leaf water on reflectance spectra must be removed to within 10%. Other effects were reduced by continuum removal and normalization of band depths. If the effects of leaf water can be compensated for, it might be possible to extend this method to remote sensing data acquired by imaging spectrometers to give estimates of nitrogen, lignin, and cellulose concentrations over large areas for use in ecosystem studies.We develop a new method for estimating the biochemistry of plant material using spectroscopy. Normalized band depths calculated from the continuum-removed reflectance spectra of dried and ground leaves were used to estimate their concentrations of nitrogen, lignin, and cellulose. Stepwise multiple linear regression was used to select wavelengths in the broad absorption features centered at 1.73 ??m, 2.10 ??m, and 2.301 ??m that were highly correlated with the chemistry of samples from eastern U.S. forests. Band depths of absorption features at these wavelengths were found to also be highly correlated with the chemistry of four other sites. A subset of data from the eastern U.S. forest sites was used to derive linear equations that were applied to the remaining data to successfully estimate their nitrogen, lignin, and cellulose concentrations. Correlations were highest for nitrogen (R2 from 0.75 to 0.94). The consistent results indicate the possibility of establishing a single equation capable of estimating the chemical concentrations in a wide variety of species from the reflectance spectra of dried leaves. The extension of this method to remote sensing was investigated. The effects of leaf water content, sensor signal-to-noise and bandpass, atmospheric effects, and background soil exposure were examined. Leaf water was found to be the greatest challenge to extending this empirical method to the analysis of fresh whole leaves and complete vegetation canopies. The influence of leaf water on reflectance spectra must be removed to within 10%. Other effects were reduced by continuum removal and normalization of band depths. If the effects of leaf water can be compensated for, it might be possible to extend this method to remote sensing data acquired by imaging spectrometers to give estimates of nitrogen, lignin, and cellulose concentrations over large areas for use in ecosystem studies.
Speech prosody impairment predicts cognitive decline in Parkinson's disease.
Rektorova, Irena; Mekyska, Jiri; Janousova, Eva; Kostalova, Milena; Eliasova, Ilona; Mrackova, Martina; Berankova, Dagmar; Necasova, Tereza; Smekal, Zdenek; Marecek, Radek
2016-08-01
Impairment of speech prosody is characteristic for Parkinson's disease (PD) and does not respond well to dopaminergic treatment. We assessed whether baseline acoustic parameters, alone or in combination with other predominantly non-dopaminergic symptoms may predict global cognitive decline as measured by the Addenbrooke's cognitive examination (ACE-R) and/or worsening of cognitive status as assessed by a detailed neuropsychological examination. Forty-four consecutive non-depressed PD patients underwent clinical and cognitive testing, and acoustic voice analysis at baseline and at the two-year follow-up. Influence of speech and other clinical parameters on worsening of the ACE-R and of the cognitive status was analyzed using linear and logistic regression. The cognitive status (classified as normal cognition, mild cognitive impairment and dementia) deteriorated in 25% of patients during the follow-up. The multivariate linear regression model consisted of the variation in range of the fundamental voice frequency (F0VR) and the REM Sleep Behavioral Disorder Screening Questionnaire (RBDSQ). These parameters explained 37.2% of the variability of the change in ACE-R. The most significant predictors in the univariate logistic regression were the speech index of rhythmicity (SPIR; p = 0.012), disease duration (p = 0.019), and the RBDSQ (p = 0.032). The multivariate regression analysis revealed that SPIR alone led to 73.2% accuracy in predicting a change in cognitive status. Combining SPIR with RBDSQ improved the prediction accuracy of SPIR alone by 7.3%. Impairment of speech prosody together with symptoms of RBD predicted rapid cognitive decline and worsening of PD cognitive status during a two-year period. Copyright © 2016 Elsevier Ltd. All rights reserved.
Regression analysis of sparse asynchronous longitudinal data.
Cao, Hongyuan; Zeng, Donglin; Fine, Jason P
2015-09-01
We consider estimation of regression models for sparse asynchronous longitudinal observations, where time-dependent responses and covariates are observed intermittently within subjects. Unlike with synchronous data, where the response and covariates are observed at the same time point, with asynchronous data, the observation times are mismatched. Simple kernel-weighted estimating equations are proposed for generalized linear models with either time invariant or time-dependent coefficients under smoothness assumptions for the covariate processes which are similar to those for synchronous data. For models with either time invariant or time-dependent coefficients, the estimators are consistent and asymptotically normal but converge at slower rates than those achieved with synchronous data. Simulation studies evidence that the methods perform well with realistic sample sizes and may be superior to a naive application of methods for synchronous data based on an ad hoc last value carried forward approach. The practical utility of the methods is illustrated on data from a study on human immunodeficiency virus.
NASA Astrophysics Data System (ADS)
Ayyoub, Abdellatif; Er-Raki, Salah; Khabba, Saïd; Merlin, Olivier; César Rodriguez, Julio; Ezzahar, Jamal; Bahlaoui, Ahmed; Chehbouni, Abdelghani
2016-04-01
The present work aims to develop a simple approach relating normalized daily sap flow (per unit of leaf area) and daily ET0 (mm/day) calculated by two methods: FAO-Penman-Monteith (FAO-PM) and Hargreaves-Samani (HARG). The data sets used for developing this approach are taken from three experimental sites (olive trees, cv. "Oleaeuropaea L.", olive trees, cv. "Arbequino" and citrus trees cv. "Clementine Afourar") conducted in the Tensift region around Marrakech, Morocco and one experimental site (pecan orchard, cv. "Caryaillinoinensis, Wangenh. K. Koch") conducted in the Yaqui Valley, northwest of Mexico). The results showed that the normalized daily sap flow (volume of transpired water per unit of leaf area) was linearly correlated with ET0 (mm per day) calculated by FAO-PM method. The coefficient of determination (R2) and the slope of this linear regression varied between 0.71 and 0.97 and between 0.30 and 0.35, respectively, depending on the type of orchards. For HARG method, the relationship between both terms is also linear but with less accuracy (R2 =0.7) as expected due to the underestimation of ET0 by this method. Afterward, the validation of the developed linear relationship was performed over an olive orchard ("Oleaeuropaea L.") where the measurements of sap flow were available for another (2004) cropping season. The scatter plot between the normalized measured and estimated sap flow based on FAO-PM method reveals a very good agreement (slope = 1, with R2 = 0.83 and RMSE=0.14 L/m2 leaf area). However, for the estimation of normalized sap flow based on HARG method, the correlation is more scattered with some underestimation (5%). A further validation wasperformed using the measurements of evapotranspiration (ET) by eddy correlation system and the results showed that the correlation between normalized measured ET and estimated normalized sap flow is best when using FAO-PM method (RMSE=0.33 L/m2 leaf area) for estimating ET0 than when using HARG method (RMSE= 0.51 L/m2 leaf area). Finally, the performance of the developed approach was compared to the traditional dual crop coefficient scheme for estimating plant transpiration. Cross-comparison of these two approaches with the measurements data gave satisfactory results with an average value of RMSE equal to about 0.37 mm/day for both approaches.
Sample size determination for logistic regression on a logit-normal distribution.
Kim, Seongho; Heath, Elisabeth; Heilbrun, Lance
2017-06-01
Although the sample size for simple logistic regression can be readily determined using currently available methods, the sample size calculation for multiple logistic regression requires some additional information, such as the coefficient of determination ([Formula: see text]) of a covariate of interest with other covariates, which is often unavailable in practice. The response variable of logistic regression follows a logit-normal distribution which can be generated from a logistic transformation of a normal distribution. Using this property of logistic regression, we propose new methods of determining the sample size for simple and multiple logistic regressions using a normal transformation of outcome measures. Simulation studies and a motivating example show several advantages of the proposed methods over the existing methods: (i) no need for [Formula: see text] for multiple logistic regression, (ii) available interim or group-sequential designs, and (iii) much smaller required sample size.
Bowen, Stephen R; Chappell, Richard J; Bentzen, Søren M; Deveau, Michael A; Forrest, Lisa J; Jeraj, Robert
2012-01-01
Purpose To quantify associations between pre-radiotherapy and post-radiotherapy PET parameters via spatially resolved regression. Materials and methods Ten canine sinonasal cancer patients underwent PET/CT scans of [18F]FDG (FDGpre), [18F]FLT (FLTpre), and [61Cu]Cu-ATSM (Cu-ATSMpre). Following radiotherapy regimens of 50 Gy in 10 fractions, veterinary patients underwent FDG PET/CT scans at three months (FDGpost). Regression of standardized uptake values in baseline FDGpre, FLTpre and Cu-ATSMpre tumour voxels to those in FDGpost images was performed for linear, log-linear, generalized-linear and mixed-fit linear models. Goodness-of-fit in regression coefficients was assessed by R2. Hypothesis testing of coefficients over the patient population was performed. Results Multivariate linear model fits of FDGpre to FDGpost were significantly positive over the population (FDGpost~0.17 FDGpre, p=0.03), and classified slopes of RECIST non-responders and responders to be different (0.37 vs. 0.07, p=0.01). Generalized-linear model fits related FDGpre to FDGpost by a linear power law (FDGpost~FDGpre0.93, p<0.001). Univariate mixture model fits of FDGpre improved R2 from 0.17 to 0.52. Neither baseline FLT PET nor Cu-ATSM PET uptake contributed statistically significant multivariate regression coefficients. Conclusions Spatially resolved regression analysis indicates that pre-treatment FDG PET uptake is most strongly associated with three-month post-treatment FDG PET uptake in this patient population, though associations are histopathology-dependent. PMID:22682748
An Analysis of COLA (Cost of Living Adjustment) Allocation within the United States Coast Guard.
1983-09-01
books Applied Linear Regression [Ref. 39], and Statistical Methods in Research and Production [Ref. 40], or any other book on regression. In the event...Indexes, Master’s Thesis, Air Force Institute of Technology, Wright-Patterson AFB, 1976. 39. Weisberg, Stanford, Applied Linear Regression , Wiley, 1980. 40
Testing hypotheses for differences between linear regression lines
Stanley J. Zarnoch
2009-01-01
Five hypotheses are identified for testing differences between simple linear regression lines. The distinctions between these hypotheses are based on a priori assumptions and illustrated with full and reduced models. The contrast approach is presented as an easy and complete method for testing for overall differences between the regressions and for making pairwise...
Graphical Description of Johnson-Neyman Outcomes for Linear and Quadratic Regression Surfaces.
ERIC Educational Resources Information Center
Schafer, William D.; Wang, Yuh-Yin
A modification of the usual graphical representation of heterogeneous regressions is described that can aid in interpreting significant regions for linear or quadratic surfaces. The standard Johnson-Neyman graph is a bivariate plot with the criterion variable on the ordinate and the predictor variable on the abscissa. Regression surfaces are drawn…
Teaching the Concept of Breakdown Point in Simple Linear Regression.
ERIC Educational Resources Information Center
Chan, Wai-Sum
2001-01-01
Most introductory textbooks on simple linear regression analysis mention the fact that extreme data points have a great influence on ordinary least-squares regression estimation; however, not many textbooks provide a rigorous mathematical explanation of this phenomenon. Suggests a way to fill this gap by teaching students the concept of breakdown…
Estimating monotonic rates from biological data using local linear regression.
Olito, Colin; White, Craig R; Marshall, Dustin J; Barneche, Diego R
2017-03-01
Accessing many fundamental questions in biology begins with empirical estimation of simple monotonic rates of underlying biological processes. Across a variety of disciplines, ranging from physiology to biogeochemistry, these rates are routinely estimated from non-linear and noisy time series data using linear regression and ad hoc manual truncation of non-linearities. Here, we introduce the R package LoLinR, a flexible toolkit to implement local linear regression techniques to objectively and reproducibly estimate monotonic biological rates from non-linear time series data, and demonstrate possible applications using metabolic rate data. LoLinR provides methods to easily and reliably estimate monotonic rates from time series data in a way that is statistically robust, facilitates reproducible research and is applicable to a wide variety of research disciplines in the biological sciences. © 2017. Published by The Company of Biologists Ltd.
Determination of stress intensity factors for interface cracks under mixed-mode loading
NASA Technical Reports Server (NTRS)
Naik, Rajiv A.; Crews, John H., Jr.
1992-01-01
A simple technique was developed using conventional finite element analysis to determine stress intensity factors, K1 and K2, for interface cracks under mixed-mode loading. This technique involves the calculation of crack tip stresses using non-singular finite elements. These stresses are then combined and used in a linear regression procedure to calculate K1 and K2. The technique was demonstrated by calculating three different bimaterial combinations. For the normal loading case, the K's were within 2.6 percent of an exact solution. The normalized K's under shear loading were shown to be related to the normalized K's under normal loading. Based on these relations, a simple equation was derived for calculating K1 and K2 for mixed-mode loading from knowledge of the K's under normal loading. The equation was verified by computing the K's for a mixed-mode case with equal and normal shear loading. The correlation between exact and finite element solutions is within 3.7 percent. This study provides a simple procedure to compute K2/K1 ratio which has been used to characterize the stress state at the crack tip for various combinations of materials and loadings. Tests conducted over a range of K2/K1 ratios could be used to fully characterize interface fracture toughness.
Phan, Xuan; Grisbrook, Tiffany L; Wernli, Kevin; Stearne, Sarah M; Davey, Paul; Ng, Leo
2017-08-01
This study aimed to determine if a quantifiable relationship exists between the peak sound amplitude and peak vertical ground reaction force (vGRF) and vertical loading rate during running. It also investigated whether differences in peak sound amplitude, contact time, lower limb kinematics, kinetics and foot strike technique existed when participants were verbally instructed to run quietly compared to their normal running. A total of 26 males completed running trials for two sound conditions: normal running and quiet running. Simple linear regressions revealed no significant relationships between impact sound and peak vGRF in the normal and quiet conditions and vertical loading rate in the normal condition. t-Tests revealed significant within-subject decreases in peak sound, peak vGRF and vertical loading rate during the quiet compared to the normal running condition. During the normal running condition, 15.4% of participants utilised a non-rearfoot strike technique compared to 76.9% in the quiet condition, which was corroborated by an increased ankle plantarflexion angle at initial contact. This study demonstrated that quieter impact sound is not directly associated with a lower peak vGRF or vertical loading rate. However, given the instructions to run quietly, participants effectively reduced peak impact sound, peak vGRF and vertical loading rate.
Locally linear regression for pose-invariant face recognition.
Chai, Xiujuan; Shan, Shiguang; Chen, Xilin; Gao, Wen
2007-07-01
The variation of facial appearance due to the viewpoint (/pose) degrades face recognition systems considerably, which is one of the bottlenecks in face recognition. One of the possible solutions is generating virtual frontal view from any given nonfrontal view to obtain a virtual gallery/probe face. Following this idea, this paper proposes a simple, but efficient, novel locally linear regression (LLR) method, which generates the virtual frontal view from a given nonfrontal face image. We first justify the basic assumption of the paper that there exists an approximate linear mapping between a nonfrontal face image and its frontal counterpart. Then, by formulating the estimation of the linear mapping as a prediction problem, we present the regression-based solution, i.e., globally linear regression. To improve the prediction accuracy in the case of coarse alignment, LLR is further proposed. In LLR, we first perform dense sampling in the nonfrontal face image to obtain many overlapped local patches. Then, the linear regression technique is applied to each small patch for the prediction of its virtual frontal patch. Through the combination of all these patches, the virtual frontal view is generated. The experimental results on the CMU PIE database show distinct advantage of the proposed method over Eigen light-field method.
Wheaton, Anne G; Perry, Geraldine S; Chapman, Daniel P; McKnight-Eily, Lela R; Presley-Cantrell, Letitia R; Croft, Janet B
2011-05-10
Over the past 50 years, the average sleep duration for adults in the United States has decreased while the prevalence of obesity and associated outcomes has increased. The objective of this study was to determine whether perceived insufficient sleep was associated with body mass index (BMI) in a national sample. We analyzed data from the 2008 Behavioral Risk Factor Surveillance System (BRFSS) survey (N=384,541) in which respondents were asked, "During the past 30 days, for about how many days have you felt you did not get enough rest or sleep?" We divided respondents into six BMI categories and used multivariable linear regression and logistic regression analyses to assess the association between BMI categories and days of insufficient sleep after adjusting for sociodemographic variables, smoking, physical activity, and frequent mental distress. Adjusted mean days of insufficient sleep ranged from 7.9 (95% confidence interval [CI]: 7.8, 8.0) days for people of normal weight to 10.5 (95% CI: 10.2, 10.9) days for those in the highest weight category (BMI≥40). Days of perceived insufficient sleep followed a linear trend across BMI categories. The likelihood of reporting ≥14 days of insufficient sleep in the previous 30 days was higher for respondents in the highest weight category than for those who were normal weight (34.9% vs. 25.2%; adjusted odds ratio=1.7 (95% CI: 1.5, 1.8]). Among U.S. adults, days of insufficient rest or sleep strongly correlated with BMI. Sleep sufficiency should be an important consideration in the assessment of the health of overweight and obese people and should be considered by developers of weight-reduction programs.
Prevalence of and risk factors for reduced serum bicarbonate in chronic kidney disease.
Raphael, Kalani L; Zhang, Yingying; Ying, Jian; Greene, Tom
2014-10-01
The prevalence of metabolic acidosis increases as glomerular filtration rate falls. However, most patients with stage 4 chronic kidney disease have normal serum bicarbonate concentration while some with stage 3 chronic kidney disease have low serum bicarbonate, suggesting that other factors contribute to generation of acidosis. The purpose of this study is to identify risk factors, other than reduced glomerular filtration rate, for reduced serum bicarbonate in chronic kidney disease. This is a cross-sectional analysis of baseline data from the Chronic Renal Insufficiency Cohort Study. Multivariable logistic and linear regression models were used to relate predictor variables to the odds of low serum bicarbonate (< 22 mM) compared with normal serum bicarbonate (22-30 mM) and the coefficients of Δ serum bicarbonate concentration. The prevalence of low serum bicarbonate at baseline was 17.3%. Lower estimated glomerular filtration rate had the strongest relationship with low serum bicarbonate. Factors associated with higher odds of low serum bicarbonate, independent of estimated glomerular filtration rate, were urinary albumin/creatinine ≥ 10 mg/g, smoking, anaemia, hyperkalaemia, non-diuretic use and higher serum albumin. These and younger age, higher waist circumference, and use of angiotensin converting enzyme inhibitors or angiotensin receptor blockers associated with negative Δ serum bicarbonate in linear regression models. Several factors not typically considered to associate with reduced serum bicarbonate in chronic kidney disease were identified including albuminuria ≥ 10 mg/g, anaemia, smoking, higher serum albumin, higher waist circumference, and use of angiotensin converting enzyme inhibitors or angiotensin receptor blockers. Future studies should explore the longitudinal effect of these factors on serum bicarbonate concentration. © 2014 Asian Pacific Society of Nephrology.
Zhu, Hang; Xue, Hao; Wang, Guangyi; Fu, Zhenhong; Liu, Jie; Shi, Yajun
2015-04-01
To explore the association between urinary microalbumin-to-creatinine ratio (ACR) and brachial-ankle pulse wave velocity (baPWV) in hypertensive patients. A total of 877 primary hypertension patients were enrolled in this trial from September 2009 to December 2012, and were randomly recruited and patients were divided into normal ACR group (ACR < 30 mg/g, n = 723), micro-albuminuria group (30 mg/g ≤ ACR < 300 mg/g, n = 136) and macro-albuminuria group (ACR ≥ 300 mg/g, n = 18). baPWV was measure by automatic pulse wave velocity measuring system. The baPWV values in patients of micro-albuminuria group and macro-albuminuria group were significantly higher than in the normal ACR group (all P < 0.05). The baPWV value of macro-albuminuria group was significantly higher than in the micro-albuminuria group (P < 0.05). Linear correlation analysis revealed that ACR was positively correlated with baPWV (r = 0.413, P < 0.01). Multiple linear regression analysis showed that ACR independently correlated with baPWV in patients with primary hypertension (β = 0.29, R(2) = 0.112, P < 0.01) after adjusting for age, sex, body mass index, systolic blood pressure, diastolic blood pressure, blood glucose, total cholesterol, low density lipoprotein, high density lipoprotein and triglyceride. Using ACR < 30 mg/g and ACR ≥ 30 mg/g as dichotomous variable, binary logistic regression analysis showed that ACR ≥ 30 mg/g was also a risk factor of the ascending baPWV in primary hypertension patients (OR: 1.73, 95% CI: 1.62-2.98) after adjusting the traditional cardiovascular risk factors. ACR is positively correlated to baPWV in primary hypertension patients, and the ascending baPWV is a risk factor of early renal dysfunction in primary hypertension patients.
Chiang, H; Chang, K-C; Kan, H-W; Wu, S-W; Tseng, M-T; Hsueh, H-W; Lin, Y-H; Chao, C-C; Hsieh, S-T
2018-07-01
The study aimed to investigate the physiology, psychophysics, pathology and their relationship in reversible nociceptive nerve degeneration, and the physiology of acute hyperalgesia. We enrolled 15 normal subjects to investigate intraepidermal nerve fibre (IENF) density, contact heat-evoked potential (CHEP) and thermal thresholds during the capsaicin-induced skin nerve degeneration-regeneration; and CHEP and thermal thresholds at capsaicin-induced acute hyperalgesia. After 2-week capsaicin treatment, IENF density of skin was markedly reduced with reduced amplitude and prolonged latency of CHEP, and increased warm and heat pain thresholds. The time courses of skin nerve regeneration and reversal of physiology and psychophysics were different: IENF density was still lower at 10 weeks after capsaicin treatment than that at baseline, whereas CHEP amplitude and warm threshold became normalized within 3 weeks after capsaicin treatment. Although CHEP amplitude and IENF density were best correlated in a multiple linear regression model, a one-phase exponential association model showed better fit than a simple linear one, that is in the regeneration phase, the slope of the regression line between CHEP amplitude and IENF density was steeper in the subgroup with lower IENF densities than in the one with higher IENF densities. During capsaicin-induced hyperalgesia, recordable rate of CHEP to 43 °C heat stimulation was higher with enhanced CHEP amplitude and pain perception compared to baseline. There were differential restoration of IENF density, CHEP and thermal thresholds, and changed CHEP-IENF relationships during skin reinnervation. CHEP can be a physiological signature of acute hyperalgesia. These observations suggested the relationship between nociceptive nerve terminals and brain responses to thermal stimuli changed during different degree of skin denervation, and CHEP to low-intensity heat stimulus can reflect the physiology of hyperalgesia. © 2018 European Pain Federation - EFIC®.
Motulsky, Harvey J; Brown, Ronald E
2006-01-01
Background Nonlinear regression, like linear regression, assumes that the scatter of data around the ideal curve follows a Gaussian or normal distribution. This assumption leads to the familiar goal of regression: to minimize the sum of the squares of the vertical or Y-value distances between the points and the curve. Outliers can dominate the sum-of-the-squares calculation, and lead to misleading results. However, we know of no practical method for routinely identifying outliers when fitting curves with nonlinear regression. Results We describe a new method for identifying outliers when fitting data with nonlinear regression. We first fit the data using a robust form of nonlinear regression, based on the assumption that scatter follows a Lorentzian distribution. We devised a new adaptive method that gradually becomes more robust as the method proceeds. To define outliers, we adapted the false discovery rate approach to handling multiple comparisons. We then remove the outliers, and analyze the data using ordinary least-squares regression. Because the method combines robust regression and outlier removal, we call it the ROUT method. When analyzing simulated data, where all scatter is Gaussian, our method detects (falsely) one or more outlier in only about 1–3% of experiments. When analyzing data contaminated with one or several outliers, the ROUT method performs well at outlier identification, with an average False Discovery Rate less than 1%. Conclusion Our method, which combines a new method of robust nonlinear regression with a new method of outlier identification, identifies outliers from nonlinear curve fits with reasonable power and few false positives. PMID:16526949
Applying Occam's Razor To The Proton Radius Puzzle
NASA Astrophysics Data System (ADS)
Higinbotham, Douglas
2016-09-01
Over the past five decades, ever more complex mathematical functions have been used to extract the radius of the proton from electron scattering data. For example, in 1963 the proton radius was extracted with linear and quadratic fits of low Q2 data (< 3 fm-2) and by 2014 a non-linear regression of two tenth order power series functions with thirty-one normalization parameters and data out to 25 fm-2 was used. But for electron scattering, the radius of the proton is determined by extracting the slope of the charge form factor at a Q2 of zero. By using higher precision data than was available in 1963 and focusing on the low Q2 data from 1974 to today, we find extrapolating functions consistently produce a proton radius of around 0.84 fm. A result that is in agreement with modern Lamb shift measurements.
Efficient robust doubly adaptive regularized regression with applications.
Karunamuni, Rohana J; Kong, Linglong; Tu, Wei
2018-01-01
We consider the problem of estimation and variable selection for general linear regression models. Regularized regression procedures have been widely used for variable selection, but most existing methods perform poorly in the presence of outliers. We construct a new penalized procedure that simultaneously attains full efficiency and maximum robustness. Furthermore, the proposed procedure satisfies the oracle properties. The new procedure is designed to achieve sparse and robust solutions by imposing adaptive weights on both the decision loss and the penalty function. The proposed method of estimation and variable selection attains full efficiency when the model is correct and, at the same time, achieves maximum robustness when outliers are present. We examine the robustness properties using the finite-sample breakdown point and an influence function. We show that the proposed estimator attains the maximum breakdown point. Furthermore, there is no loss in efficiency when there are no outliers or the error distribution is normal. For practical implementation of the proposed method, we present a computational algorithm. We examine the finite-sample and robustness properties using Monte Carlo studies. Two datasets are also analyzed.
On the use of log-transformation vs. nonlinear regression for analyzing biological power laws
Xiao, X.; White, E.P.; Hooten, M.B.; Durham, S.L.
2011-01-01
Power-law relationships are among the most well-studied functional relationships in biology. Recently the common practice of fitting power laws using linear regression (LR) on log-transformed data has been criticized, calling into question the conclusions of hundreds of studies. It has been suggested that nonlinear regression (NLR) is preferable, but no rigorous comparison of these two methods has been conducted. Using Monte Carlo simulations, we demonstrate that the error distribution determines which method performs better, with NLR better characterizing data with additive, homoscedastic, normal error and LR better characterizing data with multiplicative, heteroscedastic, lognormal error. Analysis of 471 biological power laws shows that both forms of error occur in nature. While previous analyses based on log-transformation appear to be generally valid, future analyses should choose methods based on a combination of biological plausibility and analysis of the error distribution. We provide detailed guidelines and associated computer code for doing so, including a model averaging approach for cases where the error structure is uncertain. ?? 2011 by the Ecological Society of America.
Effect of Malmquist bias on correlation studies with IRAS data base
NASA Technical Reports Server (NTRS)
Verter, Frances
1993-01-01
The relationships between galaxy properties in the sample of Trinchieri et al. (1989) are reexamined with corrections for Malmquist bias. The linear correlations are tested and linear regressions are fit for log-log plots of L(FIR), L(H-alpha), and L(B) as well as ratios of these quantities. The linear correlations for Malmquist bias are corrected using the method of Verter (1988), in which each galaxy observation is weighted by the inverse of its sampling volume. The linear regressions are corrected for Malmquist bias by a new method invented here in which each galaxy observation is weighted by its sampling volume. The results of correlation and regressions among the sample are significantly changed in the anticipated sense that the corrected correlation confidences are lower and the corrected slopes of the linear regressions are lower. The elimination of Malmquist bias eliminates the nonlinear rise in luminosity that has caused some authors to hypothesize additional components of FIR emission.
A primer for biomedical scientists on how to execute model II linear regression analysis.
Ludbrook, John
2012-04-01
1. There are two very different ways of executing linear regression analysis. One is Model I, when the x-values are fixed by the experimenter. The other is Model II, in which the x-values are free to vary and are subject to error. 2. I have received numerous complaints from biomedical scientists that they have great difficulty in executing Model II linear regression analysis. This may explain the results of a Google Scholar search, which showed that the authors of articles in journals of physiology, pharmacology and biochemistry rarely use Model II regression analysis. 3. I repeat my previous arguments in favour of using least products linear regression analysis for Model II regressions. I review three methods for executing ordinary least products (OLP) and weighted least products (WLP) regression analysis: (i) scientific calculator and/or computer spreadsheet; (ii) specific purpose computer programs; and (iii) general purpose computer programs. 4. Using a scientific calculator and/or computer spreadsheet, it is easy to obtain correct values for OLP slope and intercept, but the corresponding 95% confidence intervals (CI) are inaccurate. 5. Using specific purpose computer programs, the freeware computer program smatr gives the correct OLP regression coefficients and obtains 95% CI by bootstrapping. In addition, smatr can be used to compare the slopes of OLP lines. 6. When using general purpose computer programs, I recommend the commercial programs systat and Statistica for those who regularly undertake linear regression analysis and I give step-by-step instructions in the Supplementary Information as to how to use loss functions. © 2011 The Author. Clinical and Experimental Pharmacology and Physiology. © 2011 Blackwell Publishing Asia Pty Ltd.
Determining the response of sea level to atmospheric pressure forcing using TOPEX/POSEIDON data
NASA Technical Reports Server (NTRS)
Fu, Lee-Lueng; Pihos, Greg
1994-01-01
The static response of sea level to the forcing of atmospheric pressure, the so-called inverted barometer (IB) effect, is investigated using TOPEX/POSEIDON data. This response, characterized by the rise and fall of sea level to compensate for the change of atmospheric pressure at a rate of -1 cm/mbar, is not associated with any ocean currents and hence is normally treated as an error to be removed from sea level observation. Linear regression and spectral transfer function analyses are applied to sea level and pressure to examine the validity of the IB effect. In regions outside the tropics, the regression coefficient is found to be consistently close to the theoretical value except for the regions of western boundary currents, where the mesoscale variability interferes with the IB effect. The spectral transfer function shows near IB response at periods of 30 degrees is -0.84 +/- 0.29 cm/mbar (1 standard deviation). The deviation from = 1 cm /mbar is shown to be caused primarily by the effect of wind forcing on sea level, based on multivariate linear regression model involving both pressure and wind forcing. The regression coefficient for pressure resulting from the multivariate analysis is -0.96 +/- 0.32 cm/mbar. In the tropics the multivariate analysis fails because sea level in the tropics is primarily responding to remote wind forcing. However, after removing from the data the wind-forced sea level estimated by a dynamic model of the tropical Pacific, the pressure regression coefficient improves from -1.22 +/- 0.69 cm/mbar to -0.99 +/- 0.46 cm/mbar, clearly revealing an IB response. The result of the study suggests that with a proper removal of the effect of wind forcing the IB effect is valid in most of the open ocean at periods longer than 20 days and spatial scales larger than 500 km.
ERIC Educational Resources Information Center
Rocconi, Louis M.
2013-01-01
This study examined the differing conclusions one may come to depending upon the type of analysis chosen, hierarchical linear modeling or ordinary least squares (OLS) regression. To illustrate this point, this study examined the influences of seniors' self-reported critical thinking abilities three ways: (1) an OLS regression with the student…
ERIC Educational Resources Information Center
Rocconi, Louis M.
2011-01-01
Hierarchical linear models (HLM) solve the problems associated with the unit of analysis problem such as misestimated standard errors, heterogeneity of regression and aggregation bias by modeling all levels of interest simultaneously. Hierarchical linear modeling resolves the problem of misestimated standard errors by incorporating a unique random…
ERIC Educational Resources Information Center
Preacher, Kristopher J.; Curran, Patrick J.; Bauer, Daniel J.
2006-01-01
Simple slopes, regions of significance, and confidence bands are commonly used to evaluate interactions in multiple linear regression (MLR) models, and the use of these techniques has recently been extended to multilevel or hierarchical linear modeling (HLM) and latent curve analysis (LCA). However, conducting these tests and plotting the…
Liu, Peter Y; Takahashi, Paul Y; Roebuck, Pamela D; Iranmanesh, Ali; Veldhuis, Johannes D
2005-09-01
Pulsatile and thus total testosterone (Te) secretion declines in older men, albeit for unknown reasons. Analytical models forecast that aging may reduce the capability of endogenous luteinizing hormone (LH) pulses to stimulate Leydig cell steroidogenesis. This notion has been difficult to test experimentally. The present study used graded doses of a selective gonadotropin releasing hormone (GnRH)-receptor antagonist to yield four distinct strata of pulsatile LH release in each of 18 healthy men ages 23-72 yr. Deconvolution analysis was applied to frequently sampled LH and Te concentration time series to quantitate pulsatile Te secretion over a 16-h interval. Log-linear regression was used to relate pulsatile LH secretion to attendant pulsatile Te secretion (LH-Te drive) across the four stepwise interventions in each subject. Linear regression of the 18 individual estimates of LH-Te feedforward dose-response slopes on age disclosed a strongly negative relationship (r = -0.721, P < 0.001). Accordingly, the present data support the thesis that aging in healthy men attenuates amplitude-dependent LH drive of burst-like Te secretion. The experimental strategy of graded suppression of neuroglandular outflow may have utility in estimating dose-response adaptations in other endocrine systems.
NASA Astrophysics Data System (ADS)
Alvarez, César I.; Teodoro, Ana; Tierra, Alfonso
2017-10-01
Thin clouds in the optical remote sensing data are frequent and in most of the cases don't allow to have a pure surface data in order to calculate some indexes as Normalized Difference Vegetation Index (NDVI). This paper aims to evaluate the Automatic Cloud Removal Method (ACRM) algorithm over a high elevation city like Quito (Ecuador), with an altitude of 2800 meters above sea level, where the clouds are presented all the year. The ACRM is an algorithm that considers a linear regression between each Landsat 8 OLI band and the Cirrus band using the slope obtained with the linear regression established. This algorithm was employed without any reference image or mask to try to remove the clouds. The results of the application of the ACRM algorithm over Quito didn't show a good performance. Therefore, was considered improving this algorithm using a different slope value data (ACMR Improved). After, the NDVI computation was compared with a reference NDVI MODIS data (MOD13Q1). The ACMR Improved algorithm had a successful result when compared with the original ACRM algorithm. In the future, this Improved ACRM algorithm needs to be tested in different regions of the world with different conditions to evaluate if the algorithm works successfully for all conditions.
Analyses of Field Test Data at the Atucha-1 Spent Fuel Pools
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sitaraman, S.
A field test was conducted at the Atucha-1 spent nuclear fuel pools to validate a software package for gross defect detection that is used in conjunction with the inspection tool, Spent Fuel Neutron Counter (SFNC). A set of measurements was taken with the SFNC and the software predictions were compared with these data and analyzed. The data spanned a wide range of cooling times and a set of burnup levels leading to count rates from the several hundreds to around twenty per second. The current calibration in the software using linear fitting required the use of multiple calibration factors tomore » cover the entire range of count rates recorded. The solution to this was to use power regression data fitting to normalize the predicted response and derive one calibration factor that can be applied to the entire set of data. The resulting comparisons between the predicted and measured responses were generally good and provided a quantitative method of detecting missing fuel in virtually all situations. Since the current version of the software uses the linear calibration method, it would need to be updated with the new power regression method to make it more user-friendly for real time verification and fieldable for the range of responses that will be encountered.« less
Agreement evaluation of AVHRR and MODIS 16-day composite NDVI data sets
Ji, Lei; Gallo, Kevin P.; Eidenshink, Jeffery C.; Dwyer, John L.
2008-01-01
Satellite-derived normalized difference vegetation index (NDVI) data have been used extensively to detect and monitor vegetation conditions at regional and global levels. A combination of NDVI data sets derived from AVHRR and MODIS can be used to construct a long NDVI time series that may also be extended to VIIRS. Comparative analysis of NDVI data derived from AVHRR and MODIS is critical to understanding the data continuity through the time series. In this study, the AVHRR and MODIS 16-day composite NDVI products were compared using regression and agreement analysis methods. The analysis shows a high agreement between the AVHRR-NDVI and MODIS-NDVI observed from 2002 and 2003 for the conterminous United States, but the difference between the two data sets is appreciable. Twenty per cent of the total difference between the two data sets is due to systematic difference, with the remainder due to unsystematic difference. The systematic difference can be eliminated with a linear regression-based transformation between two data sets, and the unsystematic difference can be reduced partially by applying spatial filters to the data. We conclude that the continuity of NDVI time series from AVHRR to MODIS is satisfactory, but a linear transformation between the two sets is recommended.
Classical Testing in Functional Linear Models.
Kong, Dehan; Staicu, Ana-Maria; Maity, Arnab
2016-01-01
We extend four tests common in classical regression - Wald, score, likelihood ratio and F tests - to functional linear regression, for testing the null hypothesis, that there is no association between a scalar response and a functional covariate. Using functional principal component analysis, we re-express the functional linear model as a standard linear model, where the effect of the functional covariate can be approximated by a finite linear combination of the functional principal component scores. In this setting, we consider application of the four traditional tests. The proposed testing procedures are investigated theoretically for densely observed functional covariates when the number of principal components diverges. Using the theoretical distribution of the tests under the alternative hypothesis, we develop a procedure for sample size calculation in the context of functional linear regression. The four tests are further compared numerically for both densely and sparsely observed noisy functional data in simulation experiments and using two real data applications.
Classical Testing in Functional Linear Models
Kong, Dehan; Staicu, Ana-Maria; Maity, Arnab
2016-01-01
We extend four tests common in classical regression - Wald, score, likelihood ratio and F tests - to functional linear regression, for testing the null hypothesis, that there is no association between a scalar response and a functional covariate. Using functional principal component analysis, we re-express the functional linear model as a standard linear model, where the effect of the functional covariate can be approximated by a finite linear combination of the functional principal component scores. In this setting, we consider application of the four traditional tests. The proposed testing procedures are investigated theoretically for densely observed functional covariates when the number of principal components diverges. Using the theoretical distribution of the tests under the alternative hypothesis, we develop a procedure for sample size calculation in the context of functional linear regression. The four tests are further compared numerically for both densely and sparsely observed noisy functional data in simulation experiments and using two real data applications. PMID:28955155
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dumane, V; Rhome, R; Yuan, Y
2015-06-15
Purpose: To study the influence of dimensions of the tandem and ring applicator on bladder D2cc, rectum D2cc and sigmoid D2cc in HDR treatment planning for cervical cancer. Methods: 53 plans from 13 patients treated at our institution with the tandem and ring applicator were retrospectively reviewed. Prescription doses were one of the following: 8 Gy x 3, 7 Gy x 4 and 5.5 Gy x 5. Doses to the D2ccs of the bladder, rectum and the sigmoid were recorded. These doses were normalized to their relative prescriptions doses. Correlations between the normalized bladder D2cc, rectum D2cc and sigmoid D2ccmore » were investigated and linear regression models were developed to study the dependence of these doses on the ring diameter and the applicator angle. Results: Normalized doses to the D2cc of the bladder, rectum and sigmoid showed statistically significant correlation (P < 0.05) to the applicator angle. Significant correlation was also noted for the normalized D2cc of the rectum and the sigmoid with the ring diameter. The normalized bladder D2cc was found to decrease with applicator angle on an average by 22.65% ± 4.43% while the same for the rectum and sigmoid were found to increase on an average by 14.43% ± 1.65% and 14.01% ± 1.42% respectively. Both the rectum and sigmoid D2cc reduced with increasing ring diameter by 12.93% ± 1.95% and 11.27% ± 1.79%. No correlation was observed between the normalized bladder D2cc and the ring diameter. Conclusion: Preliminary regression models developed in this study can potentially aid in the choice of the appropriate applicator angle and ring diameter for tandem and ring implant so as to optimize doses to the bladder, rectum and sigmoid.« less
PINHEIRO, Rafael S.; CRUZ-JR, Ruy J.; ANDRAUS, Wellington; DUCATTI, Liliana; MARTINO, Rodrigo B.; NACIF, Lucas S.; ROCHA-SANTOS, Vinicius; ARANTES, Rubens M; LAI, Quirino; IBUKI, Felicia S.; ROCHA, Manoel S.; D´ALBUQUERQUE, Luiz A. C.
2017-01-01
ABSTRACT Background: Computed tomography volumetry (CTV) is a useful tool for predicting graft weights (GW) for living donor liver transplantation (LDLT). Few studies have examined the correlation between CTV and GW in normal liver parenchyma. Aim: To analyze the correlation between CTV and GW in an adult LDLT population and provide a systematic review of the existing mathematical models to calculate partial liver graft weight. Methods: Between January 2009 and January 2013, 28 consecutive donors undergoing right hepatectomy for LDLT were retrospectively reviewed. All grafts were perfused with HTK solution. Estimated graft volume was estimated by CTV and these values were compared to the actual graft weight, which was measured after liver harvesting and perfusion. Results: Median actual GW was 782.5 g, averaged 791.43±136 g and ranged from 520-1185 g. Median estimated graft volume was 927.5 ml, averaged 944.86±200.74 ml and ranged from 600-1477 ml. Linear regression of estimated graft volume and actual GW was significantly linear (GW=0.82 estimated graft volume, r2=0.98, slope=0.47, standard deviation of 0.024 and p<0.0001). Spearman Linear correlation was 0.65 with 95% CI of 0.45 - 0.99 (p<0.0001). Conclusion: The one-to-one rule did not applied in patients with normal liver parenchyma. A better estimation of graft weight could be reached by multiplying estimated graft volume by 0.82. PMID:28489167
Quantiles for Finite Mixtures of Normal Distributions
ERIC Educational Resources Information Center
Rahman, Mezbahur; Rahman, Rumanur; Pearson, Larry M.
2006-01-01
Quantiles for finite mixtures of normal distributions are computed. The difference between a linear combination of independent normal random variables and a linear combination of independent normal densities is emphasized. (Contains 3 tables and 1 figure.)
Musuku, Adrien; Tan, Aimin; Awaiye, Kayode; Trabelsi, Fethi
2013-09-01
Linear calibration is usually performed using eight to ten calibration concentration levels in regulated LC-MS bioanalysis because a minimum of six are specified in regulatory guidelines. However, we have previously reported that two-concentration linear calibration is as reliable as or even better than using multiple concentrations. The purpose of this research is to compare two-concentration with multiple-concentration linear calibration through retrospective data analysis of multiple bioanalytical projects that were conducted in an independent regulated bioanalytical laboratory. A total of 12 bioanalytical projects were randomly selected: two validations and two studies for each of the three most commonly used types of sample extraction methods (protein precipitation, liquid-liquid extraction, solid-phase extraction). When the existing data were retrospectively linearly regressed using only the lowest and the highest concentration levels, no extra batch failure/QC rejection was observed and the differences in accuracy and precision between the original multi-concentration regression and the new two-concentration linear regression are negligible. Specifically, the differences in overall mean apparent bias (square root of mean individual bias squares) are within the ranges of -0.3% to 0.7% and 0.1-0.7% for the validations and studies, respectively. The differences in mean QC concentrations are within the ranges of -0.6% to 1.8% and -0.8% to 2.5% for the validations and studies, respectively. The differences in %CV are within the ranges of -0.7% to 0.9% and -0.3% to 0.6% for the validations and studies, respectively. The average differences in study sample concentrations are within the range of -0.8% to 2.3%. With two-concentration linear regression, an average of 13% of time and cost could have been saved for each batch together with 53% of saving in the lead-in for each project (the preparation of working standard solutions, spiking, and aliquoting). Furthermore, examples are given as how to evaluate the linearity over the entire concentration range when only two concentration levels are used for linear regression. To conclude, two-concentration linear regression is accurate and robust enough for routine use in regulated LC-MS bioanalysis and it significantly saves time and cost as well. Copyright © 2013 Elsevier B.V. All rights reserved.
A Linear Regression and Markov Chain Model for the Arabian Horse Registry
1993-04-01
as a tax deduction? Yes No T-4367 68 26. Regardless of previous equine tax deductions, do you consider your current horse activities to be... (Mark one...E L T-4367 A Linear Regression and Markov Chain Model For the Arabian Horse Registry Accesion For NTIS CRA&I UT 7 4:iC=D 5 D-IC JA" LI J:13tjlC,3 lO...the Arabian Horse Registry, which needed to forecast its future registration of purebred Arabian horses . A linear regression model was utilized to
Diaphragm depth in normal subjects.
Shahgholi, Leili; Baria, Michael R; Sorenson, Eric J; Harper, Caitlin J; Watson, James C; Strommen, Jeffrey A; Boon, Andrea J
2014-05-01
Needle electromyography (EMG) of the diaphragm carries the potential risk of pneumothorax. Knowing the approximate depth of the diaphragm should increase the test's safety and accuracy. Distances from the skin to the diaphragm and from the outer surface of the rib to the diaphragm were measured using B mode ultrasound in 150 normal subjects. When measured at the lower intercostal spaces, diaphragm depth varied between 0.78 and 4.91 cm beneath the skin surface and between 0.25 and 1.48 cm below the outer surface of the rib. Using linear regression modeling, body mass index (BMI) could be used to predict diaphragm depth from the skin to within an average of 1.15 mm. Diaphragm depth from the skin can vary by more than 4 cm. When image guidance is not available to enhance accuracy and safety of diaphragm EMG, it is possible to reliably predict the depth of the diaphragm based on BMI. Copyright © 2013 Wiley Periodicals, Inc.
Stevanović, Nikola R; Perušković, Danica S; Gašić, Uroš M; Antunović, Vesna R; Lolić, Aleksandar Đ; Baošić, Rada M
2017-03-01
The objectives of this study were to gain insights into structure-retention relationships and to propose the model to estimating their retention. Chromatographic investigation of series of 36 Schiff bases and their copper(II) and nickel(II) complexes was performed under both normal- and reverse-phase conditions. Chemical structures of the compounds were characterized by molecular descriptors which are calculated from the structure and related to the chromatographic retention parameters by multiple linear regression analysis. Effects of chelation on retention parameters of investigated compounds, under normal- and reverse-phase chromatographic conditions, were analyzed by principal component analysis, quantitative structure-retention relationship and quantitative structure-activity relationship models were developed on the basis of theoretical molecular descriptors, calculated exclusively from molecular structure, and parameters of retention and lipophilicity. Copyright © 2016 John Wiley & Sons, Ltd.
Impact of LANDSAT MSS sensor differences on change detection analysis
NASA Technical Reports Server (NTRS)
Likens, W. C.; Wrigley, R. C.
1983-01-01
Some 512 by 512 pixel subwindows for simultaneously acquired scene pairs obtained by LANDSAT 2,3 and 4 multispectral band scanners were coregistered using LANDSAT 4 scenes as the base to which the other images were registered. Scattergrams between the coregistered scenes (a form of contingency analysis) were used to radiometrically compare data from the various sensors. Mode values were derived and used to visually fit a linear regression. Root mean square errors of the registration varied between .1 and 1.5 pixels. There appear to be no major problem preventing the use of LANDSAT 4 MSS with previous MSS sensors for change detection, provided the noise interference can be removed or minimized. Data normalizations for change detection should be based on the data rather than solely on calibration information. This allows simultaneous normalization of the atmosphere as well as the radiometry.
An improved multiple linear regression and data analysis computer program package
NASA Technical Reports Server (NTRS)
Sidik, S. M.
1972-01-01
NEWRAP, an improved version of a previous multiple linear regression program called RAPIER, CREDUC, and CRSPLT, allows for a complete regression analysis including cross plots of the independent and dependent variables, correlation coefficients, regression coefficients, analysis of variance tables, t-statistics and their probability levels, rejection of independent variables, plots of residuals against the independent and dependent variables, and a canonical reduction of quadratic response functions useful in optimum seeking experimentation. A major improvement over RAPIER is that all regression calculations are done in double precision arithmetic.
Douglas, R K; Nawar, S; Alamar, M C; Mouazen, A M; Coulon, F
2018-03-01
Visible and near infrared spectrometry (vis-NIRS) coupled with data mining techniques can offer fast and cost-effective quantitative measurement of total petroleum hydrocarbons (TPH) in contaminated soils. Literature showed however significant differences in the performance on the vis-NIRS between linear and non-linear calibration methods. This study compared the performance of linear partial least squares regression (PLSR) with a nonlinear random forest (RF) regression for the calibration of vis-NIRS when analysing TPH in soils. 88 soil samples (3 uncontaminated and 85 contaminated) collected from three sites located in the Niger Delta were scanned using an analytical spectral device (ASD) spectrophotometer (350-2500nm) in diffuse reflectance mode. Sequential ultrasonic solvent extraction-gas chromatography (SUSE-GC) was used as reference quantification method for TPH which equal to the sum of aliphatic and aromatic fractions ranging between C 10 and C 35 . Prior to model development, spectra were subjected to pre-processing including noise cut, maximum normalization, first derivative and smoothing. Then 65 samples were selected as calibration set and the remaining 20 samples as validation set. Both vis-NIR spectrometry and gas chromatography profiles of the 85 soil samples were subjected to RF and PLSR with leave-one-out cross-validation (LOOCV) for the calibration models. Results showed that RF calibration model with a coefficient of determination (R 2 ) of 0.85, a root means square error of prediction (RMSEP) 68.43mgkg -1 , and a residual prediction deviation (RPD) of 2.61 outperformed PLSR (R 2 =0.63, RMSEP=107.54mgkg -1 and RDP=2.55) in cross-validation. These results indicate that RF modelling approach is accounting for the nonlinearity of the soil spectral responses hence, providing significantly higher prediction accuracy compared to the linear PLSR. It is recommended to adopt the vis-NIRS coupled with RF modelling approach as a portable and cost effective method for the rapid quantification of TPH in soils. Copyright © 2017 Elsevier B.V. All rights reserved.
IN11B-1621: Quantifying How Climate Affects Vegetation in the Amazon Rainforest
NASA Technical Reports Server (NTRS)
Das, Kamalika; Kodali, Anuradha; Szubert, Marcin; Ganguly, Sangram; Bongard, Joshua
2016-01-01
Amazon droughts in 2005 and 2010 have raised serious concern about the future of the rainforest. Amazon forests are crucial because of their role as the largest carbon sink in the world which would effect the global warming phenomena with decreased photosynthesis activity. Especially, after a decline in plant growth in 1.68 million km2 forest area during the once-in-a-century severe drought in 2010, it is of primary importance to understand the relationship between different climatic variables and vegetation. In an earlier study, we have shown that non-linear models are better at capturing the relation dynamics of vegetation and climate variables such as temperature and precipitation, compared to linear models. In this research, we learn precise models between vegetation and climatic variables (temperature, precipitation) for normal conditions in the Amazon region using genetic programming based symbolic regression. This is done by removing high elevation and drought affected areas and also considering the slope of the region as one of the important factors while building the model. The model learned reveals new and interesting ways historical and current climate variables affect the vegetation at any location. MAIAC data has been used as a vegetation surrogate in our study. For temperature and precipitation, we have used TRMM and MODIS Land Surface Temperature data sets while learning the non-linear regression model. However, to generalize the model to make it independent of the data source, we perform transfer learning where we regress a regularized least squares to learn the parameters of the non-linear model using other data sources such as the precipitation and temperature from the Climatic Research Center (CRU). This new model is very similar in structure and performance compared to the original learned model and verifies the same claims about the nature of dependency between these climate variables and the vegetation in the Amazon region. As a result of this study, we are able to learn, for the very first time how exactly different climate factors influence vegetation at any location in the Amazon rainforests, independent of the specific sources from which the data has been obtained.
Quantifying How Climate Affects Vegetation in the Amazon Rainforest
NASA Astrophysics Data System (ADS)
Das, K.; Kodali, A.; Szubert, M.; Ganguly, S.; Bongard, J.
2016-12-01
Amazon droughts in 2005 and 2010 have raised serious concern about the future of the rainforest. Amazon forests are crucial because of their role as the largest carbon sink in the world which would effect the global warming phenomena with decreased photosynthesis activity. Especially, after a decline in plant growth in 1.68 million km2 forest area during the once-in-a-century severe drought in 2010, it is of primary importance to understand the relationship between different climatic variables and vegetation. In an earlier study, we have shown that non-linear models are better at capturing the relation dynamics of vegetation and climate variables such as temperature and precipitation, compared to linear models. In this research, we learn precise models between vegetation and climatic variables (temperature, precipitation) for normal conditions in the Amazon region using genetic programming based symbolic regression. This is done by removing high elevation and drought affected areas and also considering the slope of the region as one of the important factors while building the model. The model learned reveals new and interesting ways historical and current climate variables affect the vegetation at any location. MAIAC data has been used as a vegetation surrogate in our study. For temperature and precipitation, we have used TRMM and MODIS Land Surface Temperature data sets while learning the non-linear regression model. However, to generalize the model to make it independent of the data source, we perform transfer learning where we regress a regularized least squares to learn the parameters of the non-linear model using other data sources such as the precipitation and temperature from the Climatic Research Center (CRU). This new model is very similar in structure and performance compared to the original learned model and verifies the same claims about the nature of dependency between these climate variables and the vegetation in the Amazon region. As a result of this study, we are able to learn, for the very first time how exactly different climate factors influence vegetation at any location in the Amazon rainforests, independent of the specific sources from which the data has been obtained.
Asgary, S; Dinani, N Jafari; Madani, H; Mahzouni, P
2008-05-01
Artemisia aucheri is a native-growing plant which is widely used in Iranian traditional medicine. This study was designed to evaluate the effects of A. aucheri on regression of atherosclerosis in hypercholesterolemic rabbits. Twenty five rabbits were randomly divided into five groups of five each and treated 3-months as follows: 1: normal diet, 2: hypercholesterolemic diet (HCD), 3 and 4: HCD for 60 days and then normal diet and normal diet + A. aucheri (100 mg x kg(-1) x day(-1)) respectively for an additional 30 days (regression period). In the regression period dietary use of A. aucheri in group 4 significantly decreased total cholesterol, triglyceride and LDL-cholesterol, while HDL-cholesterol was significantly increased. The atherosclerotic area was significantly decreased in this group. Animals, which received only normal diet in the regression period showed no regression but rather progression of atherosclerosis. These findings suggest that A. aucheri may cause regression of atherosclerotic lesions.
NASA Astrophysics Data System (ADS)
Kutzbach, L.; Schneider, J.; Sachs, T.; Giebels, M.; Nykänen, H.; Shurpali, N. J.; Martikainen, P. J.; Alm, J.; Wilmking, M.
2007-07-01
Closed (non-steady state) chambers are widely used for quantifying carbon dioxide (CO2) fluxes between soils or low-stature canopies and the atmosphere. It is well recognised that covering a soil or vegetation by a closed chamber inherently disturbs the natural CO2 fluxes by altering the concentration gradients between the soil, the vegetation and the overlying air. Thus, the driving factors of CO2 fluxes are not constant during the closed chamber experiment, and no linear increase or decrease of CO2 concentration over time within the chamber headspace can be expected. Nevertheless, linear regression has been applied for calculating CO2 fluxes in many recent, partly influential, studies. This approach was justified by keeping the closure time short and assuming the concentration change over time to be in the linear range. Here, we test if the application of linear regression is really appropriate for estimating CO2 fluxes using closed chambers over short closure times and if the application of nonlinear regression is necessary. We developed a nonlinear exponential regression model from diffusion and photosynthesis theory. This exponential model was tested with four different datasets of CO2 flux measurements (total number: 1764) conducted at three peatland sites in Finland and a tundra site in Siberia. The flux measurements were performed using transparent chambers on vegetated surfaces and opaque chambers on bare peat surfaces. Thorough analyses of residuals demonstrated that linear regression was frequently not appropriate for the determination of CO2 fluxes by closed-chamber methods, even if closure times were kept short. The developed exponential model was well suited for nonlinear regression of the concentration over time c(t) evolution in the chamber headspace and estimation of the initial CO2 fluxes at closure time for the majority of experiments. CO2 flux estimates by linear regression can be as low as 40% of the flux estimates of exponential regression for closure times of only two minutes and even lower for longer closure times. The degree of underestimation increased with increasing CO2 flux strength and is dependent on soil and vegetation conditions which can disturb not only the quantitative but also the qualitative evaluation of CO2 flux dynamics. The underestimation effect by linear regression was observed to be different for CO2 uptake and release situations which can lead to stronger bias in the daily, seasonal and annual CO2 balances than in the individual fluxes. To avoid serious bias of CO2 flux estimates based on closed chamber experiments, we suggest further tests using published datasets and recommend the use of nonlinear regression models for future closed chamber studies.
NASA Technical Reports Server (NTRS)
Suit, W. T.; Cannaday, R. L.
1979-01-01
The longitudinal and lateral stability and control parameters for a high wing, general aviation, airplane are examined. Estimations using flight data obtained at various flight conditions within the normal range of the aircraft are presented. The estimations techniques, an output error technique (maximum likelihood) and an equation error technique (linear regression), are presented. The longitudinal static parameters are estimated from climbing, descending, and quasi steady state flight data. The lateral excitations involve a combination of rudder and ailerons. The sensitivity of the aircraft modes of motion to variations in the parameter estimates are discussed.
NASA Technical Reports Server (NTRS)
Cromwell, R. L.; Zanello, S. B.; Yarbough, P. O.; Ploutz-Snyder, R.; Taibbi, G.; Brewer, J. L.; Vizzeri, G.
2013-01-01
Mean IOP significantly increased while at 6deg HDT and returned towards pre-bed rest values upon leaving bed rest. While mean IOP increased during bed rest, it remained within the normal limits for subject safety. A diuretic shift and cardiovascular deconditioning occurs during in-bed rest, as expected. There was no demonstrable correlation between the largest change in IOP (pre/post) and cardiovascular measure changes (pre/post). Additional mixed effects linear regression modeling may reveal some subclinical physiological changes that might assist in describing the VIIP syndrome pathophysiology.
Monitoring Springs in the Mojave Desert Using Landsat Time Series Analysis
NASA Technical Reports Server (NTRS)
Potter, Christopher S.
2018-01-01
The purpose of this study, based on Landsat satellite data was to characterize variations and trends over 30 consecutive years (1985-2016) in perennial vegetation green cover at over 400 confirmed Mojave Desert spring locations. These springs were surveyed between in 2015 and 2016 on lands managed in California by the U.S. Bureau of Land Management (BLM) and on several land trusts within the Barstow, Needles, and Ridgecrest BLM Field Offices. The normalized difference vegetation index (NDVI) from July Landsat images was computed at each spring location and a trend model was first fit to the multi-year NDVI time series using least squares linear regression.Â
Kaimakamis, Evangelos; Tsara, Venetia; Bratsas, Charalambos; Sichletidis, Lazaros; Karvounis, Charalambos; Maglaveras, Nikolaos
2016-01-01
Obstructive Sleep Apnea (OSA) is a common sleep disorder requiring the time/money consuming polysomnography for diagnosis. Alternative methods for initial evaluation are sought. Our aim was the prediction of Apnea-Hypopnea Index (AHI) in patients potentially suffering from OSA based on nonlinear analysis of respiratory biosignals during sleep, a method that is related to the pathophysiology of the disorder. Patients referred to a Sleep Unit (135) underwent full polysomnography. Three nonlinear indices (Largest Lyapunov Exponent, Detrended Fluctuation Analysis and Approximate Entropy) extracted from two biosignals (airflow from a nasal cannula, thoracic movement) and one linear derived from Oxygen saturation provided input to a data mining application with contemporary classification algorithms for the creation of predictive models for AHI. A linear regression model presented a correlation coefficient of 0.77 in predicting AHI. With a cutoff value of AHI = 8, the sensitivity and specificity were 93% and 71.4% in discrimination between patients and normal subjects. The decision tree for the discrimination between patients and normal had sensitivity and specificity of 91% and 60%, respectively. Certain obtained nonlinear values correlated significantly with commonly accepted physiological parameters of people suffering from OSA. We developed a predictive model for the presence/severity of OSA using a simple linear equation and additional decision trees with nonlinear features extracted from 3 respiratory recordings. The accuracy of the methodology is high and the findings provide insight to the underlying pathophysiology of the syndrome. Reliable predictions of OSA are possible using linear and nonlinear indices from only 3 respiratory signals during sleep. The proposed models could lead to a better study of the pathophysiology of OSA and facilitate initial evaluation/follow up of suspected patients OSA utilizing a practical low cost methodology. ClinicalTrials.gov NCT01161381.
Karan, Shivesh Kishore; Samadder, Sukha Ranjan; Maiti, Subodh Kumar
2016-11-01
The objective of the present study is to monitor reclamation activity in mining areas. Monitoring of these reclaimed sites in the vicinity of mining areas and on closed Over Burden (OB) dumps is critical for improving the overall environmental condition, especially in developing countries where area around the mines are densely populated. The present study evaluated the reclamation success in the Block II area of Jharia coal field, India, using Landsat satellite images for the years 2000 and 2015. Four image processing methods (support vector machine, ratio vegetation index, enhanced vegetation index, and normalized difference vegetation index) were used to quantify the change in vegetation cover between the years 2000 and 2015. The study also evaluated the relationship between vegetation health and moisture content of the study area using remote sensing techniques. Statistical linear regression analysis revealed that Normalized Difference Vegetation Index (NDVI) coupled with Normalized Difference Moisture Index (NDMI) is the best method for vegetation monitoring in the study area when compared to other indices. A strong linear relationship (r(2) > 0.86) was found between NDVI and NDMI. An increase of 21% from 213.88 ha in 2000 to 258.9 ha in 2015 was observed in the vegetation cover of the reclaimed sites for an open cast mine, indicating satisfactory reclamation activity. NDVI results indicated that vegetation health also improved over the years. Copyright © 2016 Elsevier Ltd. All rights reserved.
ERIC Educational Resources Information Center
Bulcock, J. W.
The problem of model estimation when the data are collinear was examined. Though the ridge regression (RR) outperforms ordinary least squares (OLS) regression in the presence of acute multicollinearity, it is not a problem free technique for reducing the variance of the estimates. It is a stochastic procedure when it should be nonstochastic and it…
ERIC Educational Resources Information Center
Lee, Wan-Fung; Bulcock, Jeffrey Wilson
The purposes of this study are: (1) to demonstrate the superiority of simple ridge regression over ordinary least squares regression through theoretical argument and empirical example; (2) to modify ridge regression through use of the variance normalization criterion; and (3) to demonstrate the superiority of simple ridge regression based on the…
Bogle, Jamie M; Zapala, David A; Criter, Robin; Burkard, Robert
2013-02-01
The cervical vestibular evoked myogenic potential (cVEMP) is a reflexive change in sternocleidomastoid (SCM) muscle contraction activity thought to be mediated by a saccular vestibulo-collic reflex. CVEMP amplitude varies with the state of the afferent (vestibular) limb of the vestibulo-collic reflex pathway, as well as with the level of SCM muscle contraction. It follows that in order for cVEMP amplitude to reflect the status of the afferent portion of the reflex pathway, muscle contraction level must be controlled. Historically, this has been accomplished by volitionally controlling muscle contraction level either with the aid of a biofeedback method, or by an a posteriori method that normalizes cVEMP amplitude by the level of muscle contraction. A posteriori normalization methods make the implicit assumption that mathematical normalization precisely removes the influence of the efferent limb of the vestibulo-collic pathway. With the cVEMP, however, we are violating basic assumptions of signal averaging: specifically, the background noise and the response are not independent. The influence of this signal-averaging violation on our ability to normalize cVEMP amplitude using a posteriori methods is not well understood. The aims of this investigation were to describe the effect of muscle contraction, as measured by a prestimulus electromyogenic estimate, on cVEMP amplitude and interaural amplitude asymmetry ratio, and to evaluate the benefit of using a commonly advocated a posteriori normalization method on cVEMP amplitude and asymmetry ratio variability. Prospective, repeated-measures design using a convenience sample. Ten healthy adult participants between 25 and 61 yr of age. cVEMP responses to 500 Hz tone bursts (120 dB pSPL) for three conditions describing maximum, moderate, and minimal muscle contraction. Mean (standard deviation) cVEMP amplitude and asymmetry ratios were calculated for each muscle-contraction condition. Repeated measures analysis of variance and t-tests compared the variability in cVEMP amplitude between sides and conditions. Linear regression analyses compared asymmetry ratios. Polynomial regression analyses described the corrected and uncorrected cVEMP amplitude growth functions. While cVEMP amplitude increased with increased muscle contraction, the relationship was not linear or even proportionate. In the majority of cases, once muscle contraction reached a certain "threshold" level, cVEMP amplitude increased rapidly and then saturated. Normalizing cVEMP amplitudes did not remove the relationship between cVEMP amplitude and muscle contraction level. As muscle contraction increased, the normalized amplitude increased, and then decreased, corresponding with the observed amplitude saturation. Abnormal asymmetry ratios (based on values reported in the literature) were noted for four instances of uncorrected amplitude asymmetry at less than maximum muscle contraction levels. Amplitude normalization did not substantially change the number of observed asymmetry ratios. Because cVEMP amplitude did not typically grow proportionally with muscle contraction level, amplitude normalization did not lead to stable cVEMP amplitudes or asymmetry ratios across varying muscle contraction levels. Until we better understand the relationships between muscle contraction level, surface electromyography (EMG) estimates of muscle contraction level, and cVEMP amplitude, the application of normalization methods to correct cVEMP amplitude appears unjustified. American Academy of Audiology.
Ghassemi, Fariba; Mirshahi, Reza; Bazvand, Fatemeh; Fadakar, Kaveh; Faghihi, Houshang; Sabour, Siamak
2017-12-01
To provide normative data of foveal avascular zone (FAZ) and thickness. In this cross-sectional study both eyes of each normal subject were scanned with optical coherence tomography angiography (OCTA) for foveal superficial and deep avascular zone (FAZ) and central foveal thickness (CFT) and parafoveal thickness (PFT). Out of a total of 224 eyes of 112 volunteers with a mean age of 37.03 (12-67) years, the mean superficial FAZ area was 0.27 mm 2 , and deep FAZ area was 0.35 mm 2 ( P < 0.001), with no difference between both eyes. Females had a larger superficial (0.32 ± 0.11 mm 2 versus 0.23 ± 0.09 mm 2 ) and deep FAZ (0.40 ± 0.14 mm 2 versus 0.31 ± 0.10 mm 2 ) ( P < 0.001) than males. By multivariate linear regression analysis, in normal eyes, superficial FAZ area varied significantly with the gender, CFT, and deep FAZ. Deep FAZ varied with the gender and CFT. The gender and CFT influence the size of normal superficial and deep FAZ of capillary network.
ERIC Educational Resources Information Center
Quinino, Roberto C.; Reis, Edna A.; Bessegato, Lupercio F.
2013-01-01
This article proposes the use of the coefficient of determination as a statistic for hypothesis testing in multiple linear regression based on distributions acquired by beta sampling. (Contains 3 figures.)
Statistical power analyses using G*Power 3.1: tests for correlation and regression analyses.
Faul, Franz; Erdfelder, Edgar; Buchner, Axel; Lang, Albert-Georg
2009-11-01
G*Power is a free power analysis program for a variety of statistical tests. We present extensions and improvements of the version introduced by Faul, Erdfelder, Lang, and Buchner (2007) in the domain of correlation and regression analyses. In the new version, we have added procedures to analyze the power of tests based on (1) single-sample tetrachoric correlations, (2) comparisons of dependent correlations, (3) bivariate linear regression, (4) multiple linear regression based on the random predictor model, (5) logistic regression, and (6) Poisson regression. We describe these new features and provide a brief introduction to their scope and handling.
Rasmussen, Patrick P.; Gray, John R.; Glysson, G. Douglas; Ziegler, Andrew C.
2009-01-01
In-stream continuous turbidity and streamflow data, calibrated with measured suspended-sediment concentration data, can be used to compute a time series of suspended-sediment concentration and load at a stream site. Development of a simple linear (ordinary least squares) regression model for computing suspended-sediment concentrations from instantaneous turbidity data is the first step in the computation process. If the model standard percentage error (MSPE) of the simple linear regression model meets a minimum criterion, this model should be used to compute a time series of suspended-sediment concentrations. Otherwise, a multiple linear regression model using paired instantaneous turbidity and streamflow data is developed and compared to the simple regression model. If the inclusion of the streamflow variable proves to be statistically significant and the uncertainty associated with the multiple regression model results in an improvement over that for the simple linear model, the turbidity-streamflow multiple linear regression model should be used to compute a suspended-sediment concentration time series. The computed concentration time series is subsequently used with its paired streamflow time series to compute suspended-sediment loads by standard U.S. Geological Survey techniques. Once an acceptable regression model is developed, it can be used to compute suspended-sediment concentration beyond the period of record used in model development with proper ongoing collection and analysis of calibration samples. Regression models to compute suspended-sediment concentrations are generally site specific and should never be considered static, but they represent a set period in a continually dynamic system in which additional data will help verify any change in sediment load, type, and source.
Sano, Yuko; Kandori, Akihiko; Shima, Keisuke; Yamaguchi, Yuki; Tsuji, Toshio; Noda, Masafumi; Higashikawa, Fumiko; Yokoe, Masaru; Sakoda, Saburo
2016-06-01
We propose a novel index of Parkinson's disease (PD) finger-tapping severity, called "PDFTsi," for quantifying the severity of symptoms related to the finger tapping of PD patients with high accuracy. To validate the efficacy of PDFTsi, the finger-tapping movements of normal controls and PD patients were measured by using magnetic sensors, and 21 characteristics were extracted from the finger-tapping waveforms. To distinguish motor deterioration due to PD from that due to aging, the aging effect on finger tapping was removed from these characteristics. Principal component analysis (PCA) was applied to the age-normalized characteristics, and principal components that represented the motion properties of finger tapping were calculated. Multiple linear regression (MLR) with stepwise variable selection was applied to the principal components, and PDFTsi was calculated. The calculated PDFTsi indicates that PDFTsi has a high estimation ability, namely a mean square error of 0.45. The estimation ability of PDFTsi is higher than that of the alternative method, MLR with stepwise regression selection without PCA, namely a mean square error of 1.30. This result suggests that PDFTsi can quantify PD finger-tapping severity accurately. Furthermore, the result of interpreting a model for calculating PDFTsi indicated that motion wideness and rhythm disorder are important for estimating PD finger-tapping severity.
Davies-Venn, Evelyn; Nelson, Peggy; Souza, Pamela
2015-01-01
Some listeners with hearing loss show poor speech recognition scores in spite of using amplification that optimizes audibility. Beyond audibility, studies have suggested that suprathreshold abilities such as spectral and temporal processing may explain differences in amplified speech recognition scores. A variety of different methods has been used to measure spectral processing. However, the relationship between spectral processing and speech recognition is still inconclusive. This study evaluated the relationship between spectral processing and speech recognition in listeners with normal hearing and with hearing loss. Narrowband spectral resolution was assessed using auditory filter bandwidths estimated from simultaneous notched-noise masking. Broadband spectral processing was measured using the spectral ripple discrimination (SRD) task and the spectral ripple depth detection (SMD) task. Three different measures were used to assess unamplified and amplified speech recognition in quiet and noise. Stepwise multiple linear regression revealed that SMD at 2.0 cycles per octave (cpo) significantly predicted speech scores for amplified and unamplified speech in quiet and noise. Commonality analyses revealed that SMD at 2.0 cpo combined with SRD and equivalent rectangular bandwidth measures to explain most of the variance captured by the regression model. Results suggest that SMD and SRD may be promising clinical tools for diagnostic evaluation and predicting amplification outcomes. PMID:26233047
Davies-Venn, Evelyn; Nelson, Peggy; Souza, Pamela
2015-07-01
Some listeners with hearing loss show poor speech recognition scores in spite of using amplification that optimizes audibility. Beyond audibility, studies have suggested that suprathreshold abilities such as spectral and temporal processing may explain differences in amplified speech recognition scores. A variety of different methods has been used to measure spectral processing. However, the relationship between spectral processing and speech recognition is still inconclusive. This study evaluated the relationship between spectral processing and speech recognition in listeners with normal hearing and with hearing loss. Narrowband spectral resolution was assessed using auditory filter bandwidths estimated from simultaneous notched-noise masking. Broadband spectral processing was measured using the spectral ripple discrimination (SRD) task and the spectral ripple depth detection (SMD) task. Three different measures were used to assess unamplified and amplified speech recognition in quiet and noise. Stepwise multiple linear regression revealed that SMD at 2.0 cycles per octave (cpo) significantly predicted speech scores for amplified and unamplified speech in quiet and noise. Commonality analyses revealed that SMD at 2.0 cpo combined with SRD and equivalent rectangular bandwidth measures to explain most of the variance captured by the regression model. Results suggest that SMD and SRD may be promising clinical tools for diagnostic evaluation and predicting amplification outcomes.
Rapid detection of talcum powder in tea using FT-IR spectroscopy coupled with chemometrics
Li, Xiaoli; Zhang, Yuying; He, Yong
2016-01-01
This paper investigated the feasibility of Fourier transform infrared transmission (FT-IR) spectroscopy to detect talcum powder illegally added in tea based on chemometric methods. Firstly, 210 samples of tea powder with 13 dose levels of talcum powder were prepared for FT-IR spectra acquirement. In order to highlight the slight variations in FT-IR spectra, smoothing, normalize and standard normal variate (SNV) were employed to preprocess the raw spectra. Among them, SNV preprocessing had the best performance with high correlation of prediction (RP = 0.948) and low root mean square error of prediction (RMSEP = 0.108) of partial least squares (PLS) model. Then 18 characteristic wavenumbers were selected based on a hybrid of backward interval partial least squares (biPLS) regression, competitive adaptive reweighted sampling (CARS) algorithm and successive projections algorithm (SPA). These characteristic wavenumbers only accounted for 0.64% of the full wavenumbers. Following that, 18 characteristic wavenumbers were used to build linear and nonlinear determination models by PLS regression and extreme learning machine (ELM), respectively. The optimal model with RP = 0.963 and RMSEP = 0.137 was achieved by ELM algorithm. These results demonstrated that FT-IR spectroscopy with chemometrics could be used successfully to detect talcum powder in tea. PMID:27468701
Zhang, Jinping; Wang, Na; Xing, Xiaoyan; Yang, Zhaojun; Wang, Xin; Yang, Wenying
2016-01-01
To conduct a subanalysis of the randomized MARCH (Metformin and AcaRbose in Chinese as the initial Hypoglycemic treatment) trial to investigate whether specific characteristics are associated with the efficacy of either acarbose or metformin as initial therapy. A total of 657 type 2 diabetes patients who were randomly assigned to 48 weeks of therapy with either acarbose or metformin in the MARCH trial were divided into two groups based upon their hemoglobin A1c (HbA1c) levels at the end of follow-up: HbA1c <7% (<53 mmol/mol) and ≥7% (≥53 mmol/mol). Univariate, multivariate, and stepwise linear regression analyses were applied to identify the factors associated with treatment efficacy. Because this was a subanalysis, no measurement was performed. Univariate analysis showed that the efficacy of acarbose and metformin was influenced by HbA1c, fasting blood glucose (FBG), and 2 hour postprandial venous blood glucose (2hPPG) levels, as well as by changes in body mass index (BMI) (p ≤ 0.006). Multivariate analysis and stepwise linear regression analyses indicated that lower baseline 2hPPG values and greater changes in BMI were factors that positively influenced efficacy in both treatment groups (p ≤ 0.05). Stepwise regression model analysis also revealed that a lower baseline homeostasis model assessment-estimated insulin resistance (HOMA-IR) and higher serum insulin area under the curve (AUC) were factors positively influencing HbA1c normalization in all patients (p ≤ 0.032). Newly diagnosed type 2 diabetes patients with lower baseline 2hPPG and HOMA-IR values are more likely to achieve glucose control with acarbose or metformin treatment. Furthermore, the change in BMI after acarbose or metformin treatment is also a factor influencing HbA1c normalization. A prospective study with a larger sample size is necessary to confirm our results as well as measure β cell function and examine the influence of the patients' dietary habits.
Holsclaw, Tracy; Hallgren, Kevin A; Steyvers, Mark; Smyth, Padhraic; Atkins, David C
2015-12-01
Behavioral coding is increasingly used for studying mechanisms of change in psychosocial treatments for substance use disorders (SUDs). However, behavioral coding data typically include features that can be problematic in regression analyses, including measurement error in independent variables, non normal distributions of count outcome variables, and conflation of predictor and outcome variables with third variables, such as session length. Methodological research in econometrics has shown that these issues can lead to biased parameter estimates, inaccurate standard errors, and increased Type I and Type II error rates, yet these statistical issues are not widely known within SUD treatment research, or more generally, within psychotherapy coding research. Using minimally technical language intended for a broad audience of SUD treatment researchers, the present paper illustrates the nature in which these data issues are problematic. We draw on real-world data and simulation-based examples to illustrate how these data features can bias estimation of parameters and interpretation of models. A weighted negative binomial regression is introduced as an alternative to ordinary linear regression that appropriately addresses the data characteristics common to SUD treatment behavioral coding data. We conclude by demonstrating how to use and interpret these models with data from a study of motivational interviewing. SPSS and R syntax for weighted negative binomial regression models is included in online supplemental materials. (c) 2016 APA, all rights reserved).
Holsclaw, Tracy; Hallgren, Kevin A.; Steyvers, Mark; Smyth, Padhraic; Atkins, David C.
2015-01-01
Behavioral coding is increasingly used for studying mechanisms of change in psychosocial treatments for substance use disorders (SUDs). However, behavioral coding data typically include features that can be problematic in regression analyses, including measurement error in independent variables, non-normal distributions of count outcome variables, and conflation of predictor and outcome variables with third variables, such as session length. Methodological research in econometrics has shown that these issues can lead to biased parameter estimates, inaccurate standard errors, and increased type-I and type-II error rates, yet these statistical issues are not widely known within SUD treatment research, or more generally, within psychotherapy coding research. Using minimally-technical language intended for a broad audience of SUD treatment researchers, the present paper illustrates the nature in which these data issues are problematic. We draw on real-world data and simulation-based examples to illustrate how these data features can bias estimation of parameters and interpretation of models. A weighted negative binomial regression is introduced as an alternative to ordinary linear regression that appropriately addresses the data characteristics common to SUD treatment behavioral coding data. We conclude by demonstrating how to use and interpret these models with data from a study of motivational interviewing. SPSS and R syntax for weighted negative binomial regression models is included in supplementary materials. PMID:26098126
Engine With Regression and Neural Network Approximators Designed
NASA Technical Reports Server (NTRS)
Patnaik, Surya N.; Hopkins, Dale A.
2001-01-01
At the NASA Glenn Research Center, the NASA engine performance program (NEPP, ref. 1) and the design optimization testbed COMETBOARDS (ref. 2) with regression and neural network analysis-approximators have been coupled to obtain a preliminary engine design methodology. The solution to a high-bypass-ratio subsonic waverotor-topped turbofan engine, which is shown in the preceding figure, was obtained by the simulation depicted in the following figure. This engine is made of 16 components mounted on two shafts with 21 flow stations. The engine is designed for a flight envelope with 47 operating points. The design optimization utilized both neural network and regression approximations, along with the cascade strategy (ref. 3). The cascade used three algorithms in sequence: the method of feasible directions, the sequence of unconstrained minimizations technique, and sequential quadratic programming. The normalized optimum thrusts obtained by the three methods are shown in the following figure: the cascade algorithm with regression approximation is represented by a triangle, a circle is shown for the neural network solution, and a solid line indicates original NEPP results. The solutions obtained from both approximate methods lie within one standard deviation of the benchmark solution for each operating point. The simulation improved the maximum thrust by 5 percent. The performance of the linear regression and neural network methods as alternate engine analyzers was found to be satisfactory for the analysis and operation optimization of air-breathing propulsion engines (ref. 4).
Tumor necrosis factor- α, adiponectin and their ratio in gestational diabetes mellitus
Khosrowbeygi, Ali; Rezvanfar, Mohammad Reza; Ahmadvand, Hassan
2018-01-01
Background: It has been suggested that inflammation might be implicated in the gestational diabetes mellitus (GDM) complications, including insulin resistance. The aims of the current study were to explore maternal circulating values of TNF-α, adiponectin and the adiponectin/TNF-α ratio in women with GDM compared with normal pregnancy and their relationships with metabolic syndrome biomarkers. Methods: Forty women with GDM and 40 normal pregnant women were included in the study. Commercially available enzyme-linked immunosorbent assay methods were used to measure serum levels of TNF-α and total adiponectin. Results: Women with GDM had higher values of TNF-α (225.08±27.35 vs 115.68±12.64 pg/ml, p<0.001) and lower values of adiponectin (4.50±0.38 vs 6.37±0.59 µg/ml, p=0.003) and the adiponectin/TNF-α ratio (4.31±0.05 vs 4.80±0.07, P<0.001) than normal pregnant women. The adiponectin/TNF-α ratio showed negative correlations with insulin resistance (r=-0.68, p<0.001) and triglyceride (r=-0.39, p=0.014) and a positive correlation with insulin sensitivity (r=0.69, p<0.001). Multiple linear regression analysis showed that values of the adiponectin /TNF-α ratio were independently associated with insulin resistance. Binary logistic regression analysis showed that GDM was negatively associated with adiponectin /TNF-α ratio. Conclusions: In summary, the adiponectin/TNF-α ratio decreased significantly in GDM compared with normal pregnancy. The ratio might be an informative biomarker for assessment of pregnant women at high risk of insulin resistance and dyslipidemia and for diagnosis and therapeutic monitoring aims in GDM. PMID:29387323
Rode, Line; Kjærgaard, Hanne; Ottesen, Bent; Damm, Peter; Hegaard, Hanne K
2012-02-01
Our aim was to investigate the association between gestational weight gain (GWG) and postpartum weight retention (PWR) in pre-pregnancy underweight, normal weight, overweight or obese women, with emphasis on the American Institute of Medicine (IOM) recommendations. We performed secondary analyses on data based on questionnaires from 1,898 women from the "Smoke-free Newborn Study" conducted 1996-1999 at Hvidovre Hospital, Denmark. Relationship between GWG and PWR was examined according to BMI as a continuous variable and in four groups. Association between PWR and GWG according to IOM recommendations was tested by linear regression analysis and the association between PWR ≥ 5 kg (11 lbs) and GWG by logistic regression analysis. Mean GWG and mean PWR were constant for all BMI units until 26-27 kg/m(2). After this cut-off mean GWG and mean PWR decreased with increasing BMI. Nearly 40% of normal weight, 60% of overweight and 50% of obese women gained more than recommended during pregnancy. For normal weight and overweight women with GWG above recommendations the OR of gaining ≥ 5 kg (11 lbs) 1-year postpartum was 2.8 (95% CI 2.0-4.0) and 2.8 (95% CI 1.3-6.2, respectively) compared to women with GWG within recommendations. GWG above IOM recommendations significantly increases normal weight, overweight and obese women's risk of retaining weight 1 year after delivery. Health personnel face a challenge in prenatal counseling as 40-60% of these women gain more weight than recommended for their BMI. As GWG is potentially modifiable, our study should be followed by intervention studies focusing on GW.
NASA Astrophysics Data System (ADS)
Kutzbach, L.; Schneider, J.; Sachs, T.; Giebels, M.; Nykänen, H.; Shurpali, N. J.; Martikainen, P. J.; Alm, J.; Wilmking, M.
2007-11-01
Closed (non-steady state) chambers are widely used for quantifying carbon dioxide (CO2) fluxes between soils or low-stature canopies and the atmosphere. It is well recognised that covering a soil or vegetation by a closed chamber inherently disturbs the natural CO2 fluxes by altering the concentration gradients between the soil, the vegetation and the overlying air. Thus, the driving factors of CO2 fluxes are not constant during the closed chamber experiment, and no linear increase or decrease of CO2 concentration over time within the chamber headspace can be expected. Nevertheless, linear regression has been applied for calculating CO2 fluxes in many recent, partly influential, studies. This approach has been justified by keeping the closure time short and assuming the concentration change over time to be in the linear range. Here, we test if the application of linear regression is really appropriate for estimating CO2 fluxes using closed chambers over short closure times and if the application of nonlinear regression is necessary. We developed a nonlinear exponential regression model from diffusion and photosynthesis theory. This exponential model was tested with four different datasets of CO2 flux measurements (total number: 1764) conducted at three peatlands sites in Finland and a tundra site in Siberia. Thorough analyses of residuals demonstrated that linear regression was frequently not appropriate for the determination of CO2 fluxes by closed-chamber methods, even if closure times were kept short. The developed exponential model was well suited for nonlinear regression of the concentration over time c(t) evolution in the chamber headspace and estimation of the initial CO2 fluxes at closure time for the majority of experiments. However, a rather large percentage of the exponential regression functions showed curvatures not consistent with the theoretical model which is considered to be caused by violations of the underlying model assumptions. Especially the effects of turbulence and pressure disturbances by the chamber deployment are suspected to have caused unexplainable curvatures. CO2 flux estimates by linear regression can be as low as 40% of the flux estimates of exponential regression for closure times of only two minutes. The degree of underestimation increased with increasing CO2 flux strength and was dependent on soil and vegetation conditions which can disturb not only the quantitative but also the qualitative evaluation of CO2 flux dynamics. The underestimation effect by linear regression was observed to be different for CO2 uptake and release situations which can lead to stronger bias in the daily, seasonal and annual CO2 balances than in the individual fluxes. To avoid serious bias of CO2 flux estimates based on closed chamber experiments, we suggest further tests using published datasets and recommend the use of nonlinear regression models for future closed chamber studies.
Horton, Megan K; Blount, Benjamin C; Valentin-Blasini, Liza; Wapner, Ronald; Whyatt, Robin; Gennings, Chris; Factor-Litvak, Pam
2015-11-01
Adequate maternal thyroid function during pregnancy is necessary for normal fetal brain development, making pregnancy a critical window of vulnerability to thyroid disrupting insults. Sodium/iodide symporter (NIS) inhibitors, namely perchlorate, nitrate, and thiocyanate, have been shown individually to competitively inhibit uptake of iodine by the thyroid. Several epidemiologic studies examined the association between these individual exposures and thyroid function. Few studies have examined the effect of this chemical mixture on thyroid function during pregnancy We examined the cross sectional association between urinary perchlorate, thiocyanate and nitrate concentrations and thyroid function among healthy pregnant women living in New York City using weighted quantile sum (WQS) regression. We measured thyroid stimulating hormone (TSH) and free thyroxine (FreeT4) in blood samples; perchlorate, thiocyanate, nitrate and iodide in urine samples collected from 284 pregnant women at 12 (±2.8) weeks gestation. We examined associations between urinary analyte concentrations and TSH or FreeT4 using linear regression or WQS adjusting for gestational age, urinary iodide and creatinine. Individual analyte concentrations in urine were significantly correlated (Spearman's r 0.4-0.5, p<0.001). Linear regression analyses did not suggest associations between individual concentrations and thyroid function. The WQS revealed a significant positive association between the weighted sum of urinary concentrations of the three analytes and increased TSH. Perchlorate had the largest weight in the index, indicating the largest contribution to the WQS. Co-exposure to perchlorate, nitrate and thiocyanate may alter maternal thyroid function, specifically TSH, during pregnancy. Copyright © 2015 The Authors. Published by Elsevier Inc. All rights reserved.
Horton, Megan K.; Blount, Benjamin C.; Valentin-Blasini, Liza; Wapner, Ronald; Whyatt, Robin; Gennings, Chris; Factor-Litvak, Pam
2015-01-01
Background Adequate maternal thyroid function during pregnancy is necessary for normal fetal brain development, making pregnancy a critical window of vulnerability to thyroid disrupting insults. Sodium/iodide symporter (NIS) inhibitors, namely perchlorate, nitrate, and thiocyanate, have been shown individually to competitively inhibit uptake of iodine by the thyroid. Several epidemiologic studies examined the association between these individual exposures and thyroid function. Few studies have examined the effect of this chemical mixture on thyroid function during pregnancy. Objectives We examined the cross sectional association between urinary perchlorate, thiocyanate and nitrate concentrations and thyroid function among healthy pregnant women living in New York City using weighted quantile sum (WQS) regression. Methods We measured thyroid stimulating hormone (TSH) and free thyroxine (FreeT4) in blood samples; perchlorate, thiocyanate, nitrate and iodide in urine samples collected from 284 pregnant women at 12 (± 2.8) weeks gestation. We examined associations between urinary analyte concentrations and TSH or FreeT4 using linear regression or WQS adjusting for gestational age, urinary iodide and creatinine. Results Individual analyte concentrations in urine were significantly correlated (Spearman’s r 0.4–0.5, p < 0.001). Linear regression analyses did not suggest associations between individual concentrations and thyroid function. The WQS revealed a significant positive association between the weighted sum of urinary concentrations of the three analytes and increased TSH. Perchlorate had the largest weight in the index, indicating the largest contribution to the WQS. Conclusions Co-exposure to perchlorate, nitrate and thiocyanate may alter maternal thyroid function, specifically TSH, during pregnancy. PMID:26408806
Berglund, Lars; Garmo, Hans; Lindbäck, Johan; Svärdsudd, Kurt; Zethelius, Björn
2008-09-30
The least-squares estimator of the slope in a simple linear regression model is biased towards zero when the predictor is measured with random error. A corrected slope may be estimated by adding data from a reliability study, which comprises a subset of subjects from the main study. The precision of this corrected slope depends on the design of the reliability study and estimator choice. Previous work has assumed that the reliability study constitutes a random sample from the main study. A more efficient design is to use subjects with extreme values on their first measurement. Previously, we published a variance formula for the corrected slope, when the correction factor is the slope in the regression of the second measurement on the first. In this paper we show that both designs improve by maximum likelihood estimation (MLE). The precision gain is explained by the inclusion of data from all subjects for estimation of the predictor's variance and by the use of the second measurement for estimation of the covariance between response and predictor. The gain of MLE enhances with stronger true relationship between response and predictor and with lower precision in the predictor measurements. We present a real data example on the relationship between fasting insulin, a surrogate marker, and true insulin sensitivity measured by a gold-standard euglycaemic insulin clamp, and simulations, where the behavior of profile-likelihood-based confidence intervals is examined. MLE was shown to be a robust estimator for non-normal distributions and efficient for small sample situations. Copyright (c) 2008 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Mahaboob, B.; Venkateswarlu, B.; Sankar, J. Ravi; Balasiddamuni, P.
2017-11-01
This paper uses matrix calculus techniques to obtain Nonlinear Least Squares Estimator (NLSE), Maximum Likelihood Estimator (MLE) and Linear Pseudo model for nonlinear regression model. David Pollard and Peter Radchenko [1] explained analytic techniques to compute the NLSE. However the present research paper introduces an innovative method to compute the NLSE using principles in multivariate calculus. This study is concerned with very new optimization techniques used to compute MLE and NLSE. Anh [2] derived NLSE and MLE of a heteroscedatistic regression model. Lemcoff [3] discussed a procedure to get linear pseudo model for nonlinear regression model. In this research article a new technique is developed to get the linear pseudo model for nonlinear regression model using multivariate calculus. The linear pseudo model of Edmond Malinvaud [4] has been explained in a very different way in this paper. David Pollard et.al used empirical process techniques to study the asymptotic of the LSE (Least-squares estimation) for the fitting of nonlinear regression function in 2006. In Jae Myung [13] provided a go conceptual for Maximum likelihood estimation in his work “Tutorial on maximum likelihood estimation
Guo, Ying; Little, Roderick J; McConnell, Daniel S
2012-01-01
Covariate measurement error is common in epidemiologic studies. Current methods for correcting measurement error with information from external calibration samples are insufficient to provide valid adjusted inferences. We consider the problem of estimating the regression of an outcome Y on covariates X and Z, where Y and Z are observed, X is unobserved, but a variable W that measures X with error is observed. Information about measurement error is provided in an external calibration sample where data on X and W (but not Y and Z) are recorded. We describe a method that uses summary statistics from the calibration sample to create multiple imputations of the missing values of X in the regression sample, so that the regression coefficients of Y on X and Z and associated standard errors can be estimated using simple multiple imputation combining rules, yielding valid statistical inferences under the assumption of a multivariate normal distribution. The proposed method is shown by simulation to provide better inferences than existing methods, namely the naive method, classical calibration, and regression calibration, particularly for correction for bias and achieving nominal confidence levels. We also illustrate our method with an example using linear regression to examine the relation between serum reproductive hormone concentrations and bone mineral density loss in midlife women in the Michigan Bone Health and Metabolism Study. Existing methods fail to adjust appropriately for bias due to measurement error in the regression setting, particularly when measurement error is substantial. The proposed method corrects this deficiency.
A method for fitting regression splines with varying polynomial order in the linear mixed model.
Edwards, Lloyd J; Stewart, Paul W; MacDougall, James E; Helms, Ronald W
2006-02-15
The linear mixed model has become a widely used tool for longitudinal analysis of continuous variables. The use of regression splines in these models offers the analyst additional flexibility in the formulation of descriptive analyses, exploratory analyses and hypothesis-driven confirmatory analyses. We propose a method for fitting piecewise polynomial regression splines with varying polynomial order in the fixed effects and/or random effects of the linear mixed model. The polynomial segments are explicitly constrained by side conditions for continuity and some smoothness at the points where they join. By using a reparameterization of this explicitly constrained linear mixed model, an implicitly constrained linear mixed model is constructed that simplifies implementation of fixed-knot regression splines. The proposed approach is relatively simple, handles splines in one variable or multiple variables, and can be easily programmed using existing commercial software such as SAS or S-plus. The method is illustrated using two examples: an analysis of longitudinal viral load data from a study of subjects with acute HIV-1 infection and an analysis of 24-hour ambulatory blood pressure profiles.
Zhang, Guosheng; Huang, Kuan-Chieh; Xu, Zheng; Tzeng, Jung-Ying; Conneely, Karen N; Guan, Weihua; Kang, Jian; Li, Yun
2016-05-01
DNA methylation is a key epigenetic mark involved in both normal development and disease progression. Recent advances in high-throughput technologies have enabled genome-wide profiling of DNA methylation. However, DNA methylation profiling often employs different designs and platforms with varying resolution, which hinders joint analysis of methylation data from multiple platforms. In this study, we propose a penalized functional regression model to impute missing methylation data. By incorporating functional predictors, our model utilizes information from nonlocal probes to improve imputation quality. Here, we compared the performance of our functional model to linear regression and the best single probe surrogate in real data and via simulations. Specifically, we applied different imputation approaches to an acute myeloid leukemia dataset consisting of 194 samples and our method showed higher imputation accuracy, manifested, for example, by a 94% relative increase in information content and up to 86% more CpG sites passing post-imputation filtering. Our simulated association study further demonstrated that our method substantially improves the statistical power to identify trait-associated methylation loci. These findings indicate that the penalized functional regression model is a convenient and valuable imputation tool for methylation data, and it can boost statistical power in downstream epigenome-wide association study (EWAS). © 2016 WILEY PERIODICALS, INC.
Feng, Yuanbo; Ma, Zhan-Long; Chen, Feng; Yu, Jie; Cona, Marlein Miranda; Xie, Yi; Li, Yue; Ni, Yicheng
2013-01-01
AIM: To develop a method for studying myocardial area at risk (AAR) in ischemic heart disease in correlation with cardiac magnetic resonance imaging (cMRI). METHODS: Nine rabbits were anesthetized, intubated and subjected to occlusion and reperfusion of the left circumflex coronary artery (LCx) to induce myocardial infarction (MI). ECG-triggered cMRI with delayed enhancement was performed at 3.0 T. After euthanasia, the heart was excised with the LCx re-ligated. Bifunctional staining was performed by perfusing the aorta with a homemade red-iodized-oil (RIO) dye. The heart was then agar-embedded for ex vivo magnetic resonance imaging and sliced into 3 mm-sections. The AAR was defined by RIO-staining and digital radiography (DR). The perfusion density rate (PDR) was derived from DR for the AAR and normal myocardium. The MI was measured by in vivo delayed enhancement (iDE) and ex vivo delayed enhancement (eDE) cMRI. The AAR and MI were compared to validate the bifunctional straining for cardiac imaging research. Linear regression with Bland-Altman agreement, one way-ANOVA with Bonferroni’s multiple comparison, and paired t tests were applied for statistics. RESULTS: All rabbits tolerated well the surgical procedure and subsequent cMRI sessions. The open-chest occlusion and close-chest reperfusion of the LCx, double suture method and bifunctional staining were successfully applied in all animals. The percentage MI volumes globally (n = 6) and by slice (n = 25) were 36.59% ± 13.68% and 32.88% ± 12.38% on iDE, and 35.41% ± 12.25% and 32.40% ± 12.34% on eDE. There were no significant differences for MI determination with excellent linear regression correspondence (rglobal = 0.89; rslice = 0.9) between iDE and eDE. The percentage AAR volumes globally (n = 6) and by slice (n = 25) were 44.82% ± 15.18% and 40.04% ± 13.64% with RIO-staining, and 44.74% ± 15.98% and 40.48% ± 13.26% by DR showing high correlation in linear regression analysis (rglobal = 0.99; rslice = 1.0). The mean differences of the two AAR measurements on Bland-Altman were almost zero, indicating RIO-staining and DR were essentially equivalent or inter-replaceable. The AAR was significantly larger than MI both globally and slice-by-slice (P < 0.01). After correction with the background and the blank heart without bifunctional staining (n = 3), the PDR for the AAR and normal myocardium was 32% ± 15% and 35.5% ± 35%, respectively, which is significantly different (P < 0.001), suggesting that blood perfusion to the AAR probably by collateral circulation was only less than 10% of that in the normal myocardium. CONCLUSION: The myocardial area at risk in ischemic heart disease could be accurately determined postmortem by this novel bifunctional staining, which may substantially contribute to translational cardiac imaging research. PMID:25237621
GIS Tools to Estimate Average Annual Daily Traffic
DOT National Transportation Integrated Search
2012-06-01
This project presents five tools that were created for a geographical information system to estimate Annual Average Daily : Traffic using linear regression. Three of the tools can be used to prepare spatial data for linear regression. One tool can be...
Jose F. Negron; Willis C. Schaupp; Kenneth E. Gibson; John Anhold; Dawn Hansen; Ralph Thier; Phil Mocettini
1999-01-01
Data collected from Douglas-fir stands infected by the Douglas-fir beetle in Wyoming, Montana, Idaho, and Utah, were used to develop models to estimate amount of mortality in terms of basal area killed. Models were built using stepwise linear regression and regression tree approaches. Linear regression models using initial Douglas-fir basal area were built for all...
Ling, Ru; Liu, Jiawang
2011-12-01
To construct prediction model for health workforce and hospital beds in county hospitals of Hunan by multiple linear regression. We surveyed 16 counties in Hunan with stratified random sampling according to uniform questionnaires,and multiple linear regression analysis with 20 quotas selected by literature view was done. Independent variables in the multiple linear regression model on medical personnels in county hospitals included the counties' urban residents' income, crude death rate, medical beds, business occupancy, professional equipment value, the number of devices valued above 10 000 yuan, fixed assets, long-term debt, medical income, medical expenses, outpatient and emergency visits, hospital visits, actual available bed days, and utilization rate of hospital beds. Independent variables in the multiple linear regression model on county hospital beds included the the population of aged 65 and above in the counties, disposable income of urban residents, medical personnel of medical institutions in county area, business occupancy, the total value of professional equipment, fixed assets, long-term debt, medical income, medical expenses, outpatient and emergency visits, hospital visits, actual available bed days, utilization rate of hospital beds, and length of hospitalization. The prediction model shows good explanatory and fitting, and may be used for short- and mid-term forecasting.
Fast estimation of diffusion tensors under Rician noise by the EM algorithm.
Liu, Jia; Gasbarra, Dario; Railavo, Juha
2016-01-15
Diffusion tensor imaging (DTI) is widely used to characterize, in vivo, the white matter of the central nerve system (CNS). This biological tissue contains much anatomic, structural and orientational information of fibers in human brain. Spectral data from the displacement distribution of water molecules located in the brain tissue are collected by a magnetic resonance scanner and acquired in the Fourier domain. After the Fourier inversion, the noise distribution is Gaussian in both real and imaginary parts and, as a consequence, the recorded magnitude data are corrupted by Rician noise. Statistical estimation of diffusion leads a non-linear regression problem. In this paper, we present a fast computational method for maximum likelihood estimation (MLE) of diffusivities under the Rician noise model based on the expectation maximization (EM) algorithm. By using data augmentation, we are able to transform a non-linear regression problem into the generalized linear modeling framework, reducing dramatically the computational cost. The Fisher-scoring method is used for achieving fast convergence of the tensor parameter. The new method is implemented and applied using both synthetic and real data in a wide range of b-amplitudes up to 14,000s/mm(2). Higher accuracy and precision of the Rician estimates are achieved compared with other log-normal based methods. In addition, we extend the maximum likelihood (ML) framework to the maximum a posteriori (MAP) estimation in DTI under the aforementioned scheme by specifying the priors. We will describe how close numerically are the estimators of model parameters obtained through MLE and MAP estimation. Copyright © 2015 Elsevier B.V. All rights reserved.
Language skills and phonological awareness in children with cochlear implants and normal hearing.
Soleymani, Zahra; Mahmoodabadi, Najmeh; Nouri, Mina Mohammadi
2016-04-01
Early auditory experience plays a major role in language acquisition. Linguistic and metalinguistic abilities of children aged 5-5.5 years with cochlear implants (CIs) were compared to age-matched children with normal hearing (NH) to investigate the effect of hearing on development of these two skills. Eighteen children with NH and 18 children with CIs took part in the study. The Test of Language Development-Primary, third edition, was used to assess language and metalinguistic skills by assessment of phonological awareness (PA). Language skills and PA were then compared between groups. Hierarchical linear regression was conducted to determine whether the language skills explained the unique variance in PA. There were significant differences between children with NH and those with CIs for language skills and PA (p≤0.001). All language skills (semantics, syntax, listening, spoken language, organizing, and speaking) were uniquely predictive of PA outcome in the CI children. Linear combinations of listening and semantics and listening, semantics, and syntax correlated significantly with PA. The results show that children with CIs may have trouble with language skills and PA. Listening, semantics, and syntax, among other skills, are significant indicators of the variance in PA for children with CIs. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Wilson, Natalie R.; Norman, Laura
2018-01-01
Watershed restoration efforts seek to rejuvenate vegetation, biological diversity, and land productivity at Cienega San Bernardino, an important wetland in southeastern Arizona and northern Sonora, Mexico. Rock detention and earthen berm structures were built on the Cienega San Bernardino over the course of four decades, beginning in 1984 and continuing to the present. Previous research findings show that restoration supports and even increases vegetation health despite ongoing drought conditions in this arid watershed. However, the extent of restoration impacts is still unknown despite qualitative observations of improvement in surrounding vegetation amount and vigor. We analyzed spatial and temporal trends in vegetation greenness and soil moisture by applying the normalized difference vegetation index (NDVI) and normalized difference infrared index (NDII) to one dry summer season Landsat path/row from 1984 to 2016. The study area was divided into zones and spectral data for each zone was analyzed and compared with precipitation record using statistical measures including linear regression, Mann– Kendall test, and linear correlation. NDVI and NDII performed differently due to the presence of continued grazing and the effects of grazing on canopy cover; NDVI was better able to track changes in vegetation in areas without grazing while NDII was better at tracking changes in areas with continued grazing. Restoration impacts display higher greenness and vegetation water content levels, greater increases in greenness and water content through time, and a decoupling of vegetation greenness and water content from spring precipitation when compared to control sites in nearby tributary and upland areas. Our results confirm the potential of erosion control structures to affect areas up to 5 km downstream of restoration sites over time and to affect 1 km upstream of the sites.
Effect of static foot posture on the dynamic stiffness of foot joints during walking.
Sanchis-Sales, E; Sancho-Bru, J L; Roda-Sales, A; Pascual-Huerta, J
2018-05-01
The static foot posture has been related to the development of lower limb injuries. This study aimed to investigate the dynamic stiffness of foot joints during gait in the sagittal plane to understand the role of the static foot posture in the development of injuries. Seventy healthy adult male subjects with different static postures, assessed by the Foot Posture Index (FPI) (30 normal, 20 highly pronated and 20 highly supinated), were recruited. Kinematic and kinetic data were recorded using an optical motion capture system and a pressure platform, and dynamic stiffness at the different stages of the stance was calculated from the slopes of the linear regression on the flexion moment-angle curves. The effect of foot type on dynamic stiffness and on ranges of motion and moments was analysed using ANOVAs and post-hoc tests, and linear correlation between dynamic stiffness and FPI was also tested. Highly pronated feet showed a significantly smaller range of motion at the ankle and metatarsophalangeal joints and also a larger range of moments at the metatarsophalangeal joint than highly supinated feet. Dynamic stiffness during propulsion was significantly greater at all foot joints for highly pronated feet, with positive significant correlations with the squared FPI. Highly supinated feet showed greater dynamic stiffness than normal feet, although to a lesser extent. Highly pronated feet during normal gait experienced the greatest decrease in the dorsiflexor moments during propulsion, normal feet being the most balanced regarding work generated and absorbed. Extreme static foot postures show greater dynamic stiffness during propulsion and greater absorbed work, which increases the risk of developing injuries. The data presented may be used when designing orthotics or prostheses, and also when planning surgery that modifies joint stiffness. Copyright © 2018 Elsevier B.V. All rights reserved.
Watanabe, Hiroyuki; Miyazaki, Hiroyasu
2006-01-01
Over- and/or under-correction of QT intervals for changes in heart rate may lead to misleading conclusions and/or masking the potential of a drug to prolong the QT interval. This study examines a nonparametric regression model (Loess Smoother) to adjust the QT interval for differences in heart rate, with an improved fitness over a wide range of heart rates. 240 sets of (QT, RR) observations collected from each of 8 conscious and non-treated beagle dogs were used as the materials for investigation. The fitness of the nonparametric regression model to the QT-RR relationship was compared with four models (individual linear regression, common linear regression, and Bazett's and Fridericia's correlation models) with reference to Akaike's Information Criterion (AIC). Residuals were visually assessed. The bias-corrected AIC of the nonparametric regression model was the best of the models examined in this study. Although the parametric models did not fit, the nonparametric regression model improved the fitting at both fast and slow heart rates. The nonparametric regression model is the more flexible method compared with the parametric method. The mathematical fit for linear regression models was unsatisfactory at both fast and slow heart rates, while the nonparametric regression model showed significant improvement at all heart rates in beagle dogs.
Linear regression analysis: part 14 of a series on evaluation of scientific publications.
Schneider, Astrid; Hommel, Gerhard; Blettner, Maria
2010-11-01
Regression analysis is an important statistical method for the analysis of medical data. It enables the identification and characterization of relationships among multiple factors. It also enables the identification of prognostically relevant risk factors and the calculation of risk scores for individual prognostication. This article is based on selected textbooks of statistics, a selective review of the literature, and our own experience. After a brief introduction of the uni- and multivariable regression models, illustrative examples are given to explain what the important considerations are before a regression analysis is performed, and how the results should be interpreted. The reader should then be able to judge whether the method has been used correctly and interpret the results appropriately. The performance and interpretation of linear regression analysis are subject to a variety of pitfalls, which are discussed here in detail. The reader is made aware of common errors of interpretation through practical examples. Both the opportunities for applying linear regression analysis and its limitations are presented.
Grajeda, Laura M; Ivanescu, Andrada; Saito, Mayuko; Crainiceanu, Ciprian; Jaganath, Devan; Gilman, Robert H; Crabtree, Jean E; Kelleher, Dermott; Cabrera, Lilia; Cama, Vitaliano; Checkley, William
2016-01-01
Childhood growth is a cornerstone of pediatric research. Statistical models need to consider individual trajectories to adequately describe growth outcomes. Specifically, well-defined longitudinal models are essential to characterize both population and subject-specific growth. Linear mixed-effect models with cubic regression splines can account for the nonlinearity of growth curves and provide reasonable estimators of population and subject-specific growth, velocity and acceleration. We provide a stepwise approach that builds from simple to complex models, and account for the intrinsic complexity of the data. We start with standard cubic splines regression models and build up to a model that includes subject-specific random intercepts and slopes and residual autocorrelation. We then compared cubic regression splines vis-à-vis linear piecewise splines, and with varying number of knots and positions. Statistical code is provided to ensure reproducibility and improve dissemination of methods. Models are applied to longitudinal height measurements in a cohort of 215 Peruvian children followed from birth until their fourth year of life. Unexplained variability, as measured by the variance of the regression model, was reduced from 7.34 when using ordinary least squares to 0.81 (p < 0.001) when using a linear mixed-effect models with random slopes and a first order continuous autoregressive error term. There was substantial heterogeneity in both the intercept (p < 0.001) and slopes (p < 0.001) of the individual growth trajectories. We also identified important serial correlation within the structure of the data (ρ = 0.66; 95 % CI 0.64 to 0.68; p < 0.001), which we modeled with a first order continuous autoregressive error term as evidenced by the variogram of the residuals and by a lack of association among residuals. The final model provides a parametric linear regression equation for both estimation and prediction of population- and individual-level growth in height. We show that cubic regression splines are superior to linear regression splines for the case of a small number of knots in both estimation and prediction with the full linear mixed effect model (AIC 19,352 vs. 19,598, respectively). While the regression parameters are more complex to interpret in the former, we argue that inference for any problem depends more on the estimated curve or differences in curves rather than the coefficients. Moreover, use of cubic regression splines provides biological meaningful growth velocity and acceleration curves despite increased complexity in coefficient interpretation. Through this stepwise approach, we provide a set of tools to model longitudinal childhood data for non-statisticians using linear mixed-effect models.
Prediction of monthly rainfall in Victoria, Australia: Clusterwise linear regression approach
NASA Astrophysics Data System (ADS)
Bagirov, Adil M.; Mahmood, Arshad; Barton, Andrew
2017-05-01
This paper develops the Clusterwise Linear Regression (CLR) technique for prediction of monthly rainfall. The CLR is a combination of clustering and regression techniques. It is formulated as an optimization problem and an incremental algorithm is designed to solve it. The algorithm is applied to predict monthly rainfall in Victoria, Australia using rainfall data with five input meteorological variables over the period of 1889-2014 from eight geographically diverse weather stations. The prediction performance of the CLR method is evaluated by comparing observed and predicted rainfall values using four measures of forecast accuracy. The proposed method is also compared with the CLR using the maximum likelihood framework by the expectation-maximization algorithm, multiple linear regression, artificial neural networks and the support vector machines for regression models using computational results. The results demonstrate that the proposed algorithm outperforms other methods in most locations.
Regression Model Term Selection for the Analysis of Strain-Gage Balance Calibration Data
NASA Technical Reports Server (NTRS)
Ulbrich, Norbert Manfred; Volden, Thomas R.
2010-01-01
The paper discusses the selection of regression model terms for the analysis of wind tunnel strain-gage balance calibration data. Different function class combinations are presented that may be used to analyze calibration data using either a non-iterative or an iterative method. The role of the intercept term in a regression model of calibration data is reviewed. In addition, useful algorithms and metrics originating from linear algebra and statistics are recommended that will help an analyst (i) to identify and avoid both linear and near-linear dependencies between regression model terms and (ii) to make sure that the selected regression model of the calibration data uses only statistically significant terms. Three different tests are suggested that may be used to objectively assess the predictive capability of the final regression model of the calibration data. These tests use both the original data points and regression model independent confirmation points. Finally, data from a simplified manual calibration of the Ames MK40 balance is used to illustrate the application of some of the metrics and tests to a realistic calibration data set.
NASA Astrophysics Data System (ADS)
Mendoza, Carlos S.; Safdar, Nabile; Myers, Emmarie; Kittisarapong, Tanakorn; Rogers, Gary F.; Linguraru, Marius George
2013-02-01
Craniosynostosis (premature fusion of skull sutures) is a severe condition present in one of every 2000 newborns. Metopic craniosynostosis, accounting for 20-27% of cases, is diagnosed qualitatively in terms of skull shape abnormality, a subjective call of the surgeon. In this paper we introduce a new quantitative diagnostic feature for metopic craniosynostosis derived optimally from shape analysis of CT scans of the skull. We built a robust shape analysis pipeline that is capable of obtaining local shape differences in comparison to normal anatomy. Spatial normalization using 7-degree-of-freedom registration of the base of the skull is followed by a novel bone labeling strategy based on graph-cuts according to labeling priors. The statistical shape model built from 94 normal subjects allows matching a patient's anatomy to its most similar normal subject. Subsequently, the computation of local malformations from a normal subject allows characterization of the points of maximum malformation on each of the frontal bones adjacent to the metopic suture, and on the suture itself. Our results show that the malformations at these locations vary significantly (p<0.001) between abnormal/normal subjects and that an accurate diagnosis can be achieved using linear regression from these automatic measurements with an area under the curve for the receiver operating characteristic of 0.97.
On the impact of relatedness on SNP association analysis.
Gross, Arnd; Tönjes, Anke; Scholz, Markus
2017-12-06
When testing for SNP (single nucleotide polymorphism) associations in related individuals, observations are not independent. Simple linear regression assuming independent normally distributed residuals results in an increased type I error and the power of the test is also affected in a more complicate manner. Inflation of type I error is often successfully corrected by genomic control. However, this reduces the power of the test when relatedness is of concern. In the present paper, we derive explicit formulae to investigate how heritability and strength of relatedness contribute to variance inflation of the effect estimate of the linear model. Further, we study the consequences of variance inflation on hypothesis testing and compare the results with those of genomic control correction. We apply the developed theory to the publicly available HapMap trio data (N=129), the Sorbs (a self-contained population with N=977 characterised by a cryptic relatedness structure) and synthetic family studies with different sample sizes (ranging from N=129 to N=999) and different degrees of relatedness. We derive explicit and easily to apply approximation formulae to estimate the impact of relatedness on the variance of the effect estimate of the linear regression model. Variance inflation increases with increasing heritability. Relatedness structure also impacts the degree of variance inflation as shown for example family structures. Variance inflation is smallest for HapMap trios, followed by a synthetic family study corresponding to the trio data but with larger sample size than HapMap. Next strongest inflation is observed for the Sorbs, and finally, for a synthetic family study with a more extreme relatedness structure but with similar sample size as the Sorbs. Type I error increases rapidly with increasing inflation. However, for smaller significance levels, power increases with increasing inflation while the opposite holds for larger significance levels. When genomic control is applied, type I error is preserved while power decreases rapidly with increasing variance inflation. Stronger relatedness as well as higher heritability result in increased variance of the effect estimate of simple linear regression analysis. While type I error rates are generally inflated, the behaviour of power is more complex since power can be increased or reduced in dependence on relatedness and the heritability of the phenotype. Genomic control cannot be recommended to deal with inflation due to relatedness. Although it preserves type I error, the loss in power can be considerable. We provide a simple formula for estimating variance inflation given the relatedness structure and the heritability of a trait of interest. As a rule of thumb, variance inflation below 1.05 does not require correction and simple linear regression analysis is still appropriate.
Scoring and staging systems using cox linear regression modeling and recursive partitioning.
Lee, J W; Um, S H; Lee, J B; Mun, J; Cho, H
2006-01-01
Scoring and staging systems are used to determine the order and class of data according to predictors. Systems used for medical data, such as the Child-Turcotte-Pugh scoring and staging systems for ordering and classifying patients with liver disease, are often derived strictly from physicians' experience and intuition. We construct objective and data-based scoring/staging systems using statistical methods. We consider Cox linear regression modeling and recursive partitioning techniques for censored survival data. In particular, to obtain a target number of stages we propose cross-validation and amalgamation algorithms. We also propose an algorithm for constructing scoring and staging systems by integrating local Cox linear regression models into recursive partitioning, so that we can retain the merits of both methods such as superior predictive accuracy, ease of use, and detection of interactions between predictors. The staging system construction algorithms are compared by cross-validation evaluation of real data. The data-based cross-validation comparison shows that Cox linear regression modeling is somewhat better than recursive partitioning when there are only continuous predictors, while recursive partitioning is better when there are significant categorical predictors. The proposed local Cox linear recursive partitioning has better predictive accuracy than Cox linear modeling and simple recursive partitioning. This study indicates that integrating local linear modeling into recursive partitioning can significantly improve prediction accuracy in constructing scoring and staging systems.
Scarneciu, Camelia C; Sangeorzan, Livia; Rus, Horatiu; Scarneciu, Vlad D; Varciu, Mihai S; Andreescu, Oana; Scarneciu, Ioan
2017-01-01
This study aimed at assessing the incidence of pulmonary hypertension (PH) at newly diagnosed hyperthyroid patients and at finding a simple model showing the complex functional relation between pulmonary hypertension in hyperthyroidism and the factors causing it. The 53 hyperthyroid patients (H-group) were evaluated mainly by using an echocardiographical method and compared with 35 euthyroid (E-group) and 25 healthy people (C-group). In order to identify the factors causing pulmonary hypertension the statistical method of comparing the values of arithmetical means is used. The functional relation between the two random variables (PAPs and each of the factors determining it within our research study) can be expressed by linear or non-linear function. By applying the linear regression method described by a first-degree equation the line of regression (linear model) has been determined; by applying the non-linear regression method described by a second degree equation, a parabola-type curve of regression (non-linear or polynomial model) has been determined. We made the comparison and the validation of these two models by calculating the determination coefficient (criterion 1), the comparison of residuals (criterion 2), application of AIC criterion (criterion 3) and use of F-test (criterion 4). From the H-group, 47% have pulmonary hypertension completely reversible when obtaining euthyroidism. The factors causing pulmonary hypertension were identified: previously known- level of free thyroxin, pulmonary vascular resistance, cardiac output; new factors identified in this study- pretreatment period, age, systolic blood pressure. According to the four criteria and to the clinical judgment, we consider that the polynomial model (graphically parabola- type) is better than the linear one. The better model showing the functional relation between the pulmonary hypertension in hyperthyroidism and the factors identified in this study is given by a polynomial equation of second degree where the parabola is its graphical representation.
As a fast and effective technique, the multiple linear regression (MLR) method has been widely used in modeling and prediction of beach bacteria concentrations. Among previous works on this subject, however, several issues were insufficiently or inconsistently addressed. Those is...
A simplified competition data analysis for radioligand specific activity determination.
Venturino, A; Rivera, E S; Bergoc, R M; Caro, R A
1990-01-01
Non-linear regression and two-step linear fit methods were developed to determine the actual specific activity of 125I-ovine prolactin by radioreceptor self-displacement analysis. The experimental results obtained by the different methods are superposable. The non-linear regression method is considered to be the most adequate procedure to calculate the specific activity, but if its software is not available, the other described methods are also suitable.
Height and Weight Estimation From Anthropometric Measurements Using Machine Learning Regressions
Fernandes, Bruno J. T.; Roque, Alexandre
2018-01-01
Height and weight are measurements explored to tracking nutritional diseases, energy expenditure, clinical conditions, drug dosages, and infusion rates. Many patients are not ambulant or may be unable to communicate, and a sequence of these factors may not allow accurate estimation or measurements; in those cases, it can be estimated approximately by anthropometric means. Different groups have proposed different linear or non-linear equations which coefficients are obtained by using single or multiple linear regressions. In this paper, we present a complete study of the application of different learning models to estimate height and weight from anthropometric measurements: support vector regression, Gaussian process, and artificial neural networks. The predicted values are significantly more accurate than that obtained with conventional linear regressions. In all the cases, the predictions are non-sensitive to ethnicity, and to gender, if more than two anthropometric parameters are analyzed. The learning model analysis creates new opportunities for anthropometric applications in industry, textile technology, security, and health care. PMID:29651366
NASA Astrophysics Data System (ADS)
Samhouri, M.; Al-Ghandoor, A.; Fouad, R. H.
2009-08-01
In this study two techniques, for modeling electricity consumption of the Jordanian industrial sector, are presented: (i) multivariate linear regression and (ii) neuro-fuzzy models. Electricity consumption is modeled as function of different variables such as number of establishments, number of employees, electricity tariff, prevailing fuel prices, production outputs, capacity utilizations, and structural effects. It was found that industrial production and capacity utilization are the most important variables that have significant effect on future electrical power demand. The results showed that both the multivariate linear regression and neuro-fuzzy models are generally comparable and can be used adequately to simulate industrial electricity consumption. However, comparison that is based on the square root average squared error of data suggests that the neuro-fuzzy model performs slightly better for future prediction of electricity consumption than the multivariate linear regression model. Such results are in full agreement with similar work, using different methods, for other countries.
Carvalho, Carlos; Gomes, Danielo G.; Agoulmine, Nazim; de Souza, José Neuman
2011-01-01
This paper proposes a method based on multivariate spatial and temporal correlation to improve prediction accuracy in data reduction for Wireless Sensor Networks (WSN). Prediction of data not sent to the sink node is a technique used to save energy in WSNs by reducing the amount of data traffic. However, it may not be very accurate. Simulations were made involving simple linear regression and multiple linear regression functions to assess the performance of the proposed method. The results show a higher correlation between gathered inputs when compared to time, which is an independent variable widely used for prediction and forecasting. Prediction accuracy is lower when simple linear regression is used, whereas multiple linear regression is the most accurate one. In addition to that, our proposal outperforms some current solutions by about 50% in humidity prediction and 21% in light prediction. To the best of our knowledge, we believe that we are probably the first to address prediction based on multivariate correlation for WSN data reduction. PMID:22346626
Using the Ridge Regression Procedures to Estimate the Multiple Linear Regression Coefficients
NASA Astrophysics Data System (ADS)
Gorgees, HazimMansoor; Mahdi, FatimahAssim
2018-05-01
This article concerns with comparing the performance of different types of ordinary ridge regression estimators that have been already proposed to estimate the regression parameters when the near exact linear relationships among the explanatory variables is presented. For this situations we employ the data obtained from tagi gas filling company during the period (2008-2010). The main result we reached is that the method based on the condition number performs better than other methods since it has smaller mean square error (MSE) than the other stated methods.
Devereux, Richard B; de Simone, Giovanni; Arnett, Donna K; Best, Lyle G; Boerwinkle, Eric; Howard, Barbara V; Kitzman, Dalane; Lee, Elisa T; Mosley, Thomas H; Weder, Alan; Roman, Mary J
2012-10-15
Nomograms to predict normal aortic root diameter for body surface area (BSA) in broad ranges of age have been widely used but are limited by lack of consideration of gender effects, jumps in upper limits of aortic diameter among age strata, and data from older teenagers. Sinus of Valsalva diameter was measured by American Society of Echocardiography convention in normal-weight, nonhypertensive, nondiabetic subjects ≥15 years old without aortic valve disease from clinical or population-based samples. Analyses of covariance and linear regression with assessment of residuals identified determinants and developed predictive models for normal aortic root diameter. In 1,207 apparently normal subjects ≥15 years old (54% women), aortic root diameter was 2.1 to 4.3 cm. Aortic root diameter was strongly related to BSA and height (r = 0.48 for the 2 comparisons), age (r = 0.36), and male gender (+2.7 mm adjusted for BSA and age, p <0.001 for all comparisons). Multivariable equations using age, gender, and BSA or height predicted aortic diameter strongly (R = 0.674 for the 2 comparisons, p <0.001) with minimal relation of residuals to age or body size: for BSA 2.423 + (age [years] × 0.009) + (BSA [square meters] × 0.461) - (gender [1 = man, 2 = woman] × 0.267), SEE 0.261 cm; for height 1.519 + (age [years] × 0.010) + (height [centimeters] × 0.010) - (gender [1 = man, 2 = woman] × 0.247), SEE 0.215 cm. In conclusion, aortic root diameter is larger in men and increases with body size and age. Regression models incorporating body size, age, and gender are applicable to adolescents and adults without limitations of previous nomograms. Copyright © 2012 Elsevier Inc. All rights reserved.
Weichenthal, Scott; Ryswyk, Keith Van; Goldstein, Alon; Bagg, Scott; Shekkarizfard, Maryam; Hatzopoulou, Marianne
2016-04-01
Existing evidence suggests that ambient ultrafine particles (UFPs) (<0.1µm) may contribute to acute cardiorespiratory morbidity. However, few studies have examined the long-term health effects of these pollutants owing in part to a need for exposure surfaces that can be applied in large population-based studies. To address this need, we developed a land use regression model for UFPs in Montreal, Canada using mobile monitoring data collected from 414 road segments during the summer and winter months between 2011 and 2012. Two different approaches were examined for model development including standard multivariable linear regression and a machine learning approach (kernel-based regularized least squares (KRLS)) that learns the functional form of covariate impacts on ambient UFP concentrations from the data. The final models included parameters for population density, ambient temperature and wind speed, land use parameters (park space and open space), length of local roads and rail, and estimated annual average NOx emissions from traffic. The final multivariable linear regression model explained 62% of the spatial variation in ambient UFP concentrations whereas the KRLS model explained 79% of the variance. The KRLS model performed slightly better than the linear regression model when evaluated using an external dataset (R(2)=0.58 vs. 0.55) or a cross-validation procedure (R(2)=0.67 vs. 0.60). In general, our findings suggest that the KRLS approach may offer modest improvements in predictive performance compared to standard multivariable linear regression models used to estimate spatial variations in ambient UFPs. However, differences in predictive performance were not statistically significant when evaluated using the cross-validation procedure. Crown Copyright © 2015. Published by Elsevier Inc. All rights reserved.
Alzheimer's Disease Detection by Pseudo Zernike Moment and Linear Regression Classification.
Wang, Shui-Hua; Du, Sidan; Zhang, Yin; Phillips, Preetha; Wu, Le-Nan; Chen, Xian-Qing; Zhang, Yu-Dong
2017-01-01
This study presents an improved method based on "Gorji et al. Neuroscience. 2015" by introducing a relatively new classifier-linear regression classification. Our method selects one axial slice from 3D brain image, and employed pseudo Zernike moment with maximum order of 15 to extract 256 features from each image. Finally, linear regression classification was harnessed as the classifier. The proposed approach obtains an accuracy of 97.51%, a sensitivity of 96.71%, and a specificity of 97.73%. Our method performs better than Gorji's approach and five other state-of-the-art approaches. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
Kwan, Johnny S H; Kung, Annie W C; Sham, Pak C
2011-09-01
Selective genotyping can increase power in quantitative trait association. One example of selective genotyping is two-tail extreme selection, but simple linear regression analysis gives a biased genetic effect estimate. Here, we present a simple correction for the bias.
NASA Astrophysics Data System (ADS)
Rooper, Christopher N.; Zimmermann, Mark; Prescott, Megan M.
2017-08-01
Deep-sea coral and sponge ecosystems are widespread throughout most of Alaska's marine waters, and are associated with many different species of fishes and invertebrates. These ecosystems are vulnerable to the effects of commercial fishing activities and climate change. We compared four commonly used species distribution models (general linear models, generalized additive models, boosted regression trees and random forest models) and an ensemble model to predict the presence or absence and abundance of six groups of benthic invertebrate taxa in the Gulf of Alaska. All four model types performed adequately on training data for predicting presence and absence, with regression forest models having the best overall performance measured by the area under the receiver-operating-curve (AUC). The models also performed well on the test data for presence and absence with average AUCs ranging from 0.66 to 0.82. For the test data, ensemble models performed the best. For abundance data, there was an obvious demarcation in performance between the two regression-based methods (general linear models and generalized additive models), and the tree-based models. The boosted regression tree and random forest models out-performed the other models by a wide margin on both the training and testing data. However, there was a significant drop-off in performance for all models of invertebrate abundance ( 50%) when moving from the training data to the testing data. Ensemble model performance was between the tree-based and regression-based methods. The maps of predictions from the models for both presence and abundance agreed very well across model types, with an increase in variability in predictions for the abundance data. We conclude that where data conforms well to the modeled distribution (such as the presence-absence data and binomial distribution in this study), the four types of models will provide similar results, although the regression-type models may be more consistent with biological theory. For data with highly zero-inflated distributions and non-normal distributions such as the abundance data from this study, the tree-based methods performed better. Ensemble models that averaged predictions across the four model types, performed better than the GLM or GAM models but slightly poorer than the tree-based methods, suggesting ensemble models might be more robust to overfitting than tree methods, while mitigating some of the disadvantages in predictive performance of regression methods.
Queiroz, Valterlinda A O; Assis, Ana Marlúcia O; Pinheiro, Sandra Maria C; Ribeiro, Hugo C Ribeiro
2012-01-01
To investigate covariates that could affect the variation in mean length/age z scores in the first year of life of children born full term with normal birth weight. This was a prospective study of a cohort of mother-infant pairs recruited at public maternity units in two municipalities in the Brazilian state of Bahia, from March 2005 to October 2006. This paper reports the results for linear growth of 489 children who were followed-up for the first 12 months of their lives. A mixed-effect regression model was used to investigate the influence of covariates of mean length/age z score during the first year of life. The multivariate mixed effect analysis indicated that mothers not cohabiting with a partner (beta = 0.2347; p = 0.004) and increased duration of exclusive breastfeeding (beta = 0.0031; p < 0.001) had a positive impact, whereas mother's height less than 150 cm (beta = -0.4393; p < 0.001), birth weight of 2,500-2,999 g (beta = -0.8084; p < 0.001) and anemia in the child (beta = -0.0875; p < 0.001) all had a negative impact on the variation in estimated length/age z score. Therefore, the results of this study indicate that short maternal stature, birth weight < 3,000 g and anemia in the infant had a negative effect on linear growth during the first year of life, whereas longer duration of exclusive breastfeeding and mothers who did not cohabit with a partner had a positive effect.
Quantifying progression and regression of thrombotic risk in experimental atherosclerosis.
Palekar, Rohun U; Jallouk, Andrew P; Goette, Matthew J; Chen, Junjie; Myerson, Jacob W; Allen, John S; Akk, Antonina; Yang, Lihua; Tu, Yizheng; Miller, Mark J; Pham, Christine T N; Wickline, Samuel A; Pan, Hua
2015-07-01
Currently, there are no generally applicable noninvasive methods for defining the relationship between atherosclerotic vascular damage and risk of focal thrombosis. Herein, we demonstrate methods to delineate the progression and regression of vascular damage in response to an atherogenic diet by quantifying the in vivo accumulation of semipermeable 200-300 nm perfluorocarbon core nanoparticles (PFC-NP) in ApoE null mouse plaques with [(19)F] magnetic resonance spectroscopy (MRS). Permeability to PFC-NP remained minimal until 12 weeks on diet, then increased rapidly following 12 weeks, but regressed to baseline within 8 weeks after diet normalization. Markedly accelerated clotting (53.3% decrease in clotting time) was observed in carotid artery preparations of fat-fed mice subjected to photochemical injury as defined by the time to flow cessation. For all mice on and off diet, an inverse linear relationship was observed between the permeability to PFC-NP and accelerated thrombosis (P = 0.02). Translational feasibility for quantifying plaque permeability and vascular damage in vivo was demonstrated with clinical 3 T MRI of PFC-NP accumulating in plaques of atherosclerotic rabbits. These observations suggest that excessive permeability to PFC-NP may indicate prothrombotic risk in damaged atherosclerotic vasculature, which resolves within weeks after dietary therapy. © FASEB.
Regression analysis of sparse asynchronous longitudinal data
Cao, Hongyuan; Zeng, Donglin; Fine, Jason P.
2015-01-01
Summary We consider estimation of regression models for sparse asynchronous longitudinal observations, where time-dependent responses and covariates are observed intermittently within subjects. Unlike with synchronous data, where the response and covariates are observed at the same time point, with asynchronous data, the observation times are mismatched. Simple kernel-weighted estimating equations are proposed for generalized linear models with either time invariant or time-dependent coefficients under smoothness assumptions for the covariate processes which are similar to those for synchronous data. For models with either time invariant or time-dependent coefficients, the estimators are consistent and asymptotically normal but converge at slower rates than those achieved with synchronous data. Simulation studies evidence that the methods perform well with realistic sample sizes and may be superior to a naive application of methods for synchronous data based on an ad hoc last value carried forward approach. The practical utility of the methods is illustrated on data from a study on human immunodeficiency virus. PMID:26568699
2013-01-01
application of the Hammett equation with the constants rph in the chemistry of organophosphorus compounds, Russ. Chem. Rev. 38 (1969) 795–811. [13...of oximes and OP compounds and the ability of oximes to reactivate OP- inhibited AChE. Multiple linear regression equations were analyzed using...phosphonate pairs, 21 oxime/ phosphoramidate pairs and 12 oxime/phosphate pairs. The best linear regression equation resulting from multiple regression anal
Molenaar, Dylan; Bolsinova, Maria
2017-05-01
In generalized linear modelling of responses and response times, the observed response time variables are commonly transformed to make their distribution approximately normal. A normal distribution for the transformed response times is desirable as it justifies the linearity and homoscedasticity assumptions in the underlying linear model. Past research has, however, shown that the transformed response times are not always normal. Models have been developed to accommodate this violation. In the present study, we propose a modelling approach for responses and response times to test and model non-normality in the transformed response times. Most importantly, we distinguish between non-normality due to heteroscedastic residual variances, and non-normality due to a skewed speed factor. In a simulation study, we establish parameter recovery and the power to separate both effects. In addition, we apply the model to a real data set. © 2017 The Authors. British Journal of Mathematical and Statistical Psychology published by John Wiley & Sons Ltd on behalf of British Psychological Society.
NASA Astrophysics Data System (ADS)
Shih, C. Y.; Tsuei, Y. G.; Allemang, R. J.; Brown, D. L.
1988-10-01
A method of using the matrix Auto-Regressive Moving Average (ARMA) model in the Laplace domain for multiple-reference global parameter identification is presented. This method is particularly applicable to the area of modal analysis where high modal density exists. The method is also applicable when multiple reference frequency response functions are used to characterise linear systems. In order to facilitate the mathematical solution, the Forsythe orthogonal polynomial is used to reduce the ill-conditioning of the formulated equations and to decouple the normal matrix into two reduced matrix blocks. A Complex Mode Indicator Function (CMIF) is introduced, which can be used to determine the proper order of the rational polynomials.
Hsu, Ruey-Fen; Ho, Chi-Kung; Lu, Sheng-Nan; Chen, Shun-Sheng
2010-10-01
An objective investigation is needed to verify the existence and severity of hearing impairments resulting from work-related, noise-induced hearing loss in arbitration of medicolegal aspects. We investigated the accuracy of multiple-frequency auditory steady-state responses (Mf-ASSRs) between subjects with sensorineural hearing loss (SNHL) with and without occupational noise exposure. Cross-sectional study. Tertiary referral medical centre. Pure-tone audiometry and Mf-ASSRs were recorded in 88 subjects (34 patients had occupational noise-induced hearing loss [NIHL], 36 patients had SNHL without noise exposure, and 18 volunteers were normal controls). Inter- and intragroup comparisons were made. A predicting equation was derived using multiple linear regression analysis. ASSRs and pure-tone thresholds (PTTs) showed a strong correlation for all subjects (r = .77 ≈ .94). The relationship is demonstrated by the equationThe differences between the ASSR and PTT were significantly higher for the NIHL group than for the subjects with non-noise-induced SNHL (p < .001). Mf-ASSR is a promising tool for objectively evaluating hearing thresholds. Predictive value may be lower in subjects with occupational hearing loss. Regardless of carrier frequencies, the severity of hearing loss affects the steady-state response. Moreover, the ASSR may assist in detecting noise-induced injury of the auditory pathway. A multiple linear regression equation to accurately predict thresholds was shown that takes into consideration all effect factors.
Wear, Keith A; Nagaraja, Srinidhi; Dreher, Maureen L; Sadoughi, Saghi; Zhu, Shan; Keaveny, Tony M
2017-10-01
Clinical bone sonometers applied at the calcaneus measure broadband ultrasound attenuation and speed of sound. However, the relation of ultrasound measurements to bone strength is not well-characterized. Addressing this issue, we assessed the extent to which ultrasonic measurements convey in vitro mechanical properties in 25 human calcaneal cancellous bone specimens (approximately 2×4×2cm). Normalized broadband ultrasound attenuation, speed of sound, and broadband ultrasound backscatter were measured with 500kHz transducers. To assess mechanical properties, non-linear finite element analysis, based on micro-computed tomography images (34-micron cubic voxel), was used to estimate apparent elastic modulus, overall specimen stiffness, and apparent yield stress, with models typically having approximately 25-30 million elements. We found that ultrasound parameters were correlated with mechanical properties with R=0.70-0.82 (p<0.001). Multiple regression analysis indicated that ultrasound measurements provide additional information regarding mechanical properties beyond that provided by bone quantity alone (p≤0.05). Adding ultrasound variables to linear regression models based on bone quantity improved adjusted squared correlation coefficients from 0.65 to 0.77 (stiffness), 0.76 to 0.81 (apparent modulus), and 0.67 to 0.73 (yield stress). These results indicate that ultrasound can provide complementary (to bone quantity) information regarding mechanical behavior of cancellous bone. Published by Elsevier Inc.
A generalized multivariate regression model for modelling ocean wave heights
NASA Astrophysics Data System (ADS)
Wang, X. L.; Feng, Y.; Swail, V. R.
2012-04-01
In this study, a generalized multivariate linear regression model is developed to represent the relationship between 6-hourly ocean significant wave heights (Hs) and the corresponding 6-hourly mean sea level pressure (MSLP) fields. The model is calibrated using the ERA-Interim reanalysis of Hs and MSLP fields for 1981-2000, and is validated using the ERA-Interim reanalysis for 2001-2010 and ERA40 reanalysis of Hs and MSLP for 1958-2001. The performance of the fitted model is evaluated in terms of Pierce skill score, frequency bias index, and correlation skill score. Being not normally distributed, wave heights are subjected to a data adaptive Box-Cox transformation before being used in the model fitting. Also, since 6-hourly data are being modelled, lag-1 autocorrelation must be and is accounted for. The models with and without Box-Cox transformation, and with and without accounting for autocorrelation, are inter-compared in terms of their prediction skills. The fitted MSLP-Hs relationship is then used to reconstruct historical wave height climate from the 6-hourly MSLP fields taken from the Twentieth Century Reanalysis (20CR, Compo et al. 2011), and to project possible future wave height climates using CMIP5 model simulations of MSLP fields. The reconstructed and projected wave heights, both seasonal means and maxima, are subject to a trend analysis that allows for non-linear (polynomial) trends.
Pouchot, Jacques; Kherani, Raheem B.; Brant, Rollin; Lacaille, Diane; Lehman, Allen J.; Ensworth, Stephanie; Kopec, Jacek; Esdaile, John M.; Liang, Matthew H.
2008-01-01
Objective To estimate the minimal clinically important difference (MCID) of seven measures of fatigue in rheumatoid arthritis. Study Design and Setting A cross-sectional study design based on inter-individual comparisons was used. Six to eight subjects participated in a single meeting and completed seven fatigue questionnaires (nine sessions were organized and 61 subjects participated). After completion of the questionnaires, the subjects had five one-on-one 10-minute conversations with different people in the group to discuss their fatigue. After each conversation, each patient compared their fatigue to their conversational partner’s on a global rating. Ratings were compared to the scores of the fatigue measures to estimate the MCID. Both non-parametric and linear regression analyses were used. Results Non-parametric estimates for the MCID relative to “little more fatigue” tended to be smaller than those for “little less fatigue”. The global MCIDs estimated by linear regression were: FSS 20.2, VT 14.8, MAF 18.7, MFI 16.6, FACIT–F 15.9, CFS 9.9, RS 19.7, for normalized scores (0 to 100). The standardized MCIDs for the seven measures were roughly similar (0.67 to 0.76). Conclusion These estimates of MCID will help to interpret changes observed in a fatigue score and will be critical in estimating sample size requirements. PMID:18359189
A Skew-Normal Mixture Regression Model
ERIC Educational Resources Information Center
Liu, Min; Lin, Tsung-I
2014-01-01
A challenge associated with traditional mixture regression models (MRMs), which rest on the assumption of normally distributed errors, is determining the number of unobserved groups. Specifically, even slight deviations from normality can lead to the detection of spurious classes. The current work aims to (a) examine how sensitive the commonly…
Specialization Agreements in the Council for Mutual Economic Assistance
1988-02-01
proportions to stabilize variance (S. Weisberg, Applied Linear Regression , 2nd ed., John Wiley & Sons, New York, 1985, p. 134). If the dependent...27, 1986, p. 3. Weisberg, S., Applied Linear Regression , 2nd ed., John Wiley & Sons, New York, 1985, p. 134. Wiles, P. J., Communist International
Radio Propagation Prediction Software for Complex Mixed Path Physical Channels
2006-08-14
63 4.4.6. Applied Linear Regression Analysis in the Frequency Range 1-50 MHz 69 4.4.7. Projected Scaling to...4.4.6. Applied Linear Regression Analysis in the Frequency Range 1-50 MHz In order to construct a comprehensive numerical algorithm capable of
Due to the complexity of the processes contributing to beach bacteria concentrations, many researchers rely on statistical modeling, among which multiple linear regression (MLR) modeling is most widely used. Despite its ease of use and interpretation, there may be time dependence...
USING LINEAR AND POLYNOMIAL MODELS TO EXAMINE THE ENVIRONMENTAL STABILITY OF VIRUSES
The article presents the development of model equations for describing the fate of viral infectivity in environmental samples. Most of the models were based upon the use of a two-step linear regression approach. The first step employs regression of log base 10 transformed viral t...
Validation and application of single breath cardiac output determinations in man
NASA Technical Reports Server (NTRS)
Loeppky, J. A.; Fletcher, E. R.; Myhre, L. G.; Luft, U. C.
1986-01-01
The results of a procedure for estimating cardiac output by a single-breath technique (Qsb), obtained in healthy males during supine rest and during exercise on a bicycle ergometer, were compared with the results on cardiac output obtained by the direct Fick method (QF). The single breath maneuver consisted of a slow exhalation to near residual volume following an inspiration somewhat deeper than normal. The Qsb calculations incorporated an equation of the CO2 dissociation curve and a 'moving spline' sequential curve-fitting technique to calculate the instantaneous R from points on the original expirogram. The resulting linear regression equation indicated a 24-percent underestimation of QF by the Qsb technique. After applying a correction, the Qsb-QF relationship was improved. A subsequent study during upright rest and exercise to 80 percent of VO2(max) in 6 subjects indicated a close linear relationship between Qsb and VO2 for all 95 values obtained, with slope and intercept close to those in published studies in which invasive cardiac output measurements were used.
Huikang Wang; Luzheng Bi; Teng Teng
2017-07-01
This paper proposes a novel method of electroencephalography (EEG)-based driver emergency braking intention detection system for brain-controlled driving considering one electrode falling-off. First, whether one electrode falls off is discriminated based on EEG potentials. Then, the missing signals are estimated by using the signals collected from other channels based on multivariate linear regression. Finally, a linear decoder is applied to classify driver intentions. Experimental results show that the falling-off discrimination accuracy is 99.63% on average and the correlation coefficient and root mean squared error (RMSE) between the estimated and experimental data are 0.90 and 11.43 μV, respectively, on average. Given one electrode falls off, the system accuracy of the proposed intention prediction method is significantly higher than that of the original method (95.12% VS 79.11%) and is close to that (95.95%) of the original system under normal situations (i. e., no electrode falling-off).
NASA Astrophysics Data System (ADS)
Tautz-Weinert, J.; Watson, S. J.
2016-09-01
Effective condition monitoring techniques for wind turbines are needed to improve maintenance processes and reduce operational costs. Normal behaviour modelling of temperatures with information from other sensors can help to detect wear processes in drive trains. In a case study, modelling of bearing and generator temperatures is investigated with operational data from the SCADA systems of more than 100 turbines. The focus is here on automated training and testing on a farm level to enable an on-line system, which will detect failures without human interpretation. Modelling based on linear combinations, artificial neural networks, adaptive neuro-fuzzy inference systems, support vector machines and Gaussian process regression is compared. The selection of suitable modelling inputs is discussed with cross-correlation analyses and a sensitivity study, which reveals that the investigated modelling techniques react in different ways to an increased number of inputs. The case study highlights advantages of modelling with linear combinations and artificial neural networks in a feedforward configuration.
Finding Bayesian Optimal Designs for Nonlinear Models: A Semidefinite Programming-Based Approach.
Duarte, Belmiro P M; Wong, Weng Kee
2015-08-01
This paper uses semidefinite programming (SDP) to construct Bayesian optimal design for nonlinear regression models. The setup here extends the formulation of the optimal designs problem as an SDP problem from linear to nonlinear models. Gaussian quadrature formulas (GQF) are used to compute the expectation in the Bayesian design criterion, such as D-, A- or E-optimality. As an illustrative example, we demonstrate the approach using the power-logistic model and compare results in the literature. Additionally, we investigate how the optimal design is impacted by different discretising schemes for the design space, different amounts of uncertainty in the parameter values, different choices of GQF and different prior distributions for the vector of model parameters, including normal priors with and without correlated components. Further applications to find Bayesian D-optimal designs with two regressors for a logistic model and a two-variable generalised linear model with a gamma distributed response are discussed, and some limitations of our approach are noted.
Finding Bayesian Optimal Designs for Nonlinear Models: A Semidefinite Programming-Based Approach
Duarte, Belmiro P. M.; Wong, Weng Kee
2014-01-01
Summary This paper uses semidefinite programming (SDP) to construct Bayesian optimal design for nonlinear regression models. The setup here extends the formulation of the optimal designs problem as an SDP problem from linear to nonlinear models. Gaussian quadrature formulas (GQF) are used to compute the expectation in the Bayesian design criterion, such as D-, A- or E-optimality. As an illustrative example, we demonstrate the approach using the power-logistic model and compare results in the literature. Additionally, we investigate how the optimal design is impacted by different discretising schemes for the design space, different amounts of uncertainty in the parameter values, different choices of GQF and different prior distributions for the vector of model parameters, including normal priors with and without correlated components. Further applications to find Bayesian D-optimal designs with two regressors for a logistic model and a two-variable generalised linear model with a gamma distributed response are discussed, and some limitations of our approach are noted. PMID:26512159
Simple and multiple linear regression: sample size considerations.
Hanley, James A
2016-11-01
The suggested "two subjects per variable" (2SPV) rule of thumb in the Austin and Steyerberg article is a chance to bring out some long-established and quite intuitive sample size considerations for both simple and multiple linear regression. This article distinguishes two of the major uses of regression models that imply very different sample size considerations, neither served well by the 2SPV rule. The first is etiological research, which contrasts mean Y levels at differing "exposure" (X) values and thus tends to focus on a single regression coefficient, possibly adjusted for confounders. The second research genre guides clinical practice. It addresses Y levels for individuals with different covariate patterns or "profiles." It focuses on the profile-specific (mean) Y levels themselves, estimating them via linear compounds of regression coefficients and covariates. By drawing on long-established closed-form variance formulae that lie beneath the standard errors in multiple regression, and by rearranging them for heuristic purposes, one arrives at quite intuitive sample size considerations for both research genres. Copyright © 2016 Elsevier Inc. All rights reserved.
Jiang, Feng; Han, Ji-zhong
2018-01-01
Cross-domain collaborative filtering (CDCF) solves the sparsity problem by transferring rating knowledge from auxiliary domains. Obviously, different auxiliary domains have different importance to the target domain. However, previous works cannot evaluate effectively the significance of different auxiliary domains. To overcome this drawback, we propose a cross-domain collaborative filtering algorithm based on Feature Construction and Locally Weighted Linear Regression (FCLWLR). We first construct features in different domains and use these features to represent different auxiliary domains. Thus the weight computation across different domains can be converted as the weight computation across different features. Then we combine the features in the target domain and in the auxiliary domains together and convert the cross-domain recommendation problem into a regression problem. Finally, we employ a Locally Weighted Linear Regression (LWLR) model to solve the regression problem. As LWLR is a nonparametric regression method, it can effectively avoid underfitting or overfitting problem occurring in parametric regression methods. We conduct extensive experiments to show that the proposed FCLWLR algorithm is effective in addressing the data sparsity problem by transferring the useful knowledge from the auxiliary domains, as compared to many state-of-the-art single-domain or cross-domain CF methods. PMID:29623088
Yu, Xu; Lin, Jun-Yu; Jiang, Feng; Du, Jun-Wei; Han, Ji-Zhong
2018-01-01
Cross-domain collaborative filtering (CDCF) solves the sparsity problem by transferring rating knowledge from auxiliary domains. Obviously, different auxiliary domains have different importance to the target domain. However, previous works cannot evaluate effectively the significance of different auxiliary domains. To overcome this drawback, we propose a cross-domain collaborative filtering algorithm based on Feature Construction and Locally Weighted Linear Regression (FCLWLR). We first construct features in different domains and use these features to represent different auxiliary domains. Thus the weight computation across different domains can be converted as the weight computation across different features. Then we combine the features in the target domain and in the auxiliary domains together and convert the cross-domain recommendation problem into a regression problem. Finally, we employ a Locally Weighted Linear Regression (LWLR) model to solve the regression problem. As LWLR is a nonparametric regression method, it can effectively avoid underfitting or overfitting problem occurring in parametric regression methods. We conduct extensive experiments to show that the proposed FCLWLR algorithm is effective in addressing the data sparsity problem by transferring the useful knowledge from the auxiliary domains, as compared to many state-of-the-art single-domain or cross-domain CF methods.
Age-related apparent diffusion coefficient changes in the normal brain.
Watanabe, Memi; Sakai, Osamu; Ozonoff, Al; Kussman, Steven; Jara, Hernán
2013-02-01
To measure the mean diffusional age-related changes of the brain over the full human life span by using diffusion-weighted spin-echo single-shot echo-planar magnetic resonance (MR) imaging and sequential whole-brain apparent diffusion coefficient (ADC) histogram analysis and, secondarily, to build mathematical models of these normal age-related changes throughout human life. After obtaining institutional review board approval, a HIPAA-compliant retrospective search was conducted for brain MR imaging studies performed in 2007 for various clinical indications. Informed consent was waived. The brain data of 414 healthy subjects (189 males and 225 females; mean age, 33.7 years; age range, 2 days to 89.3 years) were obtained with diffusion-weighted spin-echo single-shot echo-planar MR imaging. ADC histograms of the whole brain were generated. ADC peak values, histogram widths, and intracranial volumes were plotted against age, and model parameters were estimated by using nonlinear regression. Four different stages were identified for aging changes in ADC peak values, as characterized by specific mathematical terms: There were age-associated exponential decays for the maturation period and the development period, a constant term for adulthood, and a linear increase for the senescence period. The age dependency of ADC peak value was simulated by using four-term six-coefficient function, including biexponential and linear terms. This model fit the data very closely (R(2) = 0.91). Brain diffusivity as a whole demonstrated age-related changes through four distinct periods of life. These results could contribute to establishing an ADC baseline of the normal brain, covering the full human life span.
ERIC Educational Resources Information Center
Bulcock, J. W.; And Others
Advantages of normalization regression estimation over ridge regression estimation are demonstrated by reference to Bloom's model of school learning. Theoretical concern centered on the structure of scholastic achievement at grade 10 in Canadian high schools. Data on 886 students were randomly sampled from the Carnegie Human Resources Data Bank.…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Horton, Megan K., E-mail: megan.horton@mssm.edu; Blount, Benjamin C.; Valentin-Blasini, Liza
Background: Adequate maternal thyroid function during pregnancy is necessary for normal fetal brain development, making pregnancy a critical window of vulnerability to thyroid disrupting insults. Sodium/iodide symporter (NIS) inhibitors, namely perchlorate, nitrate, and thiocyanate, have been shown individually to competitively inhibit uptake of iodine by the thyroid. Several epidemiologic studies examined the association between these individual exposures and thyroid function. Few studies have examined the effect of this chemical mixture on thyroid function during pregnancy Objectives: We examined the cross sectional association between urinary perchlorate, thiocyanate and nitrate concentrations and thyroid function among healthy pregnant women living in New Yorkmore » City using weighted quantile sum (WQS) regression. Methods: We measured thyroid stimulating hormone (TSH) and free thyroxine (FreeT4) in blood samples; perchlorate, thiocyanate, nitrate and iodide in urine samples collected from 284 pregnant women at 12 (±2.8) weeks gestation. We examined associations between urinary analyte concentrations and TSH or FreeT4 using linear regression or WQS adjusting for gestational age, urinary iodide and creatinine. Results: Individual analyte concentrations in urine were significantly correlated (Spearman's r 0.4–0.5, p<0.001). Linear regression analyses did not suggest associations between individual concentrations and thyroid function. The WQS revealed a significant positive association between the weighted sum of urinary concentrations of the three analytes and increased TSH. Perchlorate had the largest weight in the index, indicating the largest contribution to the WQS. Conclusions: Co-exposure to perchlorate, nitrate and thiocyanate may alter maternal thyroid function, specifically TSH, during pregnancy. - Highlights: • Perchlorate, nitrate, thiocyanate and iodide measured in maternal urine. • Thyroid function (TSH and Free T4) measured in maternal blood. • Weighted quantile sum (WQS) regression examined complex mixture effect. • WQS identified an inverse association between the exposure mixture and maternal TSH. • Perchlorate indicated as the ‘bad actor’ of the mixture.« less
Esserman, Denise A.; Moore, Charity G.; Roth, Mary T.
2009-01-01
Older community dwelling adults often take multiple medications for numerous chronic diseases. Non-adherence to these medications can have a large public health impact. Therefore, the measurement and modeling of medication adherence in the setting of polypharmacy is an important area of research. We apply a variety of different modeling techniques (standard linear regression; weighted linear regression; adjusted linear regression; naïve logistic regression; beta-binomial (BB) regression; generalized estimating equations (GEE)) to binary medication adherence data from a study in a North Carolina based population of older adults, where each medication an individual was taking was classified as adherent or non-adherent. In addition, through simulation we compare these different methods based on Type I error rates, bias, power, empirical 95% coverage, and goodness of fit. We find that estimation and inference using GEE is robust to a wide variety of scenarios and we recommend using this in the setting of polypharmacy when adherence is dichotomously measured for multiple medications per person. PMID:20414358
Genetic Programming Transforms in Linear Regression Situations
NASA Astrophysics Data System (ADS)
Castillo, Flor; Kordon, Arthur; Villa, Carlos
The chapter summarizes the use of Genetic Programming (GP) inMultiple Linear Regression (MLR) to address multicollinearity and Lack of Fit (LOF). The basis of the proposed method is applying appropriate input transforms (model respecification) that deal with these issues while preserving the information content of the original variables. The transforms are selected from symbolic regression models with optimal trade-off between accuracy of prediction and expressional complexity, generated by multiobjective Pareto-front GP. The chapter includes a comparative study of the GP-generated transforms with Ridge Regression, a variant of ordinary Multiple Linear Regression, which has been a useful and commonly employed approach for reducing multicollinearity. The advantages of GP-generated model respecification are clearly defined and demonstrated. Some recommendations for transforms selection are given as well. The application benefits of the proposed approach are illustrated with a real industrial application in one of the broadest empirical modeling areas in manufacturing - robust inferential sensors. The chapter contributes to increasing the awareness of the potential of GP in statistical model building by MLR.
Naval Research Logistics Quarterly. Volume 28. Number 3,
1981-09-01
denotes component-wise maximum. f has antone (isotone) differences on C x D if for cl < c2 and d, < d2, NAVAL RESEARCH LOGISTICS QUARTERLY VOL. 28...or negative correlations and linear or nonlinear regressions. Given are the mo- ments to order two and, for special cases, (he regression function and...data sets. We designate this bnb distribution as G - B - N(a, 0, v). The distribution admits only of positive correlation and linear regressions
Automating approximate Bayesian computation by local linear regression.
Thornton, Kevin R
2009-07-07
In several biological contexts, parameter inference often relies on computationally-intensive techniques. "Approximate Bayesian Computation", or ABC, methods based on summary statistics have become increasingly popular. A particular flavor of ABC based on using a linear regression to approximate the posterior distribution of the parameters, conditional on the summary statistics, is computationally appealing, yet no standalone tool exists to automate the procedure. Here, I describe a program to implement the method. The software package ABCreg implements the local linear-regression approach to ABC. The advantages are: 1. The code is standalone, and fully-documented. 2. The program will automatically process multiple data sets, and create unique output files for each (which may be processed immediately in R), facilitating the testing of inference procedures on simulated data, or the analysis of multiple data sets. 3. The program implements two different transformation methods for the regression step. 4. Analysis options are controlled on the command line by the user, and the program is designed to output warnings for cases where the regression fails. 5. The program does not depend on any particular simulation machinery (coalescent, forward-time, etc.), and therefore is a general tool for processing the results from any simulation. 6. The code is open-source, and modular.Examples of applying the software to empirical data from Drosophila melanogaster, and testing the procedure on simulated data, are shown. In practice, the ABCreg simplifies implementing ABC based on local-linear regression.
NASA Astrophysics Data System (ADS)
Jakubowski, J.; Stypulkowski, J. B.; Bernardeau, F. G.
2017-12-01
The first phase of the Abu Hamour drainage and storm tunnel was completed in early 2017. The 9.5 km long, 3.7 m diameter tunnel was excavated with two Earth Pressure Balance (EPB) Tunnel Boring Machines from Herrenknecht. TBM operation processes were monitored and recorded by Data Acquisition and Evaluation System. The authors coupled collected TBM drive data with available information on rock mass properties, cleansed, completed with secondary variables and aggregated by weeks and shifts. Correlations and descriptive statistics charts were examined. Multivariate Linear Regression and CART regression tree models linking TBM penetration rate (PR), penetration per revolution (PPR) and field penetration index (FPI) with TBM operational and geotechnical characteristics were performed for the conditions of the weak/soft rock of Doha. Both regression methods are interpretable and the data were screened with different computational approaches allowing enriched insight. The primary goal of the analysis was to investigate empirical relations between multiple explanatory and responding variables, to search for best subsets of explanatory variables and to evaluate the strength of linear and non-linear relations. For each of the penetration indices, a predictive model coupling both regression methods was built and validated. The resultant models appeared to be stronger than constituent ones and indicated an opportunity for more accurate and robust TBM performance predictions.
Bell, Lana M; Byrne, Sue; Thompson, Alisha; Ratnam, Nirubasini; Blair, Eve; Bulsara, Max; Jones, Timothy W; Davis, Elizabeth A
2007-02-01
Overweight/obesity in children is increasing. Incidence data for medical complications use arbitrary cutoff values for categories of overweight and obesity. Continuous relationships are seldom reported. The objective of this study is to report relationships of child body mass index (BMI) z-score as a continuous variable with the medical complications of overweight. This study is a part of the larger, prospective cohort Growth and Development Study. Children were recruited from the community through randomly selected primary schools. Overweight children seeking treatment were recruited through tertiary centers. Children aged 6-13 yr were community-recruited normal weight (n = 73), community-recruited overweight (n = 53), and overweight treatment-seeking (n = 51). Medical history, family history, and symptoms of complications of overweight were collected by interview, and physical examination was performed. Investigations included oral glucose tolerance tests, fasting lipids, and liver function tests. Adjusted regression was used to model each complication of obesity with age- and sex-specific child BMI z-scores entered as a continuous dependent variable. Adjusted logistic regression showed the proportion of children with musculoskeletal pain, obstructive sleep apnea symptoms, headaches, depression, anxiety, bullying, and acanthosis nigricans increased with child BMI z-score. Adjusted linear regression showed BMI z-score was significantly related to systolic and diastolic blood pressure, insulin during oral glucose tolerance test, total cholesterol, high-density lipoprotein, triglycerides, and alanine aminotransferase. Child's BMI z-score is independently related to complications of overweight and obesity in a linear or curvilinear fashion. Children's risks of most complications increase across the entire range of BMI values and are not defined by thresholds.
The Association of Fever with Total Mechanical Ventilation Time in Critically Ill Patients.
Park, Dong Won; Egi, Moritoki; Nishimura, Masaji; Chang, Youjin; Suh, Gee Young; Lim, Chae Man; Kim, Jae Yeol; Tada, Keiichi; Matsuo, Koichi; Takeda, Shinhiro; Tsuruta, Ryosuke; Yokoyama, Takeshi; Kim, Seon Ok; Koh, Younsuck
2016-12-01
This research aims to investigate the impact of fever on total mechanical ventilation time (TVT) in critically ill patients. Subgroup analysis was conducted using a previous prospective, multicenter observational study. We included mechanically ventilated patients for more than 24 hours from 10 Korean and 15 Japanese intensive care units (ICU), and recorded maximal body temperature under the support of mechanical ventilation (MAX(MV)). To assess the independent association of MAX(MV) with TVT, we used propensity-matched analysis in a total of 769 survived patients with medical or surgical admission, separately. Together with multiple linear regression analysis to evaluate the association between the severity of fever and TVT, the effect of MAX(MV) on ventilator-free days was also observed by quantile regression analysis in all subjects including non-survivors. After propensity score matching, a MAX(MV) ≥ 37.5°C was significantly associated with longer mean TVT by 5.4 days in medical admission, and by 1.2 days in surgical admission, compared to those with MAX(MV) of 36.5°C to 37.4°C. In multivariate linear regression analysis, patients with three categories of fever (MAX(MV) of 37.5°C to 38.4°C, 38.5°C to 39.4°C, and ≥ 39.5°C) sustained a significantly longer duration of TVT than those with normal range of MAX(MV) in both categories of ICU admission. A significant association between MAX(MV) and mechanical ventilator-free days was also observed in all enrolled subjects. Fever may be a detrimental factor to prolong TVT in mechanically ventilated patients. These findings suggest that fever in mechanically ventilated patients might be associated with worse mechanical ventilation outcome.
Spectral-Spatial Shared Linear Regression for Hyperspectral Image Classification.
Haoliang Yuan; Yuan Yan Tang
2017-04-01
Classification of the pixels in hyperspectral image (HSI) is an important task and has been popularly applied in many practical applications. Its major challenge is the high-dimensional small-sized problem. To deal with this problem, lots of subspace learning (SL) methods are developed to reduce the dimension of the pixels while preserving the important discriminant information. Motivated by ridge linear regression (RLR) framework for SL, we propose a spectral-spatial shared linear regression method (SSSLR) for extracting the feature representation. Comparing with RLR, our proposed SSSLR has the following two advantages. First, we utilize a convex set to explore the spatial structure for computing the linear projection matrix. Second, we utilize a shared structure learning model, which is formed by original data space and a hidden feature space, to learn a more discriminant linear projection matrix for classification. To optimize our proposed method, an efficient iterative algorithm is proposed. Experimental results on two popular HSI data sets, i.e., Indian Pines and Salinas demonstrate that our proposed methods outperform many SL methods.
Simple linear and multivariate regression models.
Rodríguez del Águila, M M; Benítez-Parejo, N
2011-01-01
In biomedical research it is common to find problems in which we wish to relate a response variable to one or more variables capable of describing the behaviour of the former variable by means of mathematical models. Regression techniques are used to this effect, in which an equation is determined relating the two variables. While such equations can have different forms, linear equations are the most widely used form and are easy to interpret. The present article describes simple and multiple linear regression models, how they are calculated, and how their applicability assumptions are checked. Illustrative examples are provided, based on the use of the freely accessible R program. Copyright © 2011 SEICAP. Published by Elsevier Espana. All rights reserved.
Narayanan, Neethu; Gupta, Suman; Gajbhiye, V T; Manjaiah, K M
2017-04-01
A carboxy methyl cellulose-nano organoclay (nano montmorillonite modified with 35-45 wt % dimethyl dialkyl (C 14 -C 18 ) amine (DMDA)) composite was prepared by solution intercalation method. The prepared composite was characterized by infrared spectroscopy (FTIR), X-Ray diffraction spectroscopy (XRD) and scanning electron microscopy (SEM). The composite was utilized for its pesticide sorption efficiency for atrazine, imidacloprid and thiamethoxam. The sorption data was fitted into Langmuir and Freundlich isotherms using linear and non linear methods. The linear regression method suggested best fitting of sorption data into Type II Langmuir and Freundlich isotherms. In order to avoid the bias resulting from linearization, seven different error parameters were also analyzed by non linear regression method. The non linear error analysis suggested that the sorption data fitted well into Langmuir model rather than in Freundlich model. The maximum sorption capacity, Q 0 (μg/g) was given by imidacloprid (2000) followed by thiamethoxam (1667) and atrazine (1429). The study suggests that the degree of determination of linear regression alone cannot be used for comparing the best fitting of Langmuir and Freundlich models and non-linear error analysis needs to be done to avoid inaccurate results. Copyright © 2017 Elsevier Ltd. All rights reserved.
Huang, Hui; Zhu, Zheng-Qiu; Zhou, Zheng-Guo; Chen, Ling-Shan; Zhao, Ming; Zhang, Yang; Li, Hong-Bo; Yin, Li-Ping
2016-12-08
To assess the role of time-intensity curves (TICs) of the normal peripheral zone (PZ) in the identification of biopsy-proven prostate nodules using contrast-enhanced transrectal ultrasound (CETRUS). This study included 132 patients with 134 prostate PZ nodules. Arrival time (AT), peak intensity (PI), mean transit time (MTT), area under the curve (AUC), time from peak to one half (TPH), wash in slope (WIS) and time to peak (TTP) were analyzed using multivariate linear logistic regression and receiver operating characteristic (ROC) curves to assess whether combining nodule TICs with normal PZ TICs improved the prediction of prostate cancer (PCa) aggressiveness. The PI, AUC (p < 0.001 for both), MTT and TPH (p = 0.011 and 0.040 respectively) values of the malignant nodules were significantly higher than those of the benign nodules. Incorporating the PI and AUC values (both, p < 0.001) of the normal PZ TIC, but not the MTT and TPH values (p = 0.076 and 0.159 respectively), significantly improved the AUC for prediction of malignancy (PI: 0.784-0.923; AUC: 0.758-0.891) and assessment of cancer aggressiveness (p < 0.001). Thus, all these findings indicate that incorporating normal PZ TICs with nodule TICs in CETRUS readings can improve the diagnostic accuracy for PCa and cancer aggressiveness assessment.
London Measure of Unplanned Pregnancy: guidance for its use as an outcome measure
Hall, Jennifer A; Barrett, Geraldine; Copas, Andrew; Stephenson, Judith
2017-01-01
Background The London Measure of Unplanned Pregnancy (LMUP) is a psychometrically validated measure of the degree of intention of a current or recent pregnancy. The LMUP is increasingly being used worldwide, and can be used to evaluate family planning or preconception care programs. However, beyond recommending the use of the full LMUP scale, there is no published guidance on how to use the LMUP as an outcome measure. Ordinal logistic regression has been recommended informally, but studies published to date have all used binary logistic regression and dichotomized the scale at different cut points. There is thus a need for evidence-based guidance to provide a standardized methodology for multivariate analysis and to enable comparison of results. This paper makes recommendations for the regression method for analysis of the LMUP as an outcome measure. Materials and methods Data collected from 4,244 pregnant women in Malawi were used to compare five regression methods: linear, logistic with two cut points, and ordinal logistic with either the full or grouped LMUP score. The recommendations were then tested on the original UK LMUP data. Results There were small but no important differences in the findings across the regression models. Logistic regression resulted in the largest loss of information, and assumptions were violated for the linear and ordinal logistic regression. Consequently, robust standard errors were used for linear regression and a partial proportional odds ordinal logistic regression model attempted. The latter could only be fitted for grouped LMUP score. Conclusion We recommend the linear regression model with robust standard errors to make full use of the LMUP score when analyzed as an outcome measure. Ordinal logistic regression could be considered, but a partial proportional odds model with grouped LMUP score may be required. Logistic regression is the least-favored option, due to the loss of information. For logistic regression, the cut point for un/planned pregnancy should be between nine and ten. These recommendations will standardize the analysis of LMUP data and enhance comparability of results across studies. PMID:28435343
Torres, Daiane Placido; Martins-Teixeira, Maristela Braga; Cadore, Solange; Queiroz, Helena Müller
2015-01-01
A method for the determination of total mercury in fresh fish and shrimp samples by solid sampling thermal decomposition/amalgamation atomic absorption spectrometry (TDA AAS) has been validated following international foodstuff protocols in order to fulfill the Brazilian National Residue Control Plan. The experimental parameters have been previously studied and optimized according to specific legislation on validation and inorganic contaminants in foodstuff. Linearity, sensitivity, specificity, detection and quantification limits, precision (repeatability and within-laboratory reproducibility), robustness as well as accuracy of the method have been evaluated. Linearity of response was satisfactory for the two range concentrations available on the TDA AAS equipment, between approximately 25.0 and 200.0 μg kg(-1) (square regression) and 250.0 and 2000.0 μg kg(-1) (linear regression) of mercury. The residues for both ranges were homoscedastic and independent, with normal distribution. Correlation coefficients obtained for these ranges were higher than 0.995. Limits of quantification (LOQ) and of detection of the method (LDM), based on signal standard deviation (SD) for a low-in-mercury sample, were 3.0 and 1.0 μg kg(-1), respectively. Repeatability of the method was better than 4%. Within-laboratory reproducibility achieved a relative SD better than 6%. Robustness of the current method was evaluated and pointed sample mass as a significant factor. Accuracy (assessed as the analyte recovery) was calculated on basis of the repeatability, and ranged from 89% to 99%. The obtained results showed the suitability of the present method for direct mercury measurement in fresh fish and shrimp samples and the importance of monitoring the analysis conditions for food control purposes. Additionally, the competence of this method was recognized by accreditation under the standard ISO/IEC 17025.
External contribution to urban air pollution.
Grima, Ramon; Micallef, Alfred; Colls, Jeremy J
2002-02-01
Elevated particulate matter concentrations in urban locations have normally been associated with local traffic emissions. Recently it has been suggested that such episodes are influenced to a high degree by PM10 sources external to urban areas. To further corroborate this hypothesis, linear regression was sought between PM10 concentrations measured at eight urban sites in the U.K., with particulate sulphate concentration measured at two rural sites, for the years 1993-1997. Analysis of the slopes, intercepts and correlation coefficients indicate a possible relationship between urban PM10 and rural sulphate concentrations. The influences of wind direction and of the distance of the urban from the rural sites on the values of the three statistical parameters are also explored. The value of linear regression as an analysis tool in such cases is discussed and it is shown that an analysis of the sign of the rate of change of the urban PM10 and rural sulphate concentrations provides a more realistic method of correlation. The results indicate a major influence on urban PM10 concentrations from the eastern side of the United Kingdom. Linear correlation was also sought using PM10 data from nine urban sites in London and nearby rural Rochester. Analysis of the magnitude of the gradients and intercepts together with episode correlation analysis between the two sites showed the effect of transported PM10 on the local London concentrations. This article also presents methods to estimate the influence of rural and urban PM10 sources on urban PM10 concentrations and to obtain a rough estimate of the transboundary contribution to urban air pollution from the PM10 concentration data of the urban site.
Electromyographic analyses of muscle pre-activation induced by single joint exercise.
Júnior, Valdinar A R; Bottaro, Martim; Pereira, Maria C C; Andrade, Marcelino M; P Júnior, Paulo R W; Carmo, Jake C
2010-01-01
To investigate whether performing a low-intensity, single-joint exercises for knee extensors was an efficient strategy for increasing the number of motor units recruited in the vastus lateralis muscle during a subsequent multi-joint exercises. Nine healthy male participants (23.33+/-3.46 yrs) underwent bouts of exercise in which knee extension and 45 degrees , and leg press exercises were performed in sequence. In the low-intensity bout (R30), 15 unilateral knee extensions were performed, followed by 15 repetitions of the leg presses at 30% and 60% of one maximum repetition load (1-MR), respectively. In the high-intensity bout (R60), the same sequence was performed, but the applied load was 60% of 1-MR for both exercises. A single set of 15 repetitions of the leg press at 60% of 1-MR was performed as a control exercise (CR). The surface electromyographic signals of the vastus lateralis muscle were recorded by means of a linear electrode array. The root mean square (RMS) values were determined for each repetition of the leg press, and linear regressions were calculated from these results. The slopes of the straight lines obtained were then normalized using the linear coefficients of the regression equations and compared using one-way ANOVAs for repeated measures. The slopes observed in the CR were significantly lower than those in the R30 and R60 (p<0.05). The results indicated that the recruitment of motor units was more effective when a single-joint exercise preceded the multi-joint exercise. Article registered in the Australian New Zealand Clinical Trials Registry (ANZCTR) under the number ACTRN12609000413224.
Korany, Mohamed A; Maher, Hadir M; Galal, Shereen M; Ragab, Marwa A A
2013-05-01
This manuscript discusses the application and the comparison between three statistical regression methods for handling data: parametric, nonparametric, and weighted regression (WR). These data were obtained from different chemometric methods applied to the high-performance liquid chromatography response data using the internal standard method. This was performed on a model drug Acyclovir which was analyzed in human plasma with the use of ganciclovir as internal standard. In vivo study was also performed. Derivative treatment of chromatographic response ratio data was followed by convolution of the resulting derivative curves using 8-points sin x i polynomials (discrete Fourier functions). This work studies and also compares the application of WR method and Theil's method, a nonparametric regression (NPR) method with the least squares parametric regression (LSPR) method, which is considered the de facto standard method used for regression. When the assumption of homoscedasticity is not met for analytical data, a simple and effective way to counteract the great influence of the high concentrations on the fitted regression line is to use WR method. WR was found to be superior to the method of LSPR as the former assumes that the y-direction error in the calibration curve will increase as x increases. Theil's NPR method was also found to be superior to the method of LSPR as the former assumes that errors could occur in both x- and y-directions and that might not be normally distributed. Most of the results showed a significant improvement in the precision and accuracy on applying WR and NPR methods relative to LSPR.
Verster, Joris C; Roth, Thomas
2012-03-01
There are various methods to examine driving ability. Comparisons between these methods and their relationship with actual on-road driving is often not determined. The objective of this study was to determine whether laboratory tests measuring driving-related skills could adequately predict on-the-road driving performance during normal traffic. Ninety-six healthy volunteers performed a standardized on-the-road driving test. Subjects were instructed to drive with a constant speed and steady lateral position within the right traffic lane. Standard deviation of lateral position (SDLP), i.e., the weaving of the car, was determined. The subjects also performed a psychometric test battery including the DSST, Sternberg memory scanning test, a tracking test, and a divided attention test. Difference scores from placebo for parameters of the psychometric tests and SDLP were computed and correlated with each other. A stepwise linear regression analysis determined the predictive validity of the laboratory test battery to SDLP. Stepwise regression analyses revealed that the combination of five parameters, hard tracking, tracking and reaction time of the divided attention test, and reaction time and percentage of errors of the Sternberg memory scanning test, together had a predictive validity of 33.4%. The psychometric tests in this test battery showed insufficient predictive validity to replace the on-the-road driving test during normal traffic.
1994-09-01
Institute of Technology, Wright- Patterson AFB OH, January 1994. 4. Neter, John and others. Applied Linear Regression Models. Boston: Irwin, 1989. 5...Technology, Wright-Patterson AFB OH 5 April 1994. 29. Neter, John and others. Applied Linear Regression Models. Boston: Irwin, 1989. 30. Office of
An Evaluation of the Automated Cost Estimating Integrated Tools (ACEIT) System
1989-09-01
residual and it is described as the residual divided by its standard deviation (13:App A,17). Neter, Wasserman, and Kutner, in Applied Linear Regression Models...others. Applied Linear Regression Models. Homewood IL: Irwin, 1983. 19. Raduchel, William J. "A Professional’s Perspective on User-Friendliness," Byte
A Simple and Convenient Method of Multiple Linear Regression to Calculate Iodine Molecular Constants
ERIC Educational Resources Information Center
Cooper, Paul D.
2010-01-01
A new procedure using a student-friendly least-squares multiple linear-regression technique utilizing a function within Microsoft Excel is described that enables students to calculate molecular constants from the vibronic spectrum of iodine. This method is advantageous pedagogically as it calculates molecular constants for ground and excited…
Conjoint Analysis: A Study of the Effects of Using Person Variables.
ERIC Educational Resources Information Center
Fraas, John W.; Newman, Isadore
Three statistical techniques--conjoint analysis, a multiple linear regression model, and a multiple linear regression model with a surrogate person variable--were used to estimate the relative importance of five university attributes for students in the process of selecting a college. The five attributes include: availability and variety of…
Fitting program for linear regressions according to Mahon (1996)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Trappitsch, Reto G.
2018-01-09
This program takes the users' Input data and fits a linear regression to it using the prescription presented by Mahon (1996). Compared to the commonly used York fit, this method has the correct prescription for measurement error propagation. This software should facilitate the proper fitting of measurements with a simple Interface.
How Robust Is Linear Regression with Dummy Variables?
ERIC Educational Resources Information Center
Blankmeyer, Eric
2006-01-01
Researchers in education and the social sciences make extensive use of linear regression models in which the dependent variable is continuous-valued while the explanatory variables are a combination of continuous-valued regressors and dummy variables. The dummies partition the sample into groups, some of which may contain only a few observations.…
Revisiting the Scale-Invariant, Two-Dimensional Linear Regression Method
ERIC Educational Resources Information Center
Patzer, A. Beate C.; Bauer, Hans; Chang, Christian; Bolte, Jan; Su¨lzle, Detlev
2018-01-01
The scale-invariant way to analyze two-dimensional experimental and theoretical data with statistical errors in both the independent and dependent variables is revisited by using what we call the triangular linear regression method. This is compared to the standard least-squares fit approach by applying it to typical simple sets of example data…
ERIC Educational Resources Information Center
Thompson, Russel L.
Homoscedasticity is an important assumption of linear regression. This paper explains what it is and why it is important to the researcher. Graphical and mathematical methods for testing the homoscedasticity assumption are demonstrated. Sources of homoscedasticity and types of homoscedasticity are discussed, and methods for correction are…
Common pitfalls in statistical analysis: Linear regression analysis
Aggarwal, Rakesh; Ranganathan, Priya
2017-01-01
In a previous article in this series, we explained correlation analysis which describes the strength of relationship between two continuous variables. In this article, we deal with linear regression analysis which predicts the value of one continuous variable from another. We also discuss the assumptions and pitfalls associated with this analysis. PMID:28447022
Comparison of l₁-Norm SVR and Sparse Coding Algorithms for Linear Regression.
Zhang, Qingtian; Hu, Xiaolin; Zhang, Bo
2015-08-01
Support vector regression (SVR) is a popular function estimation technique based on Vapnik's concept of support vector machine. Among many variants, the l1-norm SVR is known to be good at selecting useful features when the features are redundant. Sparse coding (SC) is a technique widely used in many areas and a number of efficient algorithms are available. Both l1-norm SVR and SC can be used for linear regression. In this brief, the close connection between the l1-norm SVR and SC is revealed and some typical algorithms are compared for linear regression. The results show that the SC algorithms outperform the Newton linear programming algorithm, an efficient l1-norm SVR algorithm, in efficiency. The algorithms are then used to design the radial basis function (RBF) neural networks. Experiments on some benchmark data sets demonstrate the high efficiency of the SC algorithms. In particular, one of the SC algorithms, the orthogonal matching pursuit is two orders of magnitude faster than a well-known RBF network designing algorithm, the orthogonal least squares algorithm.
Lemaitre, Herve; Goldman, Aaron L; Sambataro, Fabio; Verchinski, Beth A; Meyer-Lindenberg, Andreas; Weinberger, Daniel R; Mattay, Venkata S
2012-03-01
Normal aging is accompanied by global as well as regional structural changes. While these age-related changes in gray matter volume have been extensively studied, less has been done using newer morphological indexes, such as cortical thickness and surface area. To this end, we analyzed structural images of 216 healthy volunteers, ranging from 18 to 87 years of age, using a surface-based automated parcellation approach. Linear regressions of age revealed a concomitant global age-related reduction in cortical thickness, surface area and volume. Cortical thickness and volume collectively confirmed the vulnerability of the prefrontal cortex, whereas in other cortical regions, such as in the parietal cortex, thickness was the only measure sensitive to the pronounced age-related atrophy. No cortical regions showed more surface area reduction than the global average. The distinction between these morphological measures may provide valuable information to dissect age-related structural changes of the brain, with each of these indexes probably reflecting specific histological changes occurring during aging. Published by Elsevier Inc.
Ichikawa, Akio; Ono, Hiroshi; Furuta, Kenjiro; Shiotsuki, Takahiro; Shinoda, Tetsuro
2007-08-17
Juvenile hormone III (JH III) racemate was prepared from methyl (2E,6E)-farnesoate via epoxidation with 3-chloroperbenzoic acid (mCPBA). Enantioselective separation of JH III was conducted using normal-phase high-performance liquid chromatography (HPLC) on a chiral stationary phase. [(2)H(3)]Methyl (2E,6E)-farnesoate was also prepared from (2E,6E)-farnesoic acid and [(2)H(4)]methanol (methanol-d(4)) using 1-(3-dimethylaminopropyl)-3-ethylcarbodiimide hydrochloride (EDC) and 4-dimethylaminopyridine (DMAP); the conjugated double bond underwent isomerization to some degree. Epoxidation of [(2)H(3)]methyl (2E,6E)-farnesoate with mCPBA gave a novel deuterium-substituted internal standard [(2)H(3)]JH III (JH III-d(3)). The standard curve was produced by linear regression using the peak area ratios of JH III and JH III-d(3) in liquid chromatography-mass spectrometry (LC-MS).
Machine Learning to Assess Grassland Productivity in Southeastern Arizona
NASA Astrophysics Data System (ADS)
Ponce-Campos, G. E.; Heilman, P.; Armendariz, G.; Moser, E.; Archer, V.; Vaughan, R.
2015-12-01
We present preliminary results of machine learning (ML) techniques modeling the combined effects of climate, management, and inherent potential on productivity of grazed semi-arid grasslands in southeastern Arizona. Our goal is to support public land managers determine if agency management policies are meeting objectives and where to focus attention. Monitoring in the field is becoming more and more limited in space and time. Remotely sensed data cover the entire allotments and go back in time, but do not consider the key issue of species composition. By estimating expected vegetative production as a function of site potential and climatic inputs, management skill can be assessed through time, across individual allotments, and between allotments. Here we present the use of Random Forest (RF) as the main ML technique, in this case for the purpose of regression. Our response variable is the maximum annual NDVI, a surrogate for grassland productivity, as generated by the Google Earth Engine cloud computing platform based on Landsat 5, 7, and 8 datasets. PRISM 33-year normal precipitation (1980-2013) was resampled to the Landsat scale. In addition, the GRIDMET climate dataset was the source for the calculation of the annual SPEI (Standardized Precipitation Evapotranspiration Index), a drought index. We also included information about landscape position, aspect, streams, ponds, roads and fire disturbances as part of the modeling process. Our results show that in terms of variable importance, the 33-year normal precipitation, along with SPEI, are the most important features affecting grasslands productivity within the study area. The RF approach was compared to a linear regression model with the same variables. The linear model resulted in an r2 = 0.41, whereas RF showed a significant improvement with an r2 = 0.79. We continue refining the model by comparison with aerial photography and to include grazing intensity and infrastructure from units/allotments to assess the effect of management practices on vegetation production.
A geometric approach to aortic root surgical anatomy.
Contino, Monica; Mangini, Andrea; Lemma, Massimo Giovanni; Romagnoni, Claudia; Zerbi, Pietro; Gelpi, Guido; Antona, Carlo
2016-01-01
The aim of this study was the analysis of the geometrical relationships between the different structures constituting the aortic root, with particular attention to interleaflet triangles, haemodynamic ventriculo-arterial junction and functional aortic annulus in normal subjects. Sixteen formol-fixed human hearts with normal aortic roots were studied. The aortic root was isolated, sectioned at the midpoint of the non-coronary sinus, spread apart and photographed by a high-resolution digital camera. After calibration and picture resizing, the software AutoCAD 2004 was used to identify and measure all the elements of the interleaflets triangles and of the aortic root that were objects of our analysis. Multiple comparisons were performed with one-way analysis of variance for continuous data and with Kruskal-Wallis analysis for non-continuous data. Linear regression and Pearson's product correlation were used to correlate root element dimensions when appropriate. Student's t-test was used to compare means for unpaired data. Heron's formula was applied to estimate the functional aortic annular diameters. The non coronary-left coronary interleaflets triangles were larger, followed by inter-coronary and right-non-coronary ones. The apical angle is <60° and its standard deviation can be considered an asymmetry index. The sinu-tubular junction was shown to be 10% larger than the virtual basal ring (VBR). The mathematical relationship between the haemodynamic ventriculo-arterial junction and the VBR calculated by linear regression and expressed in terms of the diameter was: haemodynamic ventriculo-arterial junction = 2.29 VBR (diameter) + 47. Conservative aortic surgery is based on a better understanding of aortic root anatomy and physiology. The relationships among its elements are of paramount importance during aortic valve repair/sparing procedures and they can be useful also in echocardiographic analysis and in computed tomography reconstruction. © The Author 2015. Published by Oxford University Press on behalf of the European Association for Cardio-Thoracic Surgery. All rights reserved.
Messinger, Mindl M; Moffett, Brady S; Wilfong, Angus
2015-12-01
Obesity has been shown to affect the disposition of water-soluble medications in pediatric patients. There are no published data describing serum phenytoin concentrations in obese pediatric patients. A retrospective descriptive study was designed that included patients from 2011 to 2013 between 2 and 19 years of age who received a dose of fosphenytoin with a subsequent serum phenytoin concentration, drawn 2-4 hours postloading dose. Body mass index (BMI) was calculated and patients were categorized by BMI percentiles into underweight (<5th percentile), normal weight (5th-84th percentile), overweight (85th-94th percentile), and obese (≥95th percentile). Descriptive statistical analysis and comparisons between groups occurred to determine differences in serum phenytoin concentrations. Multivariable linear regression analysis was performed to determine the effect of body habitus on serum phenytoin concentrations. One hundred ten patients met study criteria (male 51.8%, mean age: 8.3 ± 4.9 years). Patients were normal weight (47.3%), underweight (20.9%), overweight (14.6%), and obese (17.3%). No significant differences were identified between groups in regard to patient demographics, with the exception of weight (P < 0.05). The mean fosphenytoin dose was 23.4 ± 5.7 mg Phenytoin Equivalents (PE)/kg and the serum phenytoin concentration was 22.4 ± 6.8 mg/L measured at 2.9 ± 0.6 hours after dose, and this did not vary significantly across groups (P > 0.05). Multivariable linear regression identified body habitus as a nonsignificant predictor of serum phenytoin concentrations (P > 0.05). Patients of higher BMI did not require further antiepileptic therapy as compared with patients with lower BMI (P > 0.05). Contrary to the adult population, loading dose adjustments do not seem to be required in pediatric patients. Obesity does not affect serum phenytoin concentrations in pediatric patients after intravenous bolus fosphenytoin administration.
NASA Astrophysics Data System (ADS)
Knowles, J. F.; Lestak, L.; Molotch, N. P.
2016-12-01
We evaluated the long term (1989-2012) relationship between the satellite-observed Normalized Difference Vegetation Index (NDVI), snowpack accumulation, and atmospheric demand throughout the Southern Rocky Mountain Ecoregion, USA. Deviations from this relationship were further explored during pre- and post-disturbance conditions associated with bark beetles and drought. Over the entire study area, both the snow water equivalent (SWE) and a snow aridity index (SAI), which used the SWE to normalize potential evapotranspiration (PET), were significant predictors of the long-term AVHRR NDVI, but the SAI was a better predictor of NDVI relative to SWE regardless of disturbance. Since these relationships were weaker in disturbed areas, we also introduced a metric of tree mortality, and subsequent multiple linear regression of SAI and cumulative mortality best predicted the NDVI from a pair of heavily impacted focus areas within the larger study area. The post-disturbance NDVI was systematically reduced per unit SAI in these areas, and the difference between the observed and predicted (from pre-disturbance regressions) post-disturbance NDVI was significantly correlated with the cumulative forest mortality. At the Ecoregion scale, these disturbance effects were not clearly evident, and we attribute this to spatial variability of both SAI and NDVI throughout the large study area as evidenced by spatial analysis of Moderate Resolution Imaging Spectroradiometer (MODIS)-derived data. These results constrain the expected reduction in forest productivity due to disturbance and demonstrate that this reduction can be particularly evident during drought conditions resultant from low snow accumulation during the winter. Hence, terrestrial carbon uptake may decrease non-linearly post disturbance. This work has implications for predicting the ecohydrological response to climate change in the southern Rocky Mountains, as reductions in SWE and increases in PET are predicted for this area in the future, and therefore changes in the terrestrial carbon, water, and energy cycles should be expected.
SU-E-J-83: CBCT Based Rectum and Bladder Dose Tracking in the Prostate Radiotherapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Z; Wang, J; Yang, Z
2015-06-15
Purpose: The aim of this study is to monitor the volume changes of bladder and rectum and evaluate the dosimetric changes of bladder and rectum using daily cone-beam CT for prostate radiotherapy. Methods: The data of this study were obtained from 12 patients, totally 222 CBCTs. All the volume of the bladder and the rectum on the CBCT were normalized to the bladder and the rectum on their own original CT to monitory the volume changes. To evaluate dose delivered to the OARs, volumes that receive 70Gy (V70Gy), 60Gy, 50Gy, 40Gy and 30Gy are calculated for the bladder and themore » rectum, V20Gy and V10Gy for rectum additionally. And the deviation of the mean dose to the bladder and the rectum are also chosen as the evaluation parameter. Linear regression analysis was performed to identify the mean dose change of the volume change using SPSS 19. Results: The results show that the variances of the normalize volume of the bladder and the rectum are 0.15–0.58 and 0.13–0.50. The variances of V70Gy, V60Gy, V50Gy, V40Gy and V30Gy of bladder are bigger than rectum for 11 patients. The linear regression analysis indicated a negative correlation between the volume and the mean dose of the bladder (p < 0.05). A 10% increase in bladder volume will cause 5.1% (±4.3%) reduction in mean dose. Conclusion: The bladder volume change is more significant than that for rectum for the prostate cancer patient. The volume changes of rectum are not significant except air gap in the rectum. Bladder volume varies will cause significant dose change. The bladder volume monitoring before fractional treatment delivery would be crucial for accuracy dose delivery.« less
Role of T1 mapping as a complementary tool to T2* for non-invasive cardiac iron overload assessment.
Torlasco, Camilla; Cassinerio, Elena; Roghi, Alberto; Faini, Andrea; Capecchi, Marco; Abdel-Gadir, Amna; Giannattasio, Cristina; Parati, Gianfranco; Moon, James C; Cappellini, Maria D; Pedrotti, Patrizia
2018-01-01
Iron overload-related heart failure is the principal cause of death in transfusion dependent patients, including those with Thalassemia Major. Linking cardiac siderosis measured by T2* to therapy improves outcomes. T1 mapping can also measure iron; preliminary data suggests it may have higher sensitivity for iron, particularly for early overload (the conventional cut-point for no iron by T2* is 20ms, but this is believed insensitive). We compared T1 mapping to T2* in cardiac iron overload. In a prospectively large single centre study of 138 Thalassemia Major patients and 32 healthy controls, we compared T1 mapping to dark blood and bright blood T2* acquired at 1.5T. Linear regression analysis was used to assess the association of T2* and T1. A "moving window" approach was taken to understand the strength of the association at different levels of iron overload. The relationship between T2* (here dark blood) and T1 is described by a log-log linear regression, which can be split in three different slopes: 1) T2* low, <20ms, r2 = 0.92; 2) T2* = 20-30ms, r2 = 0.48; 3) T2*>30ms, weak relationship. All subjects with T2*<20ms had low T1; among those with T2*>20ms, 38% had low T1 with most of the subjects in the T2* range 20-30ms having a low T1. In established cardiac iron overload, T1 and T2* are concordant. However, in the 20-30ms T2* range, T1 mapping appears to detect iron. These data support previous suggestions that T1 detects missed iron in 1 out of 3 subjects with normal T2*, and that T1 mapping is complementary to T2*. The clinical significance of a low T1 with normal T2* should be further investigated.
Konietschke, Frank; Libiger, Ondrej; Hothorn, Ludwig A
2012-01-01
Statistical association between a single nucleotide polymorphism (SNP) genotype and a quantitative trait in genome-wide association studies is usually assessed using a linear regression model, or, in the case of non-normally distributed trait values, using the Kruskal-Wallis test. While linear regression models assume an additive mode of inheritance via equi-distant genotype scores, Kruskal-Wallis test merely tests global differences in trait values associated with the three genotype groups. Both approaches thus exhibit suboptimal power when the underlying inheritance mode is dominant or recessive. Furthermore, these tests do not perform well in the common situations when only a few trait values are available in a rare genotype category (disbalance), or when the values associated with the three genotype categories exhibit unequal variance (variance heterogeneity). We propose a maximum test based on Marcus-type multiple contrast test for relative effect sizes. This test allows model-specific testing of either dominant, additive or recessive mode of inheritance, and it is robust against variance heterogeneity. We show how to obtain mode-specific simultaneous confidence intervals for the relative effect sizes to aid in interpreting the biological relevance of the results. Further, we discuss the use of a related all-pairwise comparisons contrast test with range preserving confidence intervals as an alternative to Kruskal-Wallis heterogeneity test. We applied the proposed maximum test to the Bogalusa Heart Study dataset, and gained a remarkable increase in the power to detect association, particularly for rare genotypes. Our simulation study also demonstrated that the proposed non-parametric tests control family-wise error rate in the presence of non-normality and variance heterogeneity contrary to the standard parametric approaches. We provide a publicly available R library nparcomp that can be used to estimate simultaneous confidence intervals or compatible multiplicity-adjusted p-values associated with the proposed maximum test.
Jiang, Rengui; Xie, Jiancang; He, Hailong; Kuo, Chun-Chao; Zhu, Jiwei; Yang, Mingxiang
2016-09-01
As one of the most popular vegetation indices to monitor terrestrial vegetation productivity, Normalized Difference Vegetation Index (NDVI) has been widely used to study the plant growth and vegetation productivity around the world, especially the dynamic response of vegetation to climate change in terms of precipitation and temperature. Alberta is the most important agricultural and forestry province and with the best climatic observation systems in Canada. However, few studies pertaining to climate change and vegetation productivity are found. The objectives of this paper therefore were to better understand impacts of climate change on vegetation productivity in Alberta using the NDVI and provide reference for policy makers and stakeholders. We investigated the following: (1) the variations of Alberta's smoothed NDVI (sNDVI, eliminated noise compared to NDVI) and two climatic variables (precipitation and temperature) using non-parametric Mann-Kendall monotonic test and Thiel-Sen's slope; (2) the relationships between sNDVI and climatic variables, and the potential predictability of sNDVI using climatic variables as predictors based on two predicted models; and (3) the use of a linear regression model and an artificial neural network calibrated by the genetic algorithm (ANN-GA) to estimate Alberta's sNDVI using precipitation and temperature as predictors. The results showed that (1) the monthly sNDVI has increased during the past 30 years and a lengthened growing season was detected; (2) vegetation productivity in northern Alberta was mainly temperature driven and the vegetation in southern Alberta was predominantly precipitation driven for the period of 1982-2011; and (3) better performances of the sNDVI-climate relationships were obtained by nonlinear model (ANN-GA) than using linear (regression) model. Similar results detected in both monthly and summer sNDVI prediction using climatic variables as predictors revealed the applicability of two models for different period of year ecologists might focus on.
NASA Astrophysics Data System (ADS)
Wu, Cheng; Zhen Yu, Jian
2018-03-01
Linear regression techniques are widely used in atmospheric science, but they are often improperly applied due to lack of consideration or inappropriate handling of measurement uncertainty. In this work, numerical experiments are performed to evaluate the performance of five linear regression techniques, significantly extending previous works by Chu and Saylor. The five techniques are ordinary least squares (OLS), Deming regression (DR), orthogonal distance regression (ODR), weighted ODR (WODR), and York regression (YR). We first introduce a new data generation scheme that employs the Mersenne twister (MT) pseudorandom number generator. The numerical simulations are also improved by (a) refining the parameterization of nonlinear measurement uncertainties, (b) inclusion of a linear measurement uncertainty, and (c) inclusion of WODR for comparison. Results show that DR, WODR and YR produce an accurate slope, but the intercept by WODR and YR is overestimated and the degree of bias is more pronounced with a low R2 XY dataset. The importance of a properly weighting parameter λ in DR is investigated by sensitivity tests, and it is found that an improper λ in DR can lead to a bias in both the slope and intercept estimation. Because the λ calculation depends on the actual form of the measurement error, it is essential to determine the exact form of measurement error in the XY data during the measurement stage. If a priori error in one of the variables is unknown, or the measurement error described cannot be trusted, DR, WODR and YR can provide the least biases in slope and intercept among all tested regression techniques. For these reasons, DR, WODR and YR are recommended for atmospheric studies when both X and Y data have measurement errors. An Igor Pro-based program (Scatter Plot) was developed to facilitate the implementation of error-in-variables regressions.
Afantitis, Antreas; Melagraki, Georgia; Sarimveis, Haralambos; Koutentis, Panayiotis A; Markopoulos, John; Igglessi-Markopoulou, Olga
2006-08-01
A quantitative-structure activity relationship was obtained by applying Multiple Linear Regression Analysis to a series of 80 1-[2-hydroxyethoxy-methyl]-6-(phenylthio) thymine (HEPT) derivatives with significant anti-HIV activity. For the selection of the best among 37 different descriptors, the Elimination Selection Stepwise Regression Method (ES-SWR) was utilized. The resulting QSAR model (R (2) (CV) = 0.8160; S (PRESS) = 0.5680) proved to be very accurate both in training and predictive stages.
Wavelet regression model in forecasting crude oil price
NASA Astrophysics Data System (ADS)
Hamid, Mohd Helmie; Shabri, Ani
2017-05-01
This study presents the performance of wavelet multiple linear regression (WMLR) technique in daily crude oil forecasting. WMLR model was developed by integrating the discrete wavelet transform (DWT) and multiple linear regression (MLR) model. The original time series was decomposed to sub-time series with different scales by wavelet theory. Correlation analysis was conducted to assist in the selection of optimal decomposed components as inputs for the WMLR model. The daily WTI crude oil price series has been used in this study to test the prediction capability of the proposed model. The forecasting performance of WMLR model were also compared with regular multiple linear regression (MLR), Autoregressive Moving Average (ARIMA) and Generalized Autoregressive Conditional Heteroscedasticity (GARCH) using root mean square errors (RMSE) and mean absolute errors (MAE). Based on the experimental results, it appears that the WMLR model performs better than the other forecasting technique tested in this study.
Partitioning sources of variation in vertebrate species richness
Boone, R.B.; Krohn, W.B.
2000-01-01
Aim: To explore biogeographic patterns of terrestrial vertebrates in Maine, USA using techniques that would describe local and spatial correlations with the environment. Location: Maine, USA. Methods: We delineated the ranges within Maine (86,156 km2) of 275 species using literature and expert review. Ranges were combined into species richness maps, and compared to geomorphology, climate, and woody plant distributions. Methods were adapted that compared richness of all vertebrate classes to each environmental correlate, rather than assessing a single explanatory theory. We partitioned variation in species richness into components using tree and multiple linear regression. Methods were used that allowed for useful comparisons between tree and linear regression results. For both methods we partitioned variation into broad-scale (spatially autocorrelated) and fine-scale (spatially uncorrelated) explained and unexplained components. By partitioning variance, and using both tree and linear regression in analyses, we explored the degree of variation in species richness for each vertebrate group that Could be explained by the relative contribution of each environmental variable. Results: In tree regression, climate variation explained richness better (92% of mean deviance explained for all species) than woody plant variation (87%) and geomorphology (86%). Reptiles were highly correlated with environmental variation (93%), followed by mammals, amphibians, and birds (each with 84-82% deviance explained). In multiple linear regression, climate was most closely associated with total vertebrate richness (78%), followed by woody plants (67%) and geomorphology (56%). Again, reptiles were closely correlated with the environment (95%), followed by mammals (73%), amphibians (63%) and birds (57%). Main conclusions: Comparing variation explained using tree and multiple linear regression quantified the importance of nonlinear relationships and local interactions between species richness and environmental variation, identifying the importance of linear relationships between reptiles and the environment, and nonlinear relationships between birds and woody plants, for example. Conservation planners should capture climatic variation in broad-scale designs; temperatures may shift during climate change, but the underlying correlations between the environment and species richness will presumably remain.
Javed, Faizan; Chan, Gregory S H; Savkin, Andrey V; Middleton, Paul M; Malouf, Philip; Steel, Elizabeth; Mackie, James; Lovell, Nigel H
2009-01-01
This paper uses non-linear support vector regression (SVR) to model the blood volume and heart rate (HR) responses in 9 hemodynamically stable kidney failure patients during hemodialysis. Using radial bias function (RBF) kernels the non-parametric models of relative blood volume (RBV) change with time as well as percentage change in HR with respect to RBV were obtained. The e-insensitivity based loss function was used for SVR modeling. Selection of the design parameters which includes capacity (C), insensitivity region (e) and the RBF kernel parameter (sigma) was made based on a grid search approach and the selected models were cross-validated using the average mean square error (AMSE) calculated from testing data based on a k-fold cross-validation technique. Linear regression was also applied to fit the curves and the AMSE was calculated for comparison with SVR. For the model based on RBV with time, SVR gave a lower AMSE for both training (AMSE=1.5) as well as testing data (AMSE=1.4) compared to linear regression (AMSE=1.8 and 1.5). SVR also provided a better fit for HR with RBV for both training as well as testing data (AMSE=15.8 and 16.4) compared to linear regression (AMSE=25.2 and 20.1).
Jasiewicz, Jan M; Allum, John H J; Middleton, James W; Barriskill, Andrew; Condie, Peter; Purcell, Brendan; Li, Raymond Che Tin
2006-12-01
We report on three different methods of gait event detection (toe-off and heel strike) using miniature linear accelerometers and angular velocity transducers in comparison to using standard pressure-sensitive foot switches. Detection was performed with normal and spinal-cord injured subjects. The detection of end contact (EC), normally toe-off, and initial contact (IC) normally, heel strike was based on either foot linear accelerations or foot sagittal angular velocity or shank sagittal angular velocity. The results showed that all three methods were as accurate as foot switches in estimating times of IC and EC for normal gait patterns. In spinal-cord injured subjects, shank angular velocity was significantly less accurate (p<0.02). We conclude that detection based on foot linear accelerations or foot angular velocity can correctly identify the timing of IC and EC events in both normal and spinal-cord injured subjects.
Di Stefano, Danilo Alessio; Arosio, Paolo
2016-01-01
Bone density at implant placement sites is one of the key factors affecting implant primary stability, which is a determinant for implant osseointegration and rehabilitation success. Site-specific bone density assessment is, therefore, of paramount importance. Recently, an implant micromotor endowed with an instantaneous torque-measuring system has been introduced. The aim of this study was to assess the reliability of this system. Five blocks with different densities (0.16, 0.26, 0.33, 0.49, and 0.65 g/cm(3)) were used. A single trained operator measured the density of one of them (0.33 g/cm(3)), by means of five different devices (20 measurements/device). The five resulting datasets were analyzed through the analysis of variance (ANOVA) model to investigate interdevice variability. As differences were not significant (P = .41), the five devices were each assigned to a different operator, who collected 20 density measurements for each block, both under irrigation (I) and without irrigation (NI). Measurements were pooled and averaged for each block, and their correlation with the actual block-density values was investigated using linear regression analysis. The possible effect of irrigation on density measurement was additionally assessed. Different devices provided reproducible, homogenous results. No significant interoperator variability was observed. Within the physiologic range of densities (> 0.30 g/cm(3)), the linear regression analysis showed a significant linear correlation between the mean torque measurements and the actual bone densities under both drilling conditions (r = 0.990 [I], r = 0.999 [NI]). Calibration lines were drawn under both conditions. Values collected under irrigation were lower than those collected without irrigation at all densities. The NI/I mean torque ratio was shown to decrease linearly with density (r = 0.998). The mean error introduced by the device-operator system was less than 10% in the range of normal jawbone density. Measurements performed with the device were linearly correlated with the blocks' bone densities. The results validate the device as an objective intraoperative tool for bone-density assessment that may contribute to proper jawbone-density evaluation and implant-insertion planning.
ERIC Educational Resources Information Center
Vasu, Ellen Storey
1978-01-01
The effects of the violation of the assumption of normality in the conditional distributions of the dependent variable, coupled with the condition of multicollinearity upon the outcome of testing the hypothesis that the regression coefficient equals zero, are investigated via a Monte Carlo study. (Author/JKS)
NASA Technical Reports Server (NTRS)
Jolly, William H.
1992-01-01
Relationships defining the ballistic limit of Space Station Freedom's (SSF) dual wall protection systems have been determined. These functions were regressed from empirical data found in Marshall Space Flight Center's (MSFC) Hypervelocity Impact Testing Summary (HITS) for the velocity range between three and seven kilometers per second. A stepwise linear least squares regression was used to determine the coefficients of several expressions that define a ballistic limit surface. Using statistical significance indicators and graphical comparisons to other limit curves, a final set of expressions is recommended for potential use in Probability of No Critical Flaw (PNCF) calculations for Space Station. The three equations listed below represent the mean curves for normal, 45 degree, and 65 degree obliquity ballistic limits, respectively, for a dual wall protection system consisting of a thin 6061-T6 aluminum bumper spaced 4.0 inches from a .125 inches thick 2219-T87 rear wall with multiple layer thermal insulation installed between the two walls. Normal obliquity is d(sub c) = 1.0514 v(exp 0.2983 t(sub 1)(exp 0.5228). Forty-five degree obliquity is d(sub c) = 0.8591 v(exp 0.0428) t(sub 1)(exp 0.2063). Sixty-five degree obliquity is d(sub c) = 0.2824 v(exp 0.1986) t(sub 1)(exp -0.3874). Plots of these curves are provided. A sensitivity study on the effects of using these new equations in the probability of no critical flaw analysis indicated a negligible increase in the performance of the dual wall protection system for SSF over the current baseline. The magnitude of the increase was 0.17 percent over 25 years on the MB-7 configuration run with the Bumper II program code.
NASA Astrophysics Data System (ADS)
Lopez, Patricia; Verkade, Jan; Weerts, Albrecht; Solomatine, Dimitri
2014-05-01
Hydrological forecasting is subject to many sources of uncertainty, including those originating in initial state, boundary conditions, model structure and model parameters. Although uncertainty can be reduced, it can never be fully eliminated. Statistical post-processing techniques constitute an often used approach to estimate the hydrological predictive uncertainty, where a model of forecast error is built using a historical record of past forecasts and observations. The present study focuses on the use of the Quantile Regression (QR) technique as a hydrological post-processor. It estimates the predictive distribution of water levels using deterministic water level forecasts as predictors. This work aims to thoroughly verify uncertainty estimates using the implementation of QR that was applied in an operational setting in the UK National Flood Forecasting System, and to inter-compare forecast quality and skill in various, differing configurations of QR. These configurations are (i) 'classical' QR, (ii) QR constrained by a requirement that quantiles do not cross, (iii) QR derived on time series that have been transformed into the Normal domain (Normal Quantile Transformation - NQT), and (iv) a piecewise linear derivation of QR models. The QR configurations are applied to fourteen hydrological stations on the Upper Severn River with different catchments characteristics. Results of each QR configuration are conditionally verified for progressively higher flood levels, in terms of commonly used verification metrics and skill scores. These include Brier's probability score (BS), the continuous ranked probability score (CRPS) and corresponding skill scores as well as the Relative Operating Characteristic score (ROCS). Reliability diagrams are also presented and analysed. The results indicate that none of the four Quantile Regression configurations clearly outperforms the others.
NASA Astrophysics Data System (ADS)
Jolly, William H.
1992-05-01
Relationships defining the ballistic limit of Space Station Freedom's (SSF) dual wall protection systems have been determined. These functions were regressed from empirical data found in Marshall Space Flight Center's (MSFC) Hypervelocity Impact Testing Summary (HITS) for the velocity range between three and seven kilometers per second. A stepwise linear least squares regression was used to determine the coefficients of several expressions that define a ballistic limit surface. Using statistical significance indicators and graphical comparisons to other limit curves, a final set of expressions is recommended for potential use in Probability of No Critical Flaw (PNCF) calculations for Space Station. The three equations listed below represent the mean curves for normal, 45 degree, and 65 degree obliquity ballistic limits, respectively, for a dual wall protection system consisting of a thin 6061-T6 aluminum bumper spaced 4.0 inches from a .125 inches thick 2219-T87 rear wall with multiple layer thermal insulation installed between the two walls. Normal obliquity is d(sub c) = 1.0514 v(exp 0.2983 t(sub 1)(exp 0.5228). Forty-five degree obliquity is d(sub c) = 0.8591 v(exp 0.0428) t(sub 1)(exp 0.2063). Sixty-five degree obliquity is d(sub c) = 0.2824 v(exp 0.1986) t(sub 1)(exp -0.3874). Plots of these curves are provided. A sensitivity study on the effects of using these new equations in the probability of no critical flaw analysis indicated a negligible increase in the performance of the dual wall protection system for SSF over the current baseline. The magnitude of the increase was 0.17 percent over 25 years on the MB-7 configuration run with the Bumper II program code.
Multi-satellites normalization of the FengYun-2s visible detectors by the MVP method
NASA Astrophysics Data System (ADS)
Li, Yuan; Rong, Zhi-guo; Zhang, Li-jun; Sun, Ling; Xu, Na
2013-08-01
After January 13, 2012, FY-2F had successfully launched, the total number of the in orbit operating FengYun-2 geostationary meteorological satellites reached three. For accurate and efficient application of multi-satellite observation data, the study of the multi-satellites normalization of the visible detector was urgent. The method required to be non-rely on the in orbit calibration. So as to validate the calibration results before and after the launch; calculate day updating surface bidirectional reflectance distribution function (BRDF); at the same time track the long-term decay phenomenon of the detector's linearity and responsivity. By research of the typical BRDF model, the normalization method was designed. Which could effectively solute the interference of surface directional reflectance characteristics, non-rely on visible detector in orbit calibration. That was the Median Vertical Plane (MVP) method. The MVP method was based on the symmetry of principal plane, which were the directional reflective properties of the general surface targets. Two geostationary satellites were taken as the endpoint of a segment, targets on the intersecting line of the segment's MVP and the earth surface could be used as a normalization reference target (NRT). Observation on the NRT by two satellites at the moment the sun passing through the MVP brought the same observation zenith, solar zenith, and opposite relative direction angle. At that time, the linear regression coefficients of the satellite output data were the required normalization coefficients. The normalization coefficients between FY-2D, FY-2E and FY-2F were calculated, and the self-test method of the normalized results was designed and realized. The results showed the differences of the responsivity between satellites could up to 10.1%(FY-2E to FY-2F); the differences of the output reflectance calculated by the broadcast calibration look-up table could up to 21.1%(FY-2D to FY-2F); the differences of the output reflectance from FY-2D and FY-2E calculated by the site experiment results reduced to 2.9%(13.6% when using the broadcast table). The normalized relative error was also calculated by the self-test method, which was less than 0.2%.
Post-processing through linear regression
NASA Astrophysics Data System (ADS)
van Schaeybroeck, B.; Vannitsem, S.
2011-03-01
Various post-processing techniques are compared for both deterministic and ensemble forecasts, all based on linear regression between forecast data and observations. In order to evaluate the quality of the regression methods, three criteria are proposed, related to the effective correction of forecast error, the optimal variability of the corrected forecast and multicollinearity. The regression schemes under consideration include the ordinary least-square (OLS) method, a new time-dependent Tikhonov regularization (TDTR) method, the total least-square method, a new geometric-mean regression (GM), a recently introduced error-in-variables (EVMOS) method and, finally, a "best member" OLS method. The advantages and drawbacks of each method are clarified. These techniques are applied in the context of the 63 Lorenz system, whose model version is affected by both initial condition and model errors. For short forecast lead times, the number and choice of predictors plays an important role. Contrarily to the other techniques, GM degrades when the number of predictors increases. At intermediate lead times, linear regression is unable to provide corrections to the forecast and can sometimes degrade the performance (GM and the best member OLS with noise). At long lead times the regression schemes (EVMOS, TDTR) which yield the correct variability and the largest correlation between ensemble error and spread, should be preferred.
Linear regression metamodeling as a tool to summarize and present simulation model results.
Jalal, Hawre; Dowd, Bryan; Sainfort, François; Kuntz, Karen M
2013-10-01
Modelers lack a tool to systematically and clearly present complex model results, including those from sensitivity analyses. The objective was to propose linear regression metamodeling as a tool to increase transparency of decision analytic models and better communicate their results. We used a simplified cancer cure model to demonstrate our approach. The model computed the lifetime cost and benefit of 3 treatment options for cancer patients. We simulated 10,000 cohorts in a probabilistic sensitivity analysis (PSA) and regressed the model outcomes on the standardized input parameter values in a set of regression analyses. We used the regression coefficients to describe measures of sensitivity analyses, including threshold and parameter sensitivity analyses. We also compared the results of the PSA to deterministic full-factorial and one-factor-at-a-time designs. The regression intercept represented the estimated base-case outcome, and the other coefficients described the relative parameter uncertainty in the model. We defined simple relationships that compute the average and incremental net benefit of each intervention. Metamodeling produced outputs similar to traditional deterministic 1-way or 2-way sensitivity analyses but was more reliable since it used all parameter values. Linear regression metamodeling is a simple, yet powerful, tool that can assist modelers in communicating model characteristics and sensitivity analyses.
Xue, Liang; Wang, Pengcheng; Wang, Lianshui; Renzi, Emily; Radivojac, Predrag; Tang, Haixu; Arnold, Randy; Zhu, Jian-Kang; Tao, W Andy
2013-08-01
Global phosphorylation changes in plants in response to environmental stress have been relatively poorly characterized to date. Here we introduce a novel mass spectrometry-based label-free quantitation method that facilitates systematic profiling plant phosphoproteome changes with high efficiency and accuracy. This method employs synthetic peptide libraries tailored specifically as internal standards for complex phosphopeptide samples and accordingly, a local normalization algorithm, LAXIC, which calculates phosphopeptide abundance normalized locally with co-eluting library peptides. Normalization was achieved in a small time frame centered to each phosphopeptide to compensate for the diverse ion suppression effect across retention time. The label-free LAXIC method was further treated with a linear regression function to accurately measure phosphoproteome responses to osmotic stress in Arabidopsis. Among 2027 unique phosphopeptides identified and 1850 quantified phosphopeptides in Arabidopsis samples, 468 regulated phosphopeptides representing 497 phosphosites have shown significant changes. Several known and novel components in the abiotic stress pathway were identified, illustrating the capability of this method to identify critical signaling events among dynamic and complex phosphorylation. Further assessment of those regulated proteins may help shed light on phosphorylation response to osmotic stress in plants.
Association between body mass index, diet and dental caries in Grade 6 boys in Medina, Saudi Arabia.
Bhayat, A; Ahmad, M S; Fadel, H T
2016-12-12
The prevalence of obesity is increasing in Saudi Arabia and although caries is associated with obesity, this association has not been investigated in Medina. This study aimed to determine the association between dental caries, body mass index (BMI) and dietary habits of 12-year-old boys from four geographically distinct schools in Medina. Mean BMI was 22.17 kg/m² (± 5.15); 41% had normal BMI, 25% were overweight and 30% were obese. The mean Decayed, Missing and Filled Teeth (DMFT) score was 1.46 (± 2.04). Those in the normal BMI range had a significantly higher prevalence of caries (57%) and DMFT score (1.92) compared with the overweight and obese groups (P < 0.05). These differences remained significant after controlling for possible confounders via linear regression. Mean BMI was significantly lower in boys with severe compared with mild or no caries. Normal and underweight participants had an almost 2 times greater risk of developing caries compared with their overweight and obese counterparts. The children had poor dietary habits and there were no significant associations between dietary variables and caries.
Aptel, Florent; Sayous, Romain; Fortoul, Vincent; Beccat, Sylvain; Denis, Philippe
2010-12-01
To evaluate and compare the regional relationships between visual field sensitivity and retinal nerve fiber layer (RNFL) thickness as measured by spectral-domain optical coherence tomography (OCT) and scanning laser polarimetry. Prospective cross-sectional study. One hundred and twenty eyes of 120 patients (40 with healthy eyes, 40 with suspected glaucoma, and 40 with glaucoma) were tested on Cirrus-OCT, GDx VCC, and standard automated perimetry. Raw data on RNFL thickness were extracted for 256 peripapillary sectors of 1.40625 degrees each for the OCT measurement ellipse and 64 peripapillary sectors of 5.625 degrees each for the GDx VCC measurement ellipse. Correlations between peripapillary RNFL thickness in 6 sectors and visual field sensitivity in the 6 corresponding areas were evaluated using linear and logarithmic regression analysis. Receiver operating curve areas were calculated for each instrument. With spectral-domain OCT, the correlations (r(2)) between RNFL thickness and visual field sensitivity ranged from 0.082 (nasal RNFL and corresponding visual field area, linear regression) to 0.726 (supratemporal RNFL and corresponding visual field area, logarithmic regression). By comparison, with GDx-VCC, the correlations ranged from 0.062 (temporal RNFL and corresponding visual field area, linear regression) to 0.362 (supratemporal RNFL and corresponding visual field area, logarithmic regression). In pairwise comparisons, these structure-function correlations were generally stronger with spectral-domain OCT than with GDx VCC and with logarithmic regression than with linear regression. The largest areas under the receiver operating curve were seen for OCT superior thickness (0.963 ± 0.022; P < .001) in eyes with glaucoma and for OCT average thickness (0.888 ± 0.072; P < .001) in eyes with suspected glaucoma. The structure-function relationship was significantly stronger with spectral-domain OCT than with scanning laser polarimetry, and was better expressed logarithmically than linearly. Measurements with these 2 instruments should not be considered to be interchangeable. Copyright © 2010 Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Rule, David L.
Several regression methods were examined within the framework of weighted structural regression (WSR), comparing their regression weight stability and score estimation accuracy in the presence of outlier contamination. The methods compared are: (1) ordinary least squares; (2) WSR ridge regression; (3) minimum risk regression; (4) minimum risk 2;…
Unit Cohesion and the Surface Navy: Does Cohesion Affect Performance
1989-12-01
v. 68, 1968. Neter, J., Wasserman, W., and Kutner, M. H., Applied Linear Regression Models, 2d ed., Boston, MA: Irwin, 1989. Rand Corporation R-2607...Neter, J., Wasserman, W., and Kutner, M. H., Applied Linear Regression Models, 2d ed., Boston, MA: Irwin, 1989. SAS User’s Guide: Basics, Version 5 ed
1990-03-01
and M.H. Knuter. Applied Linear Regression Models. Homewood IL: Richard D. Erwin Inc., 1983. Pritsker, A. Alan B. Introduction to Simulation and SLAM...Control Variates in Simulation," European Journal of Operational Research, 42: (1989). Neter, J., W. Wasserman, and M.H. Xnuter. Applied Linear Regression Models
ERIC Educational Resources Information Center
Yan, Jun; Aseltine, Robert H., Jr.; Harel, Ofer
2013-01-01
Comparing regression coefficients between models when one model is nested within another is of great practical interest when two explanations of a given phenomenon are specified as linear models. The statistical problem is whether the coefficients associated with a given set of covariates change significantly when other covariates are added into…
Calibrated Peer Review for Interpreting Linear Regression Parameters: Results from a Graduate Course
ERIC Educational Resources Information Center
Enders, Felicity B.; Jenkins, Sarah; Hoverman, Verna
2010-01-01
Biostatistics is traditionally a difficult subject for students to learn. While the mathematical aspects are challenging, it can also be demanding for students to learn the exact language to use to correctly interpret statistical results. In particular, correctly interpreting the parameters from linear regression is both a vital tool and a…
ERIC Educational Resources Information Center
Richter, Tobias
2006-01-01
Most reading time studies using naturalistic texts yield data sets characterized by a multilevel structure: Sentences (sentence level) are nested within persons (person level). In contrast to analysis of variance and multiple regression techniques, hierarchical linear models take the multilevel structure of reading time data into account. They…
Some Applied Research Concerns Using Multiple Linear Regression Analysis.
ERIC Educational Resources Information Center
Newman, Isadore; Fraas, John W.
The intention of this paper is to provide an overall reference on how a researcher can apply multiple linear regression in order to utilize the advantages that it has to offer. The advantages and some concerns expressed about the technique are examined. A number of practical ways by which researchers can deal with such concerns as…
ERIC Educational Resources Information Center
Nelson, Dean
2009-01-01
Following the Guidelines for Assessment and Instruction in Statistics Education (GAISE) recommendation to use real data, an example is presented in which simple linear regression is used to evaluate the effect of the Montreal Protocol on atmospheric concentration of chlorofluorocarbons. This simple set of data, obtained from a public archive, can…
Quantum State Tomography via Linear Regression Estimation
Qi, Bo; Hou, Zhibo; Li, Li; Dong, Daoyi; Xiang, Guoyong; Guo, Guangcan
2013-01-01
A simple yet efficient state reconstruction algorithm of linear regression estimation (LRE) is presented for quantum state tomography. In this method, quantum state reconstruction is converted into a parameter estimation problem of a linear regression model and the least-squares method is employed to estimate the unknown parameters. An asymptotic mean squared error (MSE) upper bound for all possible states to be estimated is given analytically, which depends explicitly upon the involved measurement bases. This analytical MSE upper bound can guide one to choose optimal measurement sets. The computational complexity of LRE is O(d4) where d is the dimension of the quantum state. Numerical examples show that LRE is much faster than maximum-likelihood estimation for quantum state tomography. PMID:24336519
NASA Astrophysics Data System (ADS)
Yadav, Shweta; Tandon, Ankit; Attri, Arun K.
2014-12-01
The detection of nicotine, an organic tracer for Environmental Tobacco Smoke (ETS), in the collected PM10 samples from Delhi region's ambient environment, in a appropriately designed investigation was initiated over four years (2006-2009) to: (1) Comprehend seasonal and inter-annual variations in the nicotine present in PM10; (2) Extract regression based linear trend profile manifested by nicotine in PM10; (3) Determine the non-linear trend timeline from the nicotine data, and compare it with the obtained linear trend; (4) Suggest the possible use of the designed experiment and analysis to have a qualitative appraisal of Tobacco Smoking activity in the sampling region. The PM10 samples were collected in a monthly time-series sequence at a known receptor site. Quantitative estimates of nicotine (ng m-3) were made by using a Thermal Desorption Gas Chromatography Mass Spectrometry (TD-GC/MS). The annual average concentrations of nicotine (ng m-3) were 516 ± 302 (2008) > 494 ± 301 (2009) > 438 ± 250 (2007) > 325 ± 149 (2006). The estimated linear trend of 5.4 ng m-3 month-1 corresponded to 16.3% per annum increase in the PM10 associated nicotine. The industrial production of India's tobacco index normalized to Delhi region's consumption, pegged an increase at 10.5% per annum over this period.
Applications of statistics to medical science, III. Correlation and regression.
Watanabe, Hiroshi
2012-01-01
In this third part of a series surveying medical statistics, the concepts of correlation and regression are reviewed. In particular, methods of linear regression and logistic regression are discussed. Arguments related to survival analysis will be made in a subsequent paper.
A phenomenological biological dose model for proton therapy based on linear energy transfer spectra.
Rørvik, Eivind; Thörnqvist, Sara; Stokkevåg, Camilla H; Dahle, Tordis J; Fjaera, Lars Fredrik; Ytre-Hauge, Kristian S
2017-06-01
The relative biological effectiveness (RBE) of protons varies with the radiation quality, quantified by the linear energy transfer (LET). Most phenomenological models employ a linear dependency of the dose-averaged LET (LET d ) to calculate the biological dose. However, several experiments have indicated a possible non-linear trend. Our aim was to investigate if biological dose models including non-linear LET dependencies should be considered, by introducing a LET spectrum based dose model. The RBE-LET relationship was investigated by fitting of polynomials from 1st to 5th degree to a database of 85 data points from aerobic in vitro experiments. We included both unweighted and weighted regression, the latter taking into account experimental uncertainties. Statistical testing was performed to decide whether higher degree polynomials provided better fits to the data as compared to lower degrees. The newly developed models were compared to three published LET d based models for a simulated spread out Bragg peak (SOBP) scenario. The statistical analysis of the weighted regression analysis favored a non-linear RBE-LET relationship, with the quartic polynomial found to best represent the experimental data (P = 0.010). The results of the unweighted regression analysis were on the borderline of statistical significance for non-linear functions (P = 0.053), and with the current database a linear dependency could not be rejected. For the SOBP scenario, the weighted non-linear model estimated a similar mean RBE value (1.14) compared to the three established models (1.13-1.17). The unweighted model calculated a considerably higher RBE value (1.22). The analysis indicated that non-linear models could give a better representation of the RBE-LET relationship. However, this is not decisive, as inclusion of the experimental uncertainties in the regression analysis had a significant impact on the determination and ranking of the models. As differences between the models were observed for the SOBP scenario, both non-linear LET spectrum- and linear LET d based models should be further evaluated in clinically realistic scenarios. © 2017 American Association of Physicists in Medicine.
Kwon, Junki; Choi, Jaewan; Shin, Joong Won; Lee, Jiyun; Kook, Michael S
2017-12-01
To assess the diagnostic ability of foveal avascular zone (FAZ) parameters to discriminate glaucomatous eyes with visual field defects (VFDs) in different locations (central vs. peripheral) from normal eyes. Totally, 125 participants were separated into 3 groups: normal (n=45), glaucoma with peripheral VFD (PVFD, n=45), and glaucoma with central VFD (CVFD, n=35). The FAZ area, perimeter, and circularity and parafoveal vessel density were calculated from optical coherence tomography angiography images. The diagnostic ability of the FAZ parameters and other structural parameters was determined according to glaucomatous VFD location. Associations between the FAZ parameters and central visual function were evaluated. A larger FAZ area and longer FAZ perimeter were observed in the CVFD group than in the PVFD and normal groups. The FAZ area, perimeter, and circularity were better in differentiating glaucomatous eyes with CVFDs from normal eyes [areas under the receiver operating characteristic curves (AUC), 0.78 to 0.88] than in differentiating PVFDs from normal eyes (AUC, 0.51 to 0.64). The FAZ perimeter had a similar AUC value to the circumpapillary retinal nerve fiber layer and macular ganglion cell-inner plexiform layer thickness for differentiating eyes with CVFDs from normal eyes (all P>0.05, the DeLong test). The FAZ area was significantly correlated with central visual function (β=-112.7, P=0.035, multivariate linear regression). The FAZ perimeter had good diagnostic capability in differentiating glaucomatous eyes with CVFDs from normal eyes, and may be a potential diagnostic biomarker for detecting glaucomatous patients with CVFDs.
Medenwald, Daniel; Swenne, Cees A; Frantz, Stefan; Nuding, Sebastian; Kors, Jan A; Pietzner, Diana; Tiller, Daniel; Greiser, Karin H; Kluttig, Alexander; Haerting, Johannes
2017-12-01
To assess the value of cardiac structure/function in predicting heart rate variability (HRV) and the possibly predictive value of HRV on cardiac parameters. Baseline and 4-year follow-up data from the population-based CARLA cohort were used (790 men, 646 women, aged 45-83 years at baseline and 50-87 years at follow-up). Echocardiographic and HRV recordings were performed at baseline and at follow-up. Linear regression models with a quadratic term were used. Crude and covariate adjusted estimates were calculated. Missing values were imputed by means of multiple imputation. Heart rate variability measures taken into account consisted of linear time and frequency domain [standard deviation of normal-to-normal intervals (SDNN), high-frequency power (HF), low-frequency power (LF), LF/HF ratio] and non-linear measures [detrended fluctuation analysis (DFA1), SD1, SD2, SD1/SD2 ratio]. Echocardiographic parameters considered were ventricular mass index, diastolic interventricular septum thickness, left ventricular diastolic dimension, left atrial dimension systolic (LADS), and ejection fraction (Teichholz). A negative quadratic relation between baseline LADS and change in SDNN and HF was observed. The maximum HF and SDNN change (an increase of roughly 0.02%) was predicted at LADS of 3.72 and 3.57 cm, respectively, while the majority of subjects experienced a decrease in HRV. There was no association between further echocardiographic parameters and change in HRV, and there was no evidence of a predictive value of HRV in the prediction of changes in cardiac structure. In the general population, LADS predicts 4-year alteration in SDNN and HF non-linearly. Because of the novelty of the result, analyses should be replicated in other populations. Published on behalf of the European Society of Cardiology. All rights reserved. © The Author 2017. For permissions please email: journals.permissions@oup.com.
Regression of non-linear coupling of noise in LIGO detectors
NASA Astrophysics Data System (ADS)
Da Silva Costa, C. F.; Billman, C.; Effler, A.; Klimenko, S.; Cheng, H.-P.
2018-03-01
In 2015, after their upgrade, the advanced Laser Interferometer Gravitational-Wave Observatory (LIGO) detectors started acquiring data. The effort to improve their sensitivity has never stopped since then. The goal to achieve design sensitivity is challenging. Environmental and instrumental noise couple to the detector output with different, linear and non-linear, coupling mechanisms. The noise regression method we use is based on the Wiener–Kolmogorov filter, which uses witness channels to make noise predictions. We present here how this method helped to determine complex non-linear noise couplings in the output mode cleaner and in the mirror suspension system of the LIGO detector.
García-Esquinas, Esther; Pérez-Gómez, Beatriz; Fernández, Mario Antonio; Pérez-Meixeira, Ana María; Gil, Elisa; de Paz, Concha; Iriso, Andrés; Sanz, Juan Carlos; Astray, Jenaro; Cisneros, Margot; de Santos, Amparo; Asensio, Angel; García-Sagredo, José Miguel; García, José Frutos; Vioque, Jesus; Pollán, Marina; López-Abente, Gonzalo; González, Maria José; Martínez, Mercedes; Bohigas, Pedro Arias; Pastor, Roberto; Aragonés, Nuria
2011-09-01
Although breastfeeding is the ideal way of nurturing infants, it can be a source of exposure to toxicants. This study reports the concentration of Hg, Pb and Cd in breast milk from a sample of women drawn from the general population of the Madrid Region, and explores the association between metal levels and socio-demographic factors, lifestyle habits, diet and environmental exposures, including tobacco smoke, exposure at home and occupational exposures. Breast milk was obtained from 100 women (20 mL) at around the third week postpartum. Pb, Cd and Hg levels were determined using Atomic Absorption Spectrometry. Metal levels were log-transformed due to non-normal distribution. Their association with the variables collected by questionnaire was assessed using linear regression models. Separate models were fitted for Hg, Pb and Cd, using univariate linear regression in a first step. Secondly, multivariate linear regression models were adjusted introducing potential confounders specific for each metal. Finally, a test for trend was performed in order to evaluate possible dose-response relationships between metal levels and changes in variables categories. Geometric mean Hg, Pb and Cd content in milk were 0.53 μg L(-1), 15.56 μg L(-1), and 1.31 μg L(-1), respectively. Decreases in Hg levels in older women and in those with a previous history of pregnancies and lactations suggested clearance of this metal over lifetime, though differences were not statistically significant, probably due to limited sample size. Lead concentrations increased with greater exposure to motor vehicle traffic and higher potato consumption. Increased Cd levels were associated with type of lactation and tended to increase with tobacco smoking. Surveillance for the presence of heavy metals in human milk is needed. Smoking and dietary habits are the main factors linked to heavy metal levels in breast milk. Our results reinforce the need to strengthen national food safety programs and to further promote avoidance of unhealthy behaviors such as smoking during pregnancy. Copyright © 2011 Elsevier Ltd. All rights reserved.
Hu, Rongrong; Wang, Chenkun; Gu, Yangshun; Racette, Lyne
2016-01-01
Abstract Detection of progression is paramount to the clinical management of glaucoma. Our goal is to compare the performance of standard automated perimetry (SAP), short-wavelength automated perimetry (SWAP), and frequency-doubling technology (FDT) perimetry in monitoring glaucoma progression. Longitudinal data of paired SAP, SWAP, and FDT from 113 eyes with primary open-angle glaucoma enrolled in the Diagnostic Innovations in Glaucoma Study or the African Descent and Glaucoma Evaluation Study were included. Data from all tests were expressed in comparable units by converting the sensitivity from decibels to unitless contrast sensitivity and by expressing sensitivity values in percent of mean normal based on an independent dataset of 207 healthy eyes with aging deterioration taken into consideration. Pointwise linear regression analysis was performed and 3 criteria (conservative, moderate, and liberal) were used to define progression and improvement. Global mean sensitivity (MS) was fitted with linear mixed models. No statistically significant difference in the proportion of progressing and improving eyes was observed across tests using the conservative criterion. Fewer eyes showed improvement on SAP compared to SWAP and FDT using the moderate criterion; and FDT detected less progressing eyes than SAP and SWAP using the liberal criterion. The agreement between these test types was poor. The linear mixed model showed a progressing trend of global MS overtime for SAP and SWAP, but not for FDT. The baseline estimate of SWAP MS was significantly lower than SAP MS by 21.59% of mean normal. FDT showed comparable estimation of baseline MS with SAP. SWAP and FDT do not appear to have significant benefits over SAP in monitoring glaucoma progression. SAP, SWAP, and FDT may, however, detect progression in different glaucoma eyes. PMID:26886602
Goodarzi, Mohammad; Jensen, Richard; Vander Heyden, Yvan
2012-12-01
A Quantitative Structure-Retention Relationship (QSRR) is proposed to estimate the chromatographic retention of 83 diverse drugs on a Unisphere poly butadiene (PBD) column, using isocratic elutions at pH 11.7. Previous work has generated QSRR models for them using Classification And Regression Trees (CART). In this work, Ant Colony Optimization is used as a feature selection method to find the best molecular descriptors from a large pool. In addition, several other selection methods have been applied, such as Genetic Algorithms, Stepwise Regression and the Relief method, not only to evaluate Ant Colony Optimization as a feature selection method but also to investigate its ability to find the important descriptors in QSRR. Multiple Linear Regression (MLR) and Support Vector Machines (SVMs) were applied as linear and nonlinear regression methods, respectively, giving excellent correlation between the experimental, i.e. extrapolated to a mobile phase consisting of pure water, and predicted logarithms of the retention factors of the drugs (logk(w)). The overall best model was the SVM one built using descriptors selected by ACO. Copyright © 2012 Elsevier B.V. All rights reserved.
Not Quite Normal: Consequences of Violating the Assumption of Normality in Regression Mixture Models
ERIC Educational Resources Information Center
Van Horn, M. Lee; Smith, Jessalyn; Fagan, Abigail A.; Jaki, Thomas; Feaster, Daniel J.; Masyn, Katherine; Hawkins, J. David; Howe, George
2012-01-01
Regression mixture models, which have only recently begun to be used in applied research, are a new approach for finding differential effects. This approach comes at the cost of the assumption that error terms are normally distributed within classes. This study uses Monte Carlo simulations to explore the effects of relatively minor violations of…
Source localization in an ocean waveguide using supervised machine learning.
Niu, Haiqiang; Reeves, Emma; Gerstoft, Peter
2017-09-01
Source localization in ocean acoustics is posed as a machine learning problem in which data-driven methods learn source ranges directly from observed acoustic data. The pressure received by a vertical linear array is preprocessed by constructing a normalized sample covariance matrix and used as the input for three machine learning methods: feed-forward neural networks (FNN), support vector machines (SVM), and random forests (RF). The range estimation problem is solved both as a classification problem and as a regression problem by these three machine learning algorithms. The results of range estimation for the Noise09 experiment are compared for FNN, SVM, RF, and conventional matched-field processing and demonstrate the potential of machine learning for underwater source localization.
NASA Technical Reports Server (NTRS)
Dejesusparada, N. (Principal Investigator); Verdesio, J. J.
1981-01-01
The relationship existing between Guanabara Bay water quality ground truth parameters and LANDSAT MSS video data was investigated. The parameters considered were: chorophyll content, water transparency usng the Secchi disk, salinity, and dissolved ammonia. Data from two overflights was used, and methods of processing digital data were compared. Linear and nonlinear regression analyses were utilized, comparing original data with processed data by using the correlation coefficient and the estimation mean error. It was determined that better quality data are obtained by using radiometric correction programs with a physical basis, contrast ratio, and normalization. Incidental locations of floating vegetation, changes in bottom depth, oil slicks, and ships at anchor were made.
Krell-Roesch, Janina; Ruider, Hanna; Lowe, Val J; Stokin, Gorazd B; Pink, Anna; Roberts, Rosebud O; Mielke, Michelle M; Knopman, David S; Christianson, Teresa J; Machulda, Mary M; Jack, Clifford R; Petersen, Ronald C; Geda, Yonas E
2016-07-14
One of the key research agenda of the field of aging is investigation of presymptomatic Alzheimer's disease (AD). Furthermore, abnormalities in brain glucose metabolism (as measured by FDG-PET) have been reported among cognitively normal elderly persons. However, little is known about the association of FDG-PET abnormalities with neuropsychiatric symptoms (NPS) in a population-based setting. Thus, we conducted a cross-sectional study derived from the ongoing population-based Mayo Clinic Study of Aging in order to examine the association between brain glucose metabolism and NPS among cognitively normal (CN) persons aged > 70 years. Participants underwent FDG-PET and completed the Neuropsychiatric Inventory Questionnaire (NPI-Q), Beck Depression Inventory (BDI), and Beck Anxiety Inventory (BAI). Cognitive classification was made by an expert consensus panel. We conducted multivariable logistic regression analyses to compute odds ratios (OR) and 95% confidence intervals after adjusting for age, sex, and education. For continuous variables, we used linear regression and Spearman rank-order correlations. Of 668 CN participants (median 78.1 years, 55.4% males), 205 had an abnormal FDG-PET (i.e., standardized uptake value ratio < 1.32 in AD-related regions). Abnormal FDG-PET was associated with depression as measured by NPI-Q (OR = 2.12; 1.23-3.64); the point estimate was further elevated for APOE ɛ4 carriers (OR = 2.59; 1.00-6.69), though marginally significant. Additionally, we observed a significant association between abnormal FDG-PET and depressive and anxiety symptoms when treated as continuous measures. These findings indicate that NPS, even in community-based samples, can be an important additional tool to the biomarker-based investigation of presymptomatic AD.
Morphometric body condition indices of wild Florida manatees (Trichechus manatus latirostris)
Harshaw, Lauren T.; Larkin, Iskande V.; Bonde, Robert K.; Deutsch, Charles J.; Hill, Richard C.
2016-01-01
In many species, body weight (W) increases geometrically with body length (L), so W/L3 provides a body condition index (BCI) that can be used to evaluate nutritional status once a normal range has been established. No such index has been established for Florida manatees (Trichechus manatus latirostris). This study was designed to determine a normal range of BCIs of Florida manatees by comparing W in kg with straight total length (SL), curvilinear total length (CL), and umbilical girth (UG) in m for 146 wild manatees measured during winter health assessments at three Florida locations. Small calves to large adults of SL from 1.47 to 3.23 m and W from 77 to 751 kg were compared. BCIs were significantly greater in adult females than in adult males (p < 0.05). W scaled proportionally to L3 in females but not in males, which were slimmer than females. The logarithms of W and of each linear measurement were regressed to develop amended indices that allow for sex differences. The regression slope for log W against log SL was 2.915 in females and 2.578 in males; W/SL2.915 ranged from 18.9 to 29.6 (mean 23.2) in females and from 24.6 to 37.3 (mean 29.8) in males. Some BCIs were slightly (4%), but significantly (p ≤ 0.05), higher for females in Crystal River than in Tampa Bay or Indian River, but there was no evidence of geographic variation in condition among males. These normal ranges should help evaluate the nutritional status of both wild and rehabilitating captive manatees.
Long-term changes in body weight are associated with changes in blood pressure levels.
Markus, M R P; Ittermann, T; Baumeister, S E; Troitzsch, P; Schipf, S; Lorbeer, R; Aumannn, N; Wallaschofski, H; Dörr, M; Rettig, R; Völzke, H
2015-03-01
Hypertension and obesity are highly prevalent in Western societies. We investigated the associations of changes in body weight with changes in blood pressure and with incident hypertension, incident cardiovascular events, or incident normalization of blood pressure in patients who were hypertensive at baseline, over a 5-year period. Data of men and women aged 20-81 years of the Study of Health in Pomerania were used. Changes in body weight were related to changes in blood pressure by linear regression (n = 1875) adjusted for cofounders. Incident hypertension, incident cardiovascular events, or incident blood pressure normalization in patients who were hypertensive at baseline were investigated using Poisson regression (n = 3280) models. A change of 1 kg in body weight was positively associated with a change of 0.45 mm Hg (95% confidence interval (CI): 0.34-0.55 mm Hg) in systolic blood pressure, 0.32 mm Hg (95% CI: 0.25-0.38 mm Hg) in diastolic blood pressure, and 0.36 mm Hg (95% CI: 0.29-0.43 mm Hg) in mean arterial pressure (all p-values <0.001). A 5% weight loss reduced the relative risk (RR) of incident hypertension (RRs 0.84 (95% CI: 0.79-0.89)) and incident cardiovascular events (RRs 0.81 (95% CI: 0.68-0.98)) and increased the chance of incident blood pressure normalization in patients who were hypertensive at baseline by 15% (95% CI: 7-23%). Absolute and relative changes in body weight are positively associated with changes in blood pressure levels and also affect the risk of cardiovascular events. Copyright © 2014 Elsevier B.V. All rights reserved.
Study on the social adaptation of Chinese children with down syndrome.
Wang, Yan-Xia; Mao, Shan-Shan; Xie, Chun-Hong; Qin, Yu-Feng; Zhu, Zhi-Wei; Zhan, Jian-Ying; Shao, Jie; Li, Rong; Zhao, Zheng-Yan
2007-06-30
To evaluate social adjustment and related factors among Chinese children with Down syndrome (DS). A structured interview and Peabody Picture Vocabulary Test (PPVT) were conducted with a group of 36 DS children with a mean age of 106.28 months, a group of 30 normally-developing children matched for mental age (MA) and a group of 40 normally-developing children matched for chronological age (CA). Mean scores of social adjustment were compared between the three groups, and partial correlations and stepwise multiple regression models were used to further explore related factors. There was no difference between the DS group and the MA group in terms of communication skills. However, the DS group scored much better than the MA group in self-dependence, locomotion, work skills, socialization and self-management. Children in the CA group achieved significantly higher scores in all aspects of social adjustment than the DS children. Partial correlations indicate a relationship between social adjustment and the PPVT raw score and also between social adjustment and age (significant r ranging between 0.24 and 0.92). A stepwise linear regression analysis showed that family structure was the main predictor of social adjustment. Newborn history was also a predictor of work skills, communication, socialization and self-management. Parental education was found to account for 8% of self-dependence. Maternal education explained 6% of the variation in locomotion. Although limited by the small sample size, these results indicate that Chinese DS children have better social adjustment skills when compared to their mental-age-matched normally-developing peers, but that the Chinese DS children showed aspects of adaptive development that differed from Western DS children. Analyses of factors related to social adjustment suggest that effective early intervention may improve social adaptability.
Impact of facial defect reconstruction on attractiveness and negative facial perception.
Dey, Jacob K; Ishii, Masaru; Boahene, Kofi D O; Byrne, Patrick; Ishii, Lisa E
2015-06-01
Measure the impact of facial defect reconstruction on observer-graded attractiveness and negative facial perception. Prospective, randomized, controlled experiment. One hundred twenty casual observers viewed images of faces with defects of varying sizes and locations before and after reconstruction as well as normal comparison faces. Observers rated attractiveness, defect severity, and how disfiguring, bothersome, and important to repair they considered each face. Facial defects decreased attractiveness -2.26 (95% confidence interval [CI]: -2.45, -2.08) on a 10-point scale. Mixed effects linear regression showed this attractiveness penalty varied with defect size and location, with large and central defects generating the greatest penalty. Reconstructive surgery increased attractiveness 1.33 (95% CI: 1.18, 1.47), an improvement dependent upon size and location, restoring some defect categories to near normal ranges of attractiveness. Iterated principal factor analysis indicated the disfiguring, important to repair, bothersome, and severity variables were highly correlated and measured a common domain; thus, they were combined to create the disfigured, important to repair, bothersome, severity (DIBS) factor score, representing negative facial perception. The DIBS regression showed defect faces have a 1.5 standard deviation increase in negative perception (DIBS: 1.69, 95% CI: 1.61, 1.77) compared to normal faces, which decreased by a similar magnitude after surgery (DIBS: -1.44, 95% CI: -1.49, -1.38). These findings varied with defect size and location. Surgical reconstruction of facial defects increased attractiveness and decreased negative social facial perception, an impact that varied with defect size and location. These new social perception data add to the evidence base demonstrating the value of high-quality reconstructive surgery. NA. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.
Body mass index and motor coordination: Non-linear relationships in children 6-10 years.
Lopes, V P; Malina, R M; Maia, J A R; Rodrigues, L P
2018-05-01
Given the concern for health-related consequences of an elevated body mass index (BMI; obesity), the potential consequences of a low BMI in children are often overlooked. The purpose was to evaluate the relationship between the BMI across its entire spectrum and motor coordination (MC) in children 6-10 years. Height, weight, and MC (Körperkoordinationstest für Kinder, KTK test battery) were measured in 1,912 boys and 1,826 girls of 6-10 years of age. BMI (kg/m 2 ) was calculated. KTK scores for each of the four tests were also converted to a motor quotient (MQ). One-way ANOVA was used to test differences in the BMI, individual test items, and MQ among boys and girls within age groups. Sex-specific quadratic regressions of individual KTK items and the MQ on the BMI were calculated. Girls and boys were also classified into four weight status groups using International Obesity Task Force criteria: thin, normal, overweight, and obese. Differences in specific test items and MQ between weight status groups were evaluated by age group in each sex. Thirty-one percent of the sample was overweight or obese, whereas 5% was thin. On average, normal weight children had the highest MQ in both sexes across the age range with few exceptions. Overweight/obese children had a lower MQ than normal weight and thin children. The quadratic regression lines generally presented an inverted parabolic relationship between the BMI and MC and suggested a decrease in MC with an increase in the BMI. In general, BMI shows a curvilinear, inverted parabolic relationship with MC in children 6-10 years. © 2018 John Wiley & Sons Ltd.
Allum, John H J; Cleworth, T; Honegger, Flurin
2016-07-01
We investigated how response asymmetries and deficit side response amplitudes for head accelerations used clinically to test the vestibular ocular reflex (VOR) are correlated with caloric canal paresis (CP) values. 30 patients were examined at onset of an acute unilateral peripheral vestibular deficit (aUPVD) and 3, 6, and 13 weeks later with three different VOR tests: caloric, rotating chair (ROT), and video head impulse tests (vHIT). Response changes over time were fitted with an exponential decay model and compared with using linear regression analysis. Recovery times (to within 10% of steady state) were similar for vHIT-asymmetry and CP (>10 weeks) but shorter for ROT asymmetry (<4 weeks). Regressions with CP were similar (vHIT asymmetry, R = 0.68, ROT, R = 0.62). Responses to the deficit side were also equally well correlated with CP values (R = 0.71). Specificity for vHIT and 20 degrees/s ROT deficit side responses was 100% in comparison to CP values, sensitivity was 74% for vHIT, 75% for ROT. A decrease in normal side responses occurred for ROT but not for vHIT at 3 weeks. Normal side responses were weekly correlated with CP for ROT (R = 0.49) but not for vHIT (R = 0.17). These results indicate that vHIT deficit side VOR gains are slightly better correlated with CP values than ROT, probably because of similar recovery time courses of vHIT and caloric responses and the lack of normal side vHIT changes. However, specificity and sensitivity is the same for vHIT and ROT tests.
Evaluating Differential Effects Using Regression Interactions and Regression Mixture Models
ERIC Educational Resources Information Center
Van Horn, M. Lee; Jaki, Thomas; Masyn, Katherine; Howe, George; Feaster, Daniel J.; Lamont, Andrea E.; George, Melissa R. W.; Kim, Minjung
2015-01-01
Research increasingly emphasizes understanding differential effects. This article focuses on understanding regression mixture models, which are relatively new statistical methods for assessing differential effects by comparing results to using an interactive term in linear regression. The research questions which each model answers, their…
Recent work on material interface reconstruction
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mosso, S.J.; Swartz, B.K.
1997-12-31
For the last 15 years, many Eulerian codes have relied on a series of piecewise linear interface reconstruction algorithms developed by David Youngs. In a typical Youngs` method, the material interfaces were reconstructed based upon nearly cell values of volume fractions of each material. The interfaces were locally represented by linear segments in two dimensions and by pieces of planes in three dimensions. The first step in such reconstruction was to locally approximate an interface normal. In Youngs` 3D method, a local gradient of a cell-volume-fraction function was estimated and taken to be the local interface normal. A linear interfacemore » was moved perpendicular to the now known normal until the mass behind it matched the material volume fraction for the cell in question. But for distorted or nonorthogonal meshes, the gradient normal estimate didn`t accurately match that of linear material interfaces. Moreover, curved material interfaces were also poorly represented. The authors will present some recent work in the computation of more accurate interface normals, without necessarily increasing stencil size. Their estimate of the normal is made using an iterative process that, given mass fractions for nearby cells of known but arbitrary variable density, converges in 3 or 4 passes in practice (and quadratically--like Newton`s method--in principle). The method reproduces a linear interface in both orthogonal and nonorthogonal meshes. The local linear approximation is generally 2nd-order accurate, with a 1st-order accurate normal for curved interfaces in both two and three dimensional polyhedral meshes. Recent work demonstrating the interface reconstruction for curved surfaces will /be discussed.« less
SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES
Zhu, Liping; Huang, Mian; Li, Runze
2012-01-01
This paper is concerned with quantile regression for a semiparametric regression model, in which both the conditional mean and conditional variance function of the response given the covariates admit a single-index structure. This semiparametric regression model enables us to reduce the dimension of the covariates and simultaneously retains the flexibility of nonparametric regression. Under mild conditions, we show that the simple linear quantile regression offers a consistent estimate of the index parameter vector. This is a surprising and interesting result because the single-index model is possibly misspecified under the linear quantile regression. With a root-n consistent estimate of the index vector, one may employ a local polynomial regression technique to estimate the conditional quantile function. This procedure is computationally efficient, which is very appealing in high-dimensional data analysis. We show that the resulting estimator of the quantile function performs asymptotically as efficiently as if the true value of the index vector were known. The methodologies are demonstrated through comprehensive simulation studies and an application to a real dataset. PMID:24501536
Prediction of siRNA potency using sparse logistic regression.
Hu, Wei; Hu, John
2014-06-01
RNA interference (RNAi) can modulate gene expression at post-transcriptional as well as transcriptional levels. Short interfering RNA (siRNA) serves as a trigger for the RNAi gene inhibition mechanism, and therefore is a crucial intermediate step in RNAi. There have been extensive studies to identify the sequence characteristics of potent siRNAs. One such study built a linear model using LASSO (Least Absolute Shrinkage and Selection Operator) to measure the contribution of each siRNA sequence feature. This model is simple and interpretable, but it requires a large number of nonzero weights. We have introduced a novel technique, sparse logistic regression, to build a linear model using single-position specific nucleotide compositions which has the same prediction accuracy of the linear model based on LASSO. The weights in our new model share the same general trend as those in the previous model, but have only 25 nonzero weights out of a total 84 weights, a 54% reduction compared to the previous model. Contrary to the linear model based on LASSO, our model suggests that only a few positions are influential on the efficacy of the siRNA, which are the 5' and 3' ends and the seed region of siRNA sequences. We also employed sparse logistic regression to build a linear model using dual-position specific nucleotide compositions, a task LASSO is not able to accomplish well due to its high dimensional nature. Our results demonstrate the superiority of sparse logistic regression as a technique for both feature selection and regression over LASSO in the context of siRNA design.
Predictive and mechanistic multivariate linear regression models for reaction development
Santiago, Celine B.; Guo, Jing-Yao
2018-01-01
Multivariate Linear Regression (MLR) models utilizing computationally-derived and empirically-derived physical organic molecular descriptors are described in this review. Several reports demonstrating the effectiveness of this methodological approach towards reaction optimization and mechanistic interrogation are discussed. A detailed protocol to access quantitative and predictive MLR models is provided as a guide for model development and parameter analysis. PMID:29719711
Adding a Parameter Increases the Variance of an Estimated Regression Function
ERIC Educational Resources Information Center
Withers, Christopher S.; Nadarajah, Saralees
2011-01-01
The linear regression model is one of the most popular models in statistics. It is also one of the simplest models in statistics. It has received applications in almost every area of science, engineering and medicine. In this article, the authors show that adding a predictor to a linear model increases the variance of the estimated regression…
Using nonlinear quantile regression to estimate the self-thinning boundary curve
Quang V. Cao; Thomas J. Dean
2015-01-01
The relationship between tree size (quadratic mean diameter) and tree density (number of trees per unit area) has been a topic of research and discussion for many decades. Starting with Reineke in 1933, the maximum size-density relationship, on a log-log scale, has been assumed to be linear. Several techniques, including linear quantile regression, have been employed...
Simultaneous spectrophotometric determination of salbutamol and bromhexine in tablets.
Habib, I H I; Hassouna, M E M; Zaki, G A
2005-03-01
Typical anti-mucolytic drugs called salbutamol hydrochloride and bromhexine sulfate encountered in tablets were determined simultaneously either by using linear regression at zero-crossing wavelengths of the first derivation of UV-spectra or by application of multiple linear partial least squares regression method. The results obtained by the two proposed mathematical methods were compared with those obtained by the HPLC technique.
Laurens, L M L; Wolfrum, E J
2013-12-18
One of the challenges associated with microalgal biomass characterization and the comparison of microalgal strains and conversion processes is the rapid determination of the composition of algae. We have developed and applied a high-throughput screening technology based on near-infrared (NIR) spectroscopy for the rapid and accurate determination of algal biomass composition. We show that NIR spectroscopy can accurately predict the full composition using multivariate linear regression analysis of varying lipid, protein, and carbohydrate content of algal biomass samples from three strains. We also demonstrate a high quality of predictions of an independent validation set. A high-throughput 96-well configuration for spectroscopy gives equally good prediction relative to a ring-cup configuration, and thus, spectra can be obtained from as little as 10-20 mg of material. We found that lipids exhibit a dominant, distinct, and unique fingerprint in the NIR spectrum that allows for the use of single and multiple linear regression of respective wavelengths for the prediction of the biomass lipid content. This is not the case for carbohydrate and protein content, and thus, the use of multivariate statistical modeling approaches remains necessary.
Zhang, Xin; Liu, Pan; Chen, Yuguang; Bai, Lu; Wang, Wei
2014-01-01
The primary objective of this study was to identify whether the frequency of traffic conflicts at signalized intersections can be modeled. The opposing left-turn conflicts were selected for the development of conflict predictive models. Using data collected at 30 approaches at 20 signalized intersections, the underlying distributions of the conflicts under different traffic conditions were examined. Different conflict-predictive models were developed to relate the frequency of opposing left-turn conflicts to various explanatory variables. The models considered include a linear regression model, a negative binomial model, and separate models developed for four traffic scenarios. The prediction performance of different models was compared. The frequency of traffic conflicts follows a negative binominal distribution. The linear regression model is not appropriate for the conflict frequency data. In addition, drivers behaved differently under different traffic conditions. Accordingly, the effects of conflicting traffic volumes on conflict frequency vary across different traffic conditions. The occurrences of traffic conflicts at signalized intersections can be modeled using generalized linear regression models. The use of conflict predictive models has potential to expand the uses of surrogate safety measures in safety estimation and evaluation.
Methyl-coenzyme M reductase A as an indicator to estimate methane production from dairy cows.
Aguinaga Casañas, M A; Rangkasenee, N; Krattenmacher, N; Thaller, G; Metges, C C; Kuhla, B
2015-06-01
The evaluation of greenhouse gas mitigation strategies requires the quantitative assessment of individual methane production. Because methane measurement in respiration chambers is highly accurate, but also comprises various disadvantages such as limited capacity and high costs, the establishment of an indicator for estimating methane production of individual ruminants would provide an alternative to direct methane measurement. Methyl-coenzyme M reductase is involved in methanogenesis and the subunit α of methyl-coenzyme M reductase is encoded by the mcrA gene of rumen archaea. We therefore examined the relationship between methane emissions of Holstein dairy cows measured in respiration chambers with 2 different diets (high- and medium-concentrate diet) and the mcrA DNA and mcrA cDNA abundance determined from corresponding rumen fluid samples. Whole-body methane production per kilogram of dry matter intake and mcrA DNA normalized to the abundance of the rrs gene coding for 16S rRNA correlated significantly when using qmcrA primers. Use of qmcrA primers also revealed linear correlation between mcrA DNA copy number and methane yield. Regression analyses based on normalized mcrA cDNA abundances revealed no significant linear correlation with methane production per kilogram of dry matter intake. Furthermore, the correlations between normalized mcrA DNA abundance and the rumen fluid concentration of acetic and isobutyric acid were positive, whereas the correlations with propionic and lactic acid were negative. These data suggest that the mcrA DNA approach based on qmcrA primers could potentially be a molecular proxy for methane yield after further refinement. Copyright © 2015 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.
Mao, Yuanyuan; Hu, Wenbin; Liu, Qin; Liu, Li; Li, Yuanming; Shen, Yueping
2015-08-01
To examine the dose-response relationship between gestational weight gain rate and the neonate birth weight. A total of 18 868 women with singleton gestations who delivered between January 2006 and December 2013 were included in this study. Maternal and neonate details of these women were drawn from the Perinatal Monitoring System database. Gestational weight gain rate was defined as the total weight gain during the last and first prenatal care visits divided by the interval weeks. Both Multiple logistic regression analysis and restricted cubic spline methods were performed. Confounding factors included maternal age, education, pre-pregnancy body mass index (BMI), state of residence, parity, gestational weeks of prenatal care entry, and sex of the neonate. The adjusted odds ratio for macrosomia was associated with gestational weight gain rate in lower pre-pregnancy BMI (OR = 3.15, 95% CI: 1.40-7.07), normal (OR = 3.64, 95% CI: 2.84-4.66) or overweight (OR = 2.37, 95% CI: 1.71-3.27). The odds ratios of low birth weight appeared a decrease in those women with lower pre-pregnancy BMI (OR = 0.28, 95% CI: 0.13-0.61) while the normal weight (OR = 0.37, 95% CI: 0.22-0.64) group with gestational weight gain, the rate showed an increase. Association of gestational weight gain rate for macrosomia was found a S-curve in those term delivery women (non-linearity test P < 0.000 1). However, L-curve was observed for low birth weight and gestational weight gain rate in term births (non-linearity test P < 0.000 1). A S-curve was seen between gestational weight gain rate and term delivered macrosomia while L-curve was observed among term delivered low birth weight neonates.
Standards for Standardized Logistic Regression Coefficients
ERIC Educational Resources Information Center
Menard, Scott
2011-01-01
Standardized coefficients in logistic regression analysis have the same utility as standardized coefficients in linear regression analysis. Although there has been no consensus on the best way to construct standardized logistic regression coefficients, there is now sufficient evidence to suggest a single best approach to the construction of a…
Image interpolation via regularized local linear regression.
Liu, Xianming; Zhao, Debin; Xiong, Ruiqin; Ma, Siwei; Gao, Wen; Sun, Huifang
2011-12-01
The linear regression model is a very attractive tool to design effective image interpolation schemes. Some regression-based image interpolation algorithms have been proposed in the literature, in which the objective functions are optimized by ordinary least squares (OLS). However, it is shown that interpolation with OLS may have some undesirable properties from a robustness point of view: even small amounts of outliers can dramatically affect the estimates. To address these issues, in this paper we propose a novel image interpolation algorithm based on regularized local linear regression (RLLR). Starting with the linear regression model where we replace the OLS error norm with the moving least squares (MLS) error norm leads to a robust estimator of local image structure. To keep the solution stable and avoid overfitting, we incorporate the l(2)-norm as the estimator complexity penalty. Moreover, motivated by recent progress on manifold-based semi-supervised learning, we explicitly consider the intrinsic manifold structure by making use of both measured and unmeasured data points. Specifically, our framework incorporates the geometric structure of the marginal probability distribution induced by unmeasured samples as an additional local smoothness preserving constraint. The optimal model parameters can be obtained with a closed-form solution by solving a convex optimization problem. Experimental results on benchmark test images demonstrate that the proposed method achieves very competitive performance with the state-of-the-art interpolation algorithms, especially in image edge structure preservation. © 2011 IEEE
Rosa, Erica Carine Campos Caldas; Dos Santos, Renan Renato Cruz; Fernandes, Luis Fernando Amarante; Neves, Francisco de Assis Rocha; Coelho, Michella Soares; Amato, Angelica Amorim
2018-01-01
We investigated leukocyte relative telomere length (TL) in patients with type 2 diabetes (T2D) diagnosed for no longer than five years and its association with clinical and biochemical variables. Peripheral blood leukocyte relative TL was investigated in 108 patients with T2D (87 women, 21 men) and 125 (37 women, 88 men) age-matched control subjects with normal glucose tolerance, by quantitative polymerase chain reaction. Multiple linear regression analysis was used to examine the association between relative TL and demographic, anthropometric and biochemical indicators of metabolic control among patients with T2D. Patients with T2D had a median time since diagnosis of 1 year and most were on metformin monotherapy, with satisfactory glucose control determined by HbA1c levels. Median relative TL was not different between patients with T2D and control subjects. However, multiple linear regression analyses showed that relative TL was inversely associated with time since T2D diagnosis, fasting plasma glucose levels and HbA1c levels, but not with HbA1c levels assessed in the preceding 5-12 months, after adjustment for age, sex and body mass index. This study suggests that relative TL is not shorter in patients with recently diagnosed T2D, but is inversely correlated with glucose levels, even among patients with overall satisfactory glucose control. Copyright © 2017 Elsevier B.V. All rights reserved.
Tisell, Anders; Leinhard, Olof Dahlqvist; Warntjes, Jan Bertus Marcel; Aalto, Anne; Smedby, Örjan; Landtblom, Anne-Marie; Lundberg, Peter
2013-01-01
In Multiple Sclerosis (MS) the relationship between disease process in normal-appearing white matter (NAWM) and the development of white matter lesions is not well understood. In this study we used single voxel proton ‘Quantitative Magnetic Resonance Spectroscopy’ (qMRS) to characterize the NAWM and thalamus both in atypical ‘Clinically Definite MS’ (CDMS) patients, MRIneg (N = 15) with very few lesions (two or fewer lesions), and in typical CDMS patients, MRIpos (N = 20) with lesions, in comparison with healthy control subjects (N = 20). In addition, the metabolite concentrations were also correlated with extent of brain atrophy measured using Brain Parenchymal Fraction (BPF) and severity of the disease measured using ‘Multiple Sclerosis Severity Score’ (MSSS). Elevated concentrations of glutamate and glutamine (Glx) were observed in both MS groups (MRIneg 8.12 mM, p<0.001 and MRIpos 7.96 mM p<0.001) compared to controls, 6.76 mM. Linear regressions of Glx and total creatine (tCr) with MSSS were 0.16±0.06 mM/MSSS (p = 0.02) for Glx and 0.06±0.03 mM/MSSS (p = 0.04) for tCr, respectively. Moreover, linear regressions of tCr and myo-Inositol (mIns) with BPF were −6.22±1.63 mM/BPF (p<0.001) for tCr and −7.71±2.43 mM/BPF (p = 0.003) for mIns. Furthermore, the MRIpos patients had lower N-acetylaspartate and N-acetylaspartate-glutamate (tNA) and elevated mIns concentrations in NAWM compared to both controls (tNA: p = 0.04 mIns p<0.001) and MRIneg (tNA: p = 0.03 , mIns: p = 0.002). The results suggest that Glx may be an important marker for pathology in non-lesional white matter in MS. Moreover, Glx is related to the severity of MS independent of number of lesions in the patient. In contrast, increased glial density indicated by increased mIns and decreased neuronal density indicated by the decreased tNA, were only observed in NAWM of typical CDMS patients with white matter lesions. PMID:23613944
NASA Astrophysics Data System (ADS)
Bartiko, Daniel; Chaffe, Pedro; Bonumá, Nadia
2017-04-01
Floods may be strongly affected by climate, land-use, land-cover and water infrastructure changes. However, it is common to model this process as stationary. This approach has been questioned, especially when it involves estimate of the frequency and magnitude of extreme events for designing and maintaining hydraulic structures, as those responsible for flood control and dams safety. Brazil is the third largest producer of hydroelectricity in the world and many of the country's dams are located in the Southern Region. So, it seems appropriate to investigate the presence of non-stationarity in the affluence in these plants. In our study, we used historical flood data from the Brazilian National Grid Operator (ONS) to explore trends in annual maxima in river flow of the 38 main rivers flowing to Southern Brazilian reservoirs (records range from 43 to 84 years). In the analysis, we assumed a two-parameter log-normal distribution a linear regression model was applied in order to allow for the mean to vary with time. We computed recurrence reduction factors to characterize changes in the return period of an initially estimated 100 year-flood by a log-normal stationary model. To evaluate whether or not a particular site exhibits positive trend, we only considered data series with linear regression slope coefficients that exhibit significance levels (p<0,05). The significance level was calculated using the one-sided Student's test. The trend model residuals were analyzed using the Anderson-Darling normality test, the Durbin-Watson test for the independence and the Breusch-Pagan test for heteroscedasticity. Our results showed that 22 of the 38 data series analyzed have a significant positive trend. The trends were mainly in three large basins: Iguazu, Uruguay and Paranapanema, which suffered changes in land use and flow regularization in the last years. The calculated return period for the series that presented positive trend varied from 50 to 77 years for a 100 year-flood estimated by stationary model when considering a planning horizon equal to ten years. We conclude that attention should be given for future projects developed in this area, including the incorporation of non-stationarity analysis, search for answers to such changes and incorporation of new data to increase the reliability of the estimates.
Vandewalle, Sara; Van Caenegem, Eva; Craen, Margarita; Taes, Youri; Kaufman, Jean-Marc; T'Sjoen, Guy
2018-03-28
Sex steroids are essential for sexual maturation, linear growth and bone development. However, there is no consensus on the optimal timing, dosage and dosage interval of testosterone therapy to induce pubertal development and achieve a normal adult height and bone mass in children with hypogonadism. A monozygotic monochorial male twin pair, of which one boy was diagnosed with anorchia at birth due to testicular regression syndrome was followed from the age of 3 until the age of 18 years. Low dose testosterone substitution (testosterone esters 25 mg/2 weeks) was initiated in the affected twin based on the start of pubertal development in the healthy twin and then gradually increased accordingly. Both boys were followed until age 18 and were compared as regards to linear growth, sexual maturation, bone maturation and bone development. Before puberty induction both boys had a similar weight and height. During puberty, a slightly faster weight and height gain was observed in the affected twin. Both boys ended up however, with a similar and normal (near) adult height and weight and experienced a normal development of secondary sex characteristics. At the age of 17 and 18 years, bone mineral density, body composition and volumetric bone parameters at the forearm and calf were evaluated in both boys. The affected boy had a higher lean mass and muscle cross-sectional area. The bone mineral density at the lumbar spine and whole body was similar. Trabecular and cortical volumetric bone parameters were comparable. At one cortical site (proximal radius), however, the affected twin had a smaller periosteal and endosteal circumference with a thicker cortex. In conclusion, a low dose testosterone substitution in bilateral anorchia led to a normal onset of pubertal development and (near) adult height. Furthermore, there was no difference in bone mineral density at the age of 17 and 18 years.
2016-01-01
Understanding the relationship between physiological measurements from human subjects and their demographic data is important within both the biometric and forensic domains. In this paper we explore the relationship between measurements of the human hand and a range of demographic features. We assess the ability of linear regression and machine learning classifiers to predict demographics from hand features, thereby providing evidence on both the strength of relationship and the key features underpinning this relationship. Our results show that we are able to predict sex, height, weight and foot size accurately within various data-range bin sizes, with machine learning classification algorithms out-performing linear regression in most situations. In addition, we identify the features used to provide these relationships applicable across multiple applications. PMID:27806075
Miguel-Hurtado, Oscar; Guest, Richard; Stevenage, Sarah V; Neil, Greg J; Black, Sue
2016-01-01
Understanding the relationship between physiological measurements from human subjects and their demographic data is important within both the biometric and forensic domains. In this paper we explore the relationship between measurements of the human hand and a range of demographic features. We assess the ability of linear regression and machine learning classifiers to predict demographics from hand features, thereby providing evidence on both the strength of relationship and the key features underpinning this relationship. Our results show that we are able to predict sex, height, weight and foot size accurately within various data-range bin sizes, with machine learning classification algorithms out-performing linear regression in most situations. In addition, we identify the features used to provide these relationships applicable across multiple applications.
Kumar, K Vasanth; Porkodi, K; Rocha, F
2008-01-15
A comparison of linear and non-linear regression method in selecting the optimum isotherm was made to the experimental equilibrium data of basic red 9 sorption by activated carbon. The r(2) was used to select the best fit linear theoretical isotherm. In the case of non-linear regression method, six error functions namely coefficient of determination (r(2)), hybrid fractional error function (HYBRID), Marquardt's percent standard deviation (MPSD), the average relative error (ARE), sum of the errors squared (ERRSQ) and sum of the absolute errors (EABS) were used to predict the parameters involved in the two and three parameter isotherms and also to predict the optimum isotherm. Non-linear regression was found to be a better way to obtain the parameters involved in the isotherms and also the optimum isotherm. For two parameter isotherm, MPSD was found to be the best error function in minimizing the error distribution between the experimental equilibrium data and predicted isotherms. In the case of three parameter isotherm, r(2) was found to be the best error function to minimize the error distribution structure between experimental equilibrium data and theoretical isotherms. The present study showed that the size of the error function alone is not a deciding factor to choose the optimum isotherm. In addition to the size of error function, the theory behind the predicted isotherm should be verified with the help of experimental data while selecting the optimum isotherm. A coefficient of non-determination, K(2) was explained and was found to be very useful in identifying the best error function while selecting the optimum isotherm.
Probing star formation relations of mergers and normal galaxies across the CO ladder
NASA Astrophysics Data System (ADS)
Greve, Thomas R.
We examine integrated luminosity relations between the IR continuum and the CO rotational ladder observed for local (ultra) luminous infra-red galaxies ((U)LIRGs, L IR >= 1011 M⊙) and normal star forming galaxies in the context of radiation pressure regulated star formation proposed by Andrews & Thompson (2011). This can account for the normalization and linear slopes of the luminosity relations (log L IR = α log L'CO + β) of both low- and high-J CO lines observed for normal galaxies. Super-linear slopes occur for galaxy samples with significantly different dense gas fractions. Local (U)LIRGs are observed to have sub-linear high-J (J up > 6) slopes or, equivalently, increasing L COhigh-J /L IR with L IR. In the extreme ISM conditions of local (U)LIRGs, the high-J CO lines no longer trace individual hot spots of star formation (which gave rise to the linear slopes for normal galaxies) but a more widespread warm and dense gas phase mechanically heated by powerful supernovae-driven turbulence and shocks.
Applied Multiple Linear Regression: A General Research Strategy
ERIC Educational Resources Information Center
Smith, Brandon B.
1969-01-01
Illustrates some of the basic concepts and procedures for using regression analysis in experimental design, analysis of variance, analysis of covariance, and curvilinear regression. Applications to evaluation of instruction and vocational education programs are illustrated. (GR)
Khalil, Mohamed H.; Shebl, Mostafa K.; Kosba, Mohamed A.; El-Sabrout, Karim; Zaki, Nesma
2016-01-01
Aim: This research was conducted to determine the most affecting parameters on hatchability of indigenous and improved local chickens’ eggs. Materials and Methods: Five parameters were studied (fertility, early and late embryonic mortalities, shape index, egg weight, and egg weight loss) on four strains, namely Fayoumi, Alexandria, Matrouh, and Montazah. Multiple linear regression was performed on the studied parameters to determine the most influencing one on hatchability. Results: The results showed significant differences in commercial and scientific hatchability among strains. Alexandria strain has the highest significant commercial hatchability (80.70%). Regarding the studied strains, highly significant differences in hatching chick weight among strains were observed. Using multiple linear regression analysis, fertility made the greatest percent contribution (71.31%) to hatchability, and the lowest percent contributions were made by shape index and egg weight loss. Conclusion: A prediction of hatchability using multiple regression analysis could be a good tool to improve hatchability percentage in chickens. PMID:27651666
Predicting recycling behaviour: Comparison of a linear regression model and a fuzzy logic model.
Vesely, Stepan; Klöckner, Christian A; Dohnal, Mirko
2016-03-01
In this paper we demonstrate that fuzzy logic can provide a better tool for predicting recycling behaviour than the customarily used linear regression. To show this, we take a set of empirical data on recycling behaviour (N=664), which we randomly divide into two halves. The first half is used to estimate a linear regression model of recycling behaviour, and to develop a fuzzy logic model of recycling behaviour. As the first comparison, the fit of both models to the data included in estimation of the models (N=332) is evaluated. As the second comparison, predictive accuracy of both models for "new" cases (hold-out data not included in building the models, N=332) is assessed. In both cases, the fuzzy logic model significantly outperforms the regression model in terms of fit. To conclude, when accurate predictions of recycling and possibly other environmental behaviours are needed, fuzzy logic modelling seems to be a promising technique. Copyright © 2015 Elsevier Ltd. All rights reserved.
Bennett, Bradley C; Husby, Chad E
2008-03-28
Botanical pharmacopoeias are non-random subsets of floras, with some taxonomic groups over- or under-represented. Moerman [Moerman, D.E., 1979. Symbols and selectivity: a statistical analysis of Native American medical ethnobotany, Journal of Ethnopharmacology 1, 111-119] introduced linear regression/residual analysis to examine these patterns. However, regression, the commonly-employed analysis, suffers from several statistical flaws. We use contingency table and binomial analyses to examine patterns of Shuar medicinal plant use (from Amazonian Ecuador). We first analyzed the Shuar data using Moerman's approach, modified to better meet requirements of linear regression analysis. Second, we assessed the exact randomization contingency table test for goodness of fit. Third, we developed a binomial model to test for non-random selection of plants in individual families. Modified regression models (which accommodated assumptions of linear regression) reduced R(2) to from 0.59 to 0.38, but did not eliminate all problems associated with regression analyses. Contingency table analyses revealed that the entire flora departs from the null model of equal proportions of medicinal plants in all families. In the binomial analysis, only 10 angiosperm families (of 115) differed significantly from the null model. These 10 families are largely responsible for patterns seen at higher taxonomic levels. Contingency table and binomial analyses offer an easy and statistically valid alternative to the regression approach.
An Application to the Prediction of LOD Change Based on General Regression Neural Network
NASA Astrophysics Data System (ADS)
Zhang, X. H.; Wang, Q. J.; Zhu, J. J.; Zhang, H.
2011-07-01
Traditional prediction of the LOD (length of day) change was based on linear models, such as the least square model and the autoregressive technique, etc. Due to the complex non-linear features of the LOD variation, the performances of the linear model predictors are not fully satisfactory. This paper applies a non-linear neural network - general regression neural network (GRNN) model to forecast the LOD change, and the results are analyzed and compared with those obtained with the back propagation neural network and other models. The comparison shows that the performance of the GRNN model in the prediction of the LOD change is efficient and feasible.
Drop-Weight Impact Test on U-Shape Concrete Specimens with Statistical and Regression Analyses
Zhu, Xue-Chao; Zhu, Han; Li, Hao-Ran
2015-01-01
According to the principle and method of drop-weight impact test, the impact resistance of concrete was measured using self-designed U-shape specimens and a newly designed drop-weight impact test apparatus. A series of drop-weight impact tests were carried out with four different masses of drop hammers (0.875, 0.8, 0.675 and 0.5 kg). The test results show that the impact resistance results fail to follow a normal distribution. As expected, U-shaped specimens can predetermine the location of the cracks very well. It is also easy to record the cracks propagation during the test. The maximum of coefficient of variation in this study is 31.2%; it is lower than the values obtained from the American Concrete Institute (ACI) impact tests in the literature. By regression analysis, the linear relationship between the first-crack and ultimate failure impact resistance is good. It can suggested that a minimum number of specimens is required to reliably measure the properties of the material based on the observed levels of variation. PMID:28793540
Prediction of human gait parameters from temporal measures of foot-ground contact
NASA Technical Reports Server (NTRS)
Breit, G. A.; Whalen, R. T.
1997-01-01
Investigation of the influence of human physical activity on bone functional adaptation requires long-term histories of gait-related ground reaction force (GRF). Towards a simpler portable GRF measurement, we hypothesized that: 1) the reciprocal of foot-ground contact time (1/tc); or 2) the reciprocal of stride-period-normalized contact time (T/tc) predict peak vertical and horizontal GRF, loading rates, and horizontal speed during gait. GRF data were collected from 24 subjects while they walked and ran at a variety of speeds. Linear regression and ANCOVA determined the dependence of gait parameters on 1/tc and T/tc, and prediction SE. All parameters were significantly correlated to 1/tc and T/tc. The closest pooled relationship existed between peak running vertical GRF and T/tc (r2 = 0.896; SE = 3.6%) and improved with subject-specific regression (r2 = 0.970; SE = 2.2%). We conclude that temporal measures can predict force parameters of gait and may represent an alternative to direct GRF measurements for determining daily histories of habitual lower limb loading quantities necessary to quantify a bone remodeling stimulus.
DOT National Transportation Integrated Search
2016-09-01
We consider the problem of solving mixed random linear equations with k components. This is the noiseless setting of mixed linear regression. The goal is to estimate multiple linear models from mixed samples in the case where the labels (which sample...
Linear regression techniques for use in the EC tracer method of secondary organic aerosol estimation
NASA Astrophysics Data System (ADS)
Saylor, Rick D.; Edgerton, Eric S.; Hartsell, Benjamin E.
A variety of linear regression techniques and simple slope estimators are evaluated for use in the elemental carbon (EC) tracer method of secondary organic carbon (OC) estimation. Linear regression techniques based on ordinary least squares are not suitable for situations where measurement uncertainties exist in both regressed variables. In the past, regression based on the method of Deming [1943. Statistical Adjustment of Data. Wiley, London] has been the preferred choice for EC tracer method parameter estimation. In agreement with Chu [2005. Stable estimate of primary OC/EC ratios in the EC tracer method. Atmospheric Environment 39, 1383-1392], we find that in the limited case where primary non-combustion OC (OC non-comb) is assumed to be zero, the ratio of averages (ROA) approach provides a stable and reliable estimate of the primary OC-EC ratio, (OC/EC) pri. In contrast with Chu [2005. Stable estimate of primary OC/EC ratios in the EC tracer method. Atmospheric Environment 39, 1383-1392], however, we find that the optimal use of Deming regression (and the more general York et al. [2004. Unified equations for the slope, intercept, and standard errors of the best straight line. American Journal of Physics 72, 367-375] regression) provides excellent results as well. For the more typical case where OC non-comb is allowed to obtain a non-zero value, we find that regression based on the method of York is the preferred choice for EC tracer method parameter estimation. In the York regression technique, detailed information on uncertainties in the measurement of OC and EC is used to improve the linear best fit to the given data. If only limited information is available on the relative uncertainties of OC and EC, then Deming regression should be used. On the other hand, use of ROA in the estimation of secondary OC, and thus the assumption of a zero OC non-comb value, generally leads to an overestimation of the contribution of secondary OC to total measured OC.
Chouabe, C; Amsellem, J; Espinosa, L; Ribaux, P; Blaineau, S; Mégas, P; Bonvallet, R
2002-04-01
Recent studies indicate that regression of left ventricular hypertrophy normalizes membrane ionic current abnormalities. This work was designed to determine whether regression of right ventricular hypertrophy induced by permanent high-altitude exposure (4,500 m, 20 days) in adult rats also normalizes changes of ventricular myocyte electrophysiology. According to the current data, prolonged action potential, decreased transient outward current density, and increased inward sodium/calcium exchange current density normalized 20 days after the end of altitude exposure, whereas right ventricular hypertrophy evidenced by both the right ventricular weight-to-heart weight ratio and the right ventricular free wall thickness measurement normalized 40 days after the end of altitude exposure. This morphological normalization occurred at both the level of muscular tissue, as shown by the decrease toward control values of some myocyte parameters (perimeter, capacitance, and width), and the level of the interstitial collagenous connective tissue. In the chronic high-altitude hypoxia model, the regression of right ventricular hypertrophy would not be a prerequisite for normalization of ventricular electrophysiological abnormalities.
Statistical Methods for Generalized Linear Models with Covariates Subject to Detection Limits.
Bernhardt, Paul W; Wang, Huixia J; Zhang, Daowen
2015-05-01
Censored observations are a common occurrence in biomedical data sets. Although a large amount of research has been devoted to estimation and inference for data with censored responses, very little research has focused on proper statistical procedures when predictors are censored. In this paper, we consider statistical methods for dealing with multiple predictors subject to detection limits within the context of generalized linear models. We investigate and adapt several conventional methods and develop a new multiple imputation approach for analyzing data sets with predictors censored due to detection limits. We establish the consistency and asymptotic normality of the proposed multiple imputation estimator and suggest a computationally simple and consistent variance estimator. We also demonstrate that the conditional mean imputation method often leads to inconsistent estimates in generalized linear models, while several other methods are either computationally intensive or lead to parameter estimates that are biased or more variable compared to the proposed multiple imputation estimator. In an extensive simulation study, we assess the bias and variability of different approaches within the context of a logistic regression model and compare variance estimation methods for the proposed multiple imputation estimator. Lastly, we apply several methods to analyze the data set from a recently-conducted GenIMS study.
ERIC Educational Resources Information Center
Zu, Jiyun; Yuan, Ke-Hai
2012-01-01
In the nonequivalent groups with anchor test (NEAT) design, the standard error of linear observed-score equating is commonly estimated by an estimator derived assuming multivariate normality. However, real data are seldom normally distributed, causing this normal estimator to be inconsistent. A general estimator, which does not rely on the…
Yang, Xiaowei; Nie, Kun
2008-03-15
Longitudinal data sets in biomedical research often consist of large numbers of repeated measures. In many cases, the trajectories do not look globally linear or polynomial, making it difficult to summarize the data or test hypotheses using standard longitudinal data analysis based on various linear models. An alternative approach is to apply the approaches of functional data analysis, which directly target the continuous nonlinear curves underlying discretely sampled repeated measures. For the purposes of data exploration, many functional data analysis strategies have been developed based on various schemes of smoothing, but fewer options are available for making causal inferences regarding predictor-outcome relationships, a common task seen in hypothesis-driven medical studies. To compare groups of curves, two testing strategies with good power have been proposed for high-dimensional analysis of variance: the Fourier-based adaptive Neyman test and the wavelet-based thresholding test. Using a smoking cessation clinical trial data set, this paper demonstrates how to extend the strategies for hypothesis testing into the framework of functional linear regression models (FLRMs) with continuous functional responses and categorical or continuous scalar predictors. The analysis procedure consists of three steps: first, apply the Fourier or wavelet transform to the original repeated measures; then fit a multivariate linear model in the transformed domain; and finally, test the regression coefficients using either adaptive Neyman or thresholding statistics. Since a FLRM can be viewed as a natural extension of the traditional multiple linear regression model, the development of this model and computational tools should enhance the capacity of medical statistics for longitudinal data.
NASA Astrophysics Data System (ADS)
Gonçalves, Karen dos Santos; Winkler, Mirko S.; Benchimol-Barbosa, Paulo Roberto; de Hoogh, Kees; Artaxo, Paulo Eduardo; de Souza Hacon, Sandra; Schindler, Christian; Künzli, Nino
2018-07-01
Epidemiological studies generally use particulate matter measurements with diameter less 2.5 μm (PM2.5) from monitoring networks. Satellite aerosol optical depth (AOD) data has considerable potential in predicting PM2.5 concentrations, and thus provides an alternative method for producing knowledge regarding the level of pollution and its health impact in areas where no ground PM2.5 measurements are available. This is the case in the Brazilian Amazon rainforest region where forest fires are frequent sources of high pollution. In this study, we applied a non-linear model for predicting PM2.5 concentration from AOD retrievals using interaction terms between average temperature, relative humidity, sine, cosine of date in a period of 365,25 days and the square of the lagged relative residual. Regression performance statistics were tested comparing the goodness of fit and R2 based on results from linear regression and non-linear regression for six different models. The regression results for non-linear prediction showed the best performance, explaining on average 82% of the daily PM2.5 concentrations when considering the whole period studied. In the context of Amazonia, it was the first study predicting PM2.5 concentrations using the latest high-resolution AOD products also in combination with the testing of a non-linear model performance. Our results permitted a reliable prediction considering the AOD-PM2.5 relationship and set the basis for further investigations on air pollution impacts in the complex context of Brazilian Amazon Region.
Classification of the height and flexibility of the medial longitudinal arch of the foot.
Nilsson, Mette Kjærgaard; Friis, Rikke; Michaelsen, Maria Skjoldahl; Jakobsen, Patrick Abildgaard; Nielsen, Rasmus Oestergaard
2012-02-17
The risk of developing injuries during standing work may vary between persons with different foot types. High arched and low arched feet, as well as rigid and flexible feet, are considered to have different injury profiles, while those with normal arches may sustain fewer injuries. However, the cut-off values for maximum values (subtalar position during weight-bearing) and range of motion (ROM) values (difference between subtalar neutral and subtalar resting position in a weight-bearing condition) for the medial longitudinal arch (MLA) are largely unknown. The purpose of this study was to identify cut-off values for maximum values and ROM of the MLA of the foot during static tests and to identify factors influencing foot posture. The participants consisted of 254 volunteers from Central and Northern Denmark (198 m/56 f; age 39.0 ± 11.7 years; BMI 27.3 ± 4.7 kg/m2). Navicular height (NH), longitudinal arch angle (LAA) and Feiss line (FL) were measured for either the left or the right foot in a subtalar neutral position and subtalar resting position. Maximum values and ROM were calculated for each test. The 95% and 68% prediction intervals were used as cut-off limits. Multiple regression analysis was used to detect influencing factors on foot posture. The 68% cut-off values for maximum MLA values and MLA ROM for NH were 3.6 to 5.5 cm and 0.6 to 1.8 cm, respectively, without taking into account the influence of other variables. Normal maximum LAA values were between 131 and 152° and normal LAA ROM was between -1 and 13°. Normal maximum FL values were between -2.6 and -1.2 cm and normal FL ROM was between -0.1 and 0.9 cm. Results from the multivariate linear regression revealed an association between foot size with FL, LAA, and navicular drop. The cut-off values presented in this study can be used to categorize people performing standing work into groups of different foot arch types. The results of this study are important for investigating a possible link between arch height and arch movement and the development of injuries.
Herrera-Guzmán, I; Peña-Casanova, J; Lara, J P; Gudayol-Ferré, E; Böhm, P
2004-08-01
The assessment of visual perception and cognition forms an important part of any general cognitive evaluation. We have studied the possible influence of age, sex, and education on a normal elderly Spanish population (90 healthy subjects) in performance in visual perception tasks. To evaluate visual perception and cognition, we have used the subjects performance with The Visual Object and Space Perception Battery (VOSP). The test consists of 8 subtests: 4 measure visual object perception (Incomplete Letters, Silhouettes, Object Decision, and Progressive Silhouettes) while the other 4 measure visual space perception (Dot Counting, Position Discrimination, Number Location, and Cube Analysis). The statistical procedures employed were either simple or multiple linear regression analyses (subtests with normal distribution) and Mann-Whitney tests, followed by ANOVA with Scheffe correction (subtests without normal distribution). Age and sex were found to be significant modifying factors in the Silhouettes, Object Decision, Progressive Silhouettes, Position Discrimination, and Cube Analysis subtests. Educational level was found to be a significant predictor of function for the Silhouettes and Object Decision subtests. The results of the sample were adjusted in line with the differences observed. Our study also offers preliminary normative data for the administration of the VOSP to an elderly Spanish population. The results are discussed and compared with similar studies performed in different cultural backgrounds.
Schmitt, Christopher J.; McKee, Michael J.
2016-01-01
Lead (Pb) and calcium (Ca) concentrations were measured in fillet samples of longear sunfish (Lepomis megalotis) and redhorse suckers (Moxostoma spp.) collected in 2005–2012 from the Big River, which drains a historical mining area in southeastern Missouri and where a consumption advisory is in effect due to elevated Pb concentrations in fish. Lead tends to accumulated in Ca-rich tissues such as bone and scale. Concentrations of Pb in fish muscle are typically low, but can become elevated in fillets from Pb-contaminated sites depending in part on how much bone, scale, and skin is included in the sample. We used analysis-of-covariance to normalize Pb concentration to the geometric mean Ca concentration (415 ug/g wet weight, ww), which reduced variation between taxa, sites, and years, as was the number of samples that exceeded Missouri consumption advisory threshold (300 ng/g ww). Concentrations of Pb in 2005–2012 were lower than in the past, especially after Ca-normalization, but the consumption advisory is still warranted because concentrations were >300 ng/g ww in samples of both taxa from contaminated sites. For monitoring purposes, a simple linear regression model is proposed for estimating Ca-normalized Pb concentrations in fillets from Pb:Ca molar ratios as a way of reducing the effects of differing preparation methods on fillet Pb variation.
In-Vivo Fluorescence Spectroscopy Of Normal And Atherosclerotic Arteries
NASA Astrophysics Data System (ADS)
Deckelbaum, Lawrence I.; Sarembock, Ian J.; Stetz, Mark L.; O'Brien, Kenneth M.; Cutruzzola, Francis W.; Gmitro, Arthur F.; Ezekowitz, Michael D.
1988-06-01
Laser-induced fluorescence spectroscopy can discriminate atherosclerotic from normal arteries in-vitro and may thus potentially guide laser angioplasty. To evaluate the feasibility of laser-induced fluorescence spectroscopy in a living blood-filled arterial system we performed fiberoptic laser-induced fluorescence spectroscopy in a rabbit model of focal femoral atherosclerosis. A laser-induced fluorescence spectroscopy score was derived from stepwise linear regression analysis of in-vitro spectra to distinguish normal aorta (score>0) from atherosclerotic femoral artery (score<0). A 400 u silica fiber, coupled to a helium cadmium laser and optical multichannel analyzer, was inserted through a 5F catheter to induce and record in-vivo fluorescence from femoral and aortoiliac arteries. Arterial spectra could be recorded in all animals (n=10: 5 occlusions, 5 stenoses). Blood spectra were of low intensity and were easily distinguished from arterial spectra. The scores (mean ± SEM) for the in-vivo spectra were -0.69 +/- 0.29 for artherosclerotic femoral, and +0.54 ±. 0.15 for normal aorta (p<.01 p=NS compared to in-vitro spectra). In-vitro, a fiber tip to tissue distance <50 u was necessary for adequate arterial LIFS in blood. At larger distances low intensity blood spectra were recorded (1/20 the intensity of tissue spectra). Thus, fiberoptic laser-induced fluorescence spectroscopy can be sucessfully performed in a blood filled artery provided the fiber tip is approximated to the tissue.
Schmitt, Christopher J; McKee, Michael J
2016-11-01
Lead (Pb) and calcium (Ca) concentrations were measured in fillet samples of longear sunfish (Lepomis megalotis) and redhorse suckers (Moxostoma spp.) collected in 2005-2012 from the Big River, which drains a historical mining area in southeastern Missouri and where a consumption advisory is in effect due to elevated Pb concentrations in fish. Lead tends to accumulated in Ca-rich tissues such as bone and scale. Concentrations of Pb in fish muscle are typically low, but can become elevated in fillets from Pb-contaminated sites depending in part on how much bone, scale, and skin is included in the sample. We used analysis-of-covariance to normalize Pb concentration to the geometric mean Ca concentration (415 ug/g wet weight, ww), which reduced variation between taxa, sites, and years, as was the number of samples that exceeded Missouri consumption advisory threshold (300 ng/g ww). Concentrations of Pb in 2005-2012 were lower than in the past, especially after Ca-normalization, but the consumption advisory is still warranted because concentrations were >300 ng/g ww in samples of both taxa from contaminated sites. For monitoring purposes, a simple linear regression model is proposed for estimating Ca-normalized Pb concentrations in fillets from Pb:Ca molar ratios as a way of reducing the effects of differing preparation methods on fillet Pb variation.
Hosseinifard, Behshad; Moradi, Mohammad Hassan; Rostami, Reza
2013-03-01
Diagnosing depression in the early curable stages is very important and may even save the life of a patient. In this paper, we study nonlinear analysis of EEG signal for discriminating depression patients and normal controls. Forty-five unmedicated depressed patients and 45 normal subjects were participated in this study. Power of four EEG bands and four nonlinear features including detrended fluctuation analysis (DFA), higuchi fractal, correlation dimension and lyapunov exponent were extracted from EEG signal. For discriminating the two groups, k-nearest neighbor, linear discriminant analysis and logistic regression as the classifiers are then used. Highest classification accuracy of 83.3% is obtained by correlation dimension and LR classifier among other nonlinear features. For further improvement, all nonlinear features are combined and applied to classifiers. A classification accuracy of 90% is achieved by all nonlinear features and LR classifier. In all experiments, genetic algorithm is employed to select the most important features. The proposed technique is compared and contrasted with the other reported methods and it is demonstrated that by combining nonlinear features, the performance is enhanced. This study shows that nonlinear analysis of EEG can be a useful method for discriminating depressed patients and normal subjects. It is suggested that this analysis may be a complementary tool to help psychiatrists for diagnosing depressed patients. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Shi, Jun; Zheng, Yong-Ping; Huang, Qing-Hua; Chen, Xin
2008-03-01
The aim of this study is to demonstrate the feasibility of using the continuous signals about the thickness and pennation angle changes of muscles detected in real-time from ultrasound images, named as sonomyography (SMG), to characterize muscles under isometric contraction, along with synchronized surface electromyography (EMG) and generated torque signals. The right biceps brachii muscles of seven normal young adult subjects were tested. We observed that exponential functions could well represent the relationships between the normalized EMG root-mean-square (RMS) and the torque, the RMS and the muscle deformation SMG, and the RMS and the pennation angle SMG for the data of the contraction phase, with exponent coefficients of 0.0341 +/- 0.0148 (Mean SD), 0.0619 +/- 0.0273, and 0.0266 +/- 0.0076, respectively. In addition, the preliminary results also demonstrated linear relationships between the normalized torque and the muscle deformation as well as the pennation angle with the ratios of 9 .79 +/- 3.01 and 2.02 +/- 0.53, respectively. The overall mean R2 for the regressions was approximately 0.9 and the overall mean relative root mean square error (RRMSE) smaller than 15%. The potential values of SMG together with EMG to provide a more comprehensive assessment for the muscle functions should be further investigated with more subjects and more muscle groups.
Kretsch, M J; Fong, A K; Green, M W
1999-03-01
To examine behavioral and body size influences on the underreporting of energy intake by obese and normal-weight women. Seven-day estimated food records were kept by subjects before they participated in a 49-day residential study. Self-reported energy intake was compared with energy intake required to maintain a stable body weight during the residential study (reference standard). Energy intake bias and its relationship to various body size and behavioral measures were examined. Twenty-two, healthy, normal-weight (mean body mass index [BMI] = 21.3) and obese (mean BMI = 34.2) women aged 22 to 42 years were studied. Analysis of variance, paired t test, simple linear regression, and Pearson correlation analyses were conducted. Mean energy intake from self-reported food records was underreported by normal-weight (-9.7%) and obese (-19.4%) women. BMI correlated inversely with the energy intake difference for normal-weight women (r = -.67, P = .02), whereas the Beck Depression Inventory correlated positively with the energy intake difference for obese women (r = .73, P < .01). CONCLUSION/APPLICATIONS: Results suggest that body size and behavioral traits play a role in the ability of women to accurately self-report energy intake. BMI appears to be predictive of underreporting of energy intake by normal-weight women, whereas emotional factors related to depression appear to be more determinant of underreporting for obese women. Understanding causative factors of the underreporting phenomenon will help practicing dietitians to devise appropriate and realistic diet intervention plans that clients can follow to achieve meaningful change.
Senn, Stephen; Graf, Erika; Caputo, Angelika
2007-12-30
Stratifying and matching by the propensity score are increasingly popular approaches to deal with confounding in medical studies investigating effects of a treatment or exposure. A more traditional alternative technique is the direct adjustment for confounding in regression models. This paper discusses fundamental differences between the two approaches, with a focus on linear regression and propensity score stratification, and identifies points to be considered for an adequate comparison. The treatment estimators are examined for unbiasedness and efficiency. This is illustrated in an application to real data and supplemented by an investigation on properties of the estimators for a range of underlying linear models. We demonstrate that in specific circumstances the propensity score estimator is identical to the effect estimated from a full linear model, even if it is built on coarser covariate strata than the linear model. As a consequence the coarsening property of the propensity score-adjustment for a one-dimensional confounder instead of a high-dimensional covariate-may be viewed as a way to implement a pre-specified, richly parametrized linear model. We conclude that the propensity score estimator inherits the potential for overfitting and that care should be taken to restrict covariates to those relevant for outcome. Copyright (c) 2007 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Dong, Sheng; Dapino, Marcelo J.
2015-04-01
Ultrasonic lubrication has been proven effective in reducing dynamic friction. This paper investigates the relationship between friction reduction, power consumption, linear velocity, and normal stress. A modified pin-on-disc tribometer was adopted as the experimental set-up, and a Labview system was utilized for signal generation and data acquisition. Friction reduction was quantified for 0.21 to 5.31 W of electric power, 50 to 200 mm/s of linear velocity, and 23 to 70 MPa of normal stress. Friction reduction near 100% can be achieved under certain conditions. Lower linear velocity and higher electric power result in greater friction reduction, while normal stress has little effect on friction reduction. Contour plots of friction reduction, power consumption, linear velocity, and normal stress were created. An efficiency coefficient was proposed to calculate power requirements for a certain friction reduction or reduced friction for a given electric power.
Quantifying progression and regression of thrombotic risk in experimental atherosclerosis
Palekar, Rohun U.; Jallouk, Andrew P.; Goette, Matthew J.; Chen, Junjie; Myerson, Jacob W.; Allen, John S.; Akk, Antonina; Yang, Lihua; Tu, Yizheng; Miller, Mark J.; Pham, Christine T. N.; Wickline, Samuel A.; Pan, Hua
2015-01-01
Currently, there are no generally applicable noninvasive methods for defining the relationship between atherosclerotic vascular damage and risk of focal thrombosis. Herein, we demonstrate methods to delineate the progression and regression of vascular damage in response to an atherogenic diet by quantifying the in vivo accumulation of semipermeable 200–300 nm perfluorocarbon core nanoparticles (PFC-NP) in ApoE null mouse plaques with [19F] magnetic resonance spectroscopy (MRS). Permeability to PFC-NP remained minimal until 12 weeks on diet, then increased rapidly following 12 weeks, but regressed to baseline within 8 weeks after diet normalization. Markedly accelerated clotting (53.3% decrease in clotting time) was observed in carotid artery preparations of fat-fed mice subjected to photochemical injury as defined by the time to flow cessation. For all mice on and off diet, an inverse linear relationship was observed between the permeability to PFC-NP and accelerated thrombosis (P = 0.02). Translational feasibility for quantifying plaque permeability and vascular damage in vivo was demonstrated with clinical 3 T MRI of PFC-NP accumulating in plaques of atherosclerotic rabbits. These observations suggest that excessive permeability to PFC-NP may indicate prothrombotic risk in damaged atherosclerotic vasculature, which resolves within weeks after dietary therapy.—Palekar, R. U., Jallouk, A. P., Goette, M. J., Chen, J., Myerson, J. W., Allen, J. S., Akk, A., Yang, L., Tu, Y., Miller, M. J., Pham, C. T. N., Wickline, S. A., Pan, H. Quantifying progression and regression of thrombotic risk in experimental atherosclerosis. PMID:25857553
Spencer, Monique E; Jain, Alka; Matteini, Amy; Beamer, Brock A; Wang, Nae-Yuh; Leng, Sean X; Punjabi, Naresh M; Walston, Jeremy D; Fedarko, Neal S
2010-08-01
Neopterin, a GTP metabolite expressed by macrophages, is a marker of immune activation. We hypothesize that levels of this serum marker alter with donor age, reflecting increased chronic immune activation in normal aging. In addition to age, we assessed gender, race, body mass index (BMI), and percentage of body fat (%fat) as potential covariates. Serum was obtained from 426 healthy participants whose age ranged from 18 to 87 years. Anthropometric measures included %fat and BMI. Neopterin concentrations were measured by competitive ELISA. The paired associations between neopterin and age, BMI, or %fat were analyzed by Spearman's correlation or by linear regression of log-transformed neopterin, whereas overall associations were modeled by multiple regression of log-transformed neopterin as a function of age, gender, race, BMI, %fat, and interaction terms. Across all participants, neopterin exhibited a positive association with age, BMI, and %fat. Multiple regression modeling of neopterin in women and men as a function of age, BMI, and race revealed that each covariate contributed significantly to neopterin values and that optimal modeling required an interaction term between race and BMI. The covariate %fat was highly correlated with BMI and could be substituted for BMI to yield similar regression coefficients. The association of age and gender with neopterin levels and their modification by race, BMI, or %fat reflect the biology underlying chronic immune activation and perhaps gender differences in disease incidence, morbidity, and mortality.
The long-solved problem of the best-fit straight line: application to isotopic mixing lines
NASA Astrophysics Data System (ADS)
Wehr, Richard; Saleska, Scott R.
2017-01-01
It has been almost 50 years since York published an exact and general solution for the best-fit straight line to independent points with normally distributed errors in both x and y. York's solution is highly cited in the geophysical literature but almost unknown outside of it, so that there has been no ebb in the tide of books and papers wrestling with the problem. Much of the post-1969 literature on straight-line fitting has sown confusion not merely by its content but by its very existence. The optimal least-squares fit is already known; the problem is already solved. Here we introduce the non-specialist reader to York's solution and demonstrate its application in the interesting case of the isotopic mixing line, an analytical tool widely used to determine the isotopic signature of trace gas sources for the study of biogeochemical cycles. The most commonly known linear regression methods - ordinary least-squares regression (OLS), geometric mean regression (GMR), and orthogonal distance regression (ODR) - have each been recommended as the best method for fitting isotopic mixing lines. In fact, OLS, GMR, and ODR are all special cases of York's solution that are valid only under particular measurement conditions, and those conditions do not hold in general for isotopic mixing lines. Using Monte Carlo simulations, we quantify the biases in OLS, GMR, and ODR under various conditions and show that York's general - and convenient - solution is always the least biased.
Graph-based normalization and whitening for non-linear data analysis.
Aaron, Catherine
2006-01-01
In this paper we construct a graph-based normalization algorithm for non-linear data analysis. The principle of this algorithm is to get a spherical average neighborhood with unit radius. First we present a class of global dispersion measures used for "global normalization"; we then adapt these measures using a weighted graph to build a local normalization called "graph-based" normalization. Then we give details of the graph-based normalization algorithm and illustrate some results. In the second part we present a graph-based whitening algorithm built by analogy between the "global" and the "local" problem.
Analysis of Different Hyperspectral Variables for Diagnosing Leaf Nitrogen Accumulation in Wheat.
Tan, Changwei; Du, Ying; Zhou, Jian; Wang, Dunliang; Luo, Ming; Zhang, Yongjian; Guo, Wenshan
2018-01-01
Hyperspectral remote sensing is a rapid non-destructive method for diagnosing nitrogen status in wheat crops. In this study, a quantitative correlation was associated with following parameters: leaf nitrogen accumulation (LNA), raw hyperspectral reflectance, first-order differential hyperspectra, and hyperspectral characteristics of wheat. In this study, integrated linear regression of LNA was obtained with raw hyperspectral reflectance (measurement wavelength = 790.4 nm). Furthermore, an exponential regression of LNA was obtained with first-order differential hyperspectra (measurement wavelength = 831.7 nm). Coefficients ( R 2 ) were 0.813 and 0.847; root mean squared errors (RMSE) were 2.02 g·m -2 and 1.72 g·m -2 ; and relative errors (RE) were 25.97% and 20.85%, respectively. Both the techniques were considered as optimal in the diagnoses of wheat LNA. Nevertheless, the better one was the new normalized variable (SD r - SD b )/(SD r + SD b ) , which was based on vegetation indices of R 2 = 0.935, RMSE = 0.98, and RE = 11.25%. In addition, (SD r - SD b )/(SD r + SD b ) was reliable in the application of a different cultivar or even wheat grown elsewhere. This indicated a superior fit and better performance for (SD r - SD b )/(SD r + SD b ) . For diagnosing LNA in wheat, the newly normalized variable (SD r - SD b )/(SD r + SD b ) was more effective than the previously reported data of raw hyperspectral reflectance, first-order differential hyperspectra, and red-edge parameters.
Prediction equations for maximal respiratory pressures of Brazilian adolescents.
Mendes, Raquel E F; Campos, Tania F; Macêdo, Thalita M F; Borja, Raíssa O; Parreira, Verônica F; Mendonça, Karla M P P
2013-01-01
The literature emphasizes the need for studies to provide reference values and equations able to predict respiratory muscle strength of Brazilian subjects at different ages and from different regions of Brazil. To develop prediction equations for maximal respiratory pressures (MRP) of Brazilian adolescents. In total, 182 healthy adolescents (98 boys and 84 girls) aged between 12 and 18 years, enrolled in public and private schools in the city of Natal-RN, were evaluated using an MVD300 digital manometer (Globalmed®) according to a standardized protocol. Statistical analysis was performed using SPSS Statistics 17.0 software, with a significance level of 5%. Data normality was verified using the Kolmogorov-Smirnov test, and descriptive analysis results were expressed as the mean and standard deviation. To verify the correlation between the MRP and the independent variables (age, weight, height and sex), the Pearson correlation test was used. To obtain the prediction equations, stepwise multiple linear regression was used. The variables height, weight and sex were correlated to MRP. However, weight and sex explained part of the variability of MRP, and the regression analysis in this study indicated that these variables contributed significantly in predicting maximal inspiratory pressure, and only sex contributed significantly to maximal expiratory pressure. This study provides reference values and two models of prediction equations for maximal inspiratory and expiratory pressures and sets the necessary normal lower limits for the assessment of the respiratory muscle strength of Brazilian adolescents.
Non-Linear Approach in Kinesiology Should Be Preferred to the Linear--A Case of Basketball.
Trninić, Marko; Jeličić, Mario; Papić, Vladan
2015-07-01
In kinesiology, medicine, biology and psychology, in which research focus is on dynamical self-organized systems, complex connections exist between variables. Non-linear nature of complex systems has been discussed and explained by the example of non-linear anthropometric predictors of performance in basketball. Previous studies interpreted relations between anthropometric features and measures of effectiveness in basketball by (a) using linear correlation models, and by (b) including all basketball athletes in the same sample of participants regardless of their playing position. In this paper the significance and character of linear and non-linear relations between simple anthropometric predictors (AP) and performance criteria consisting of situation-related measures of effectiveness (SE) in basketball were determined and evaluated. The sample of participants consisted of top-level junior basketball players divided in three groups according to their playing time (8 minutes and more per game) and playing position: guards (N = 42), forwards (N = 26) and centers (N = 40). Linear (general model) and non-linear (general model) regression models were calculated simultaneously and separately for each group. The conclusion is viable: non-linear regressions are frequently superior to linear correlations when interpreting actual association logic among research variables.
Fenske, Nora; Burns, Jacob; Hothorn, Torsten; Rehfuess, Eva A.
2013-01-01
Background Most attempts to address undernutrition, responsible for one third of global child deaths, have fallen behind expectations. This suggests that the assumptions underlying current modelling and intervention practices should be revisited. Objective We undertook a comprehensive analysis of the determinants of child stunting in India, and explored whether the established focus on linear effects of single risks is appropriate. Design Using cross-sectional data for children aged 0–24 months from the Indian National Family Health Survey for 2005/2006, we populated an evidence-based diagram of immediate, intermediate and underlying determinants of stunting. We modelled linear, non-linear, spatial and age-varying effects of these determinants using additive quantile regression for four quantiles of the Z-score of standardized height-for-age and logistic regression for stunting and severe stunting. Results At least one variable within each of eleven groups of determinants was significantly associated with height-for-age in the 35% Z-score quantile regression. The non-modifiable risk factors child age and sex, and the protective factors household wealth, maternal education and BMI showed the largest effects. Being a twin or multiple birth was associated with dramatically decreased height-for-age. Maternal age, maternal BMI, birth order and number of antenatal visits influenced child stunting in non-linear ways. Findings across the four quantile and two logistic regression models were largely comparable. Conclusions Our analysis confirms the multifactorial nature of child stunting. It emphasizes the need to pursue a systems-based approach and to consider non-linear effects, and suggests that differential effects across the height-for-age distribution do not play a major role. PMID:24223839
Fenske, Nora; Burns, Jacob; Hothorn, Torsten; Rehfuess, Eva A
2013-01-01
Most attempts to address undernutrition, responsible for one third of global child deaths, have fallen behind expectations. This suggests that the assumptions underlying current modelling and intervention practices should be revisited. We undertook a comprehensive analysis of the determinants of child stunting in India, and explored whether the established focus on linear effects of single risks is appropriate. Using cross-sectional data for children aged 0-24 months from the Indian National Family Health Survey for 2005/2006, we populated an evidence-based diagram of immediate, intermediate and underlying determinants of stunting. We modelled linear, non-linear, spatial and age-varying effects of these determinants using additive quantile regression for four quantiles of the Z-score of standardized height-for-age and logistic regression for stunting and severe stunting. At least one variable within each of eleven groups of determinants was significantly associated with height-for-age in the 35% Z-score quantile regression. The non-modifiable risk factors child age and sex, and the protective factors household wealth, maternal education and BMI showed the largest effects. Being a twin or multiple birth was associated with dramatically decreased height-for-age. Maternal age, maternal BMI, birth order and number of antenatal visits influenced child stunting in non-linear ways. Findings across the four quantile and two logistic regression models were largely comparable. Our analysis confirms the multifactorial nature of child stunting. It emphasizes the need to pursue a systems-based approach and to consider non-linear effects, and suggests that differential effects across the height-for-age distribution do not play a major role.
Aqil, Muhammad; Kita, Ichiro; Yano, Akira; Nishiyama, Soichi
2007-10-01
Traditionally, the multiple linear regression technique has been one of the most widely used models in simulating hydrological time series. However, when the nonlinear phenomenon is significant, the multiple linear will fail to develop an appropriate predictive model. Recently, neuro-fuzzy systems have gained much popularity for calibrating the nonlinear relationships. This study evaluated the potential of a neuro-fuzzy system as an alternative to the traditional statistical regression technique for the purpose of predicting flow from a local source in a river basin. The effectiveness of the proposed identification technique was demonstrated through a simulation study of the river flow time series of the Citarum River in Indonesia. Furthermore, in order to provide the uncertainty associated with the estimation of river flow, a Monte Carlo simulation was performed. As a comparison, a multiple linear regression analysis that was being used by the Citarum River Authority was also examined using various statistical indices. The simulation results using 95% confidence intervals indicated that the neuro-fuzzy model consistently underestimated the magnitude of high flow while the low and medium flow magnitudes were estimated closer to the observed data. The comparison of the prediction accuracy of the neuro-fuzzy and linear regression methods indicated that the neuro-fuzzy approach was more accurate in predicting river flow dynamics. The neuro-fuzzy model was able to improve the root mean square error (RMSE) and mean absolute percentage error (MAPE) values of the multiple linear regression forecasts by about 13.52% and 10.73%, respectively. Considering its simplicity and efficiency, the neuro-fuzzy model is recommended as an alternative tool for modeling of flow dynamics in the study area.
González-Aparicio, I; Hidalgo, J; Baklanov, A; Padró, A; Santa-Coloma, O
2013-07-01
There is extensive evidence of the negative impacts on health linked to the rise of the regional background of particulate matter (PM) 10 levels. These levels are often increased over urban areas becoming one of the main air pollution concerns. This is the case on the Bilbao metropolitan area, Spain. This study describes a data-driven model to diagnose PM10 levels in Bilbao at hourly intervals. The model is built with a training period of 7-year historical data covering different urban environments (inland, city centre and coastal sites). The explanatory variables are quantitative-log [NO2], temperature, short-wave incoming radiation, wind speed and direction, specific humidity, hour and vehicle intensity-and qualitative-working days/weekends, season (winter/summer), the hour (from 00 to 23 UTC) and precipitation/no precipitation. Three different linear regression models are compared: simple linear regression; linear regression with interaction terms (INT); and linear regression with interaction terms following the Sawa's Bayesian Information Criteria (INT-BIC). Each type of model is calculated selecting two different periods: the training (it consists of 6 years) and the testing dataset (it consists of 1 year). The results of each type of model show that the INT-BIC-based model (R(2) = 0.42) is the best. Results were R of 0.65, 0.63 and 0.60 for the city centre, inland and coastal sites, respectively, a level of confidence similar to the state-of-the art methodology. The related error calculated for longer time intervals (monthly or seasonal means) diminished significantly (R of 0.75-0.80 for monthly means and R of 0.80 to 0.98 at seasonally means) with respect to shorter periods.
O'Leary, Neil; Chauhan, Balwantray C; Artes, Paul H
2012-10-01
To establish a method for estimating the overall statistical significance of visual field deterioration from an individual patient's data, and to compare its performance to pointwise linear regression. The Truncated Product Method was used to calculate a statistic S that combines evidence of deterioration from individual test locations in the visual field. The overall statistical significance (P value) of visual field deterioration was inferred by comparing S with its permutation distribution, derived from repeated reordering of the visual field series. Permutation of pointwise linear regression (PoPLR) and pointwise linear regression were evaluated in data from patients with glaucoma (944 eyes, median mean deviation -2.9 dB, interquartile range: -6.3, -1.2 dB) followed for more than 4 years (median 10 examinations over 8 years). False-positive rates were estimated from randomly reordered series of this dataset, and hit rates (proportion of eyes with significant deterioration) were estimated from the original series. The false-positive rates of PoPLR were indistinguishable from the corresponding nominal significance levels and were independent of baseline visual field damage and length of follow-up. At P < 0.05, the hit rates of PoPLR were 12, 29, and 42%, at the fifth, eighth, and final examinations, respectively, and at matching specificities they were consistently higher than those of pointwise linear regression. In contrast to population-based progression analyses, PoPLR provides a continuous estimate of statistical significance for visual field deterioration individualized to a particular patient's data. This allows close control over specificity, essential for monitoring patients in clinical practice and in clinical trials.
ERIC Educational Resources Information Center
Liou, Pey-Yan
2009-01-01
The current study examines three regression models: OLS (ordinary least square) linear regression, Poisson regression, and negative binomial regression for analyzing count data. Simulation results show that the OLS regression model performed better than the others, since it did not produce more false statistically significant relationships than…
Boisen, Mogens Karsbøl; Dehlendorff, Christian; Linnemann, Dorte; Schultz, Nicolai Aagaard; Jensen, Benny Vittrup; Høgdall, Estrid Vilma Solyom; Johansen, Julia Sidenius
2015-12-29
Archival formalin-fixed paraffin-embedded (FFPE) cancer tissue samples are a readily available resource for microRNA (miRNA) biomarker identification. No established standard for reference miRNAs in FFPE tissue exists. We sought to identify stable reference miRNAs for normalization of miRNA expression in FFPE tissue samples from patients with colorectal (CRC) and pancreatic (PC) cancer and to quantify the variability associated with sample age and fixation. High-throughput miRNA profiling results from 203 CRC and 256 PC FFPE samples as well as from 37 paired frozen/FFPE samples from nine other CRC tumors (methodological samples) were used. Candidate reference miRNAs were identified by their correlation with global mean expression. The stability of reference genes was analyzed according to published methods. The association between sample age and global mean miRNA expression was tested using linear regression. Variability was described using correlation coefficients and linear mixed effects models. Normalization effects were determined by changes in standard deviation and by hierarchical clustering. We created lists of 20 miRNAs with the best correlation to global mean expression in each cancer type. Nine of these miRNAs were present in both lists, and miR-103a-3p was the most stable reference miRNA for both CRC and PC FFPE tissue. The optimal number of reference miRNAs was 4 in CRC and 10 in PC. Sample age had a significant effect on global miRNA expression in PC (50% reduction over 20 years) but not in CRC. Formalin fixation for 2-6 days decreased miRNA expression 30-65%. Normalization using global mean expression reduced variability for technical and biological replicates while normalization using the expression of the identified reference miRNAs reduced variability only for biological replicates. Normalization only had a minor impact on clustering results. We identified suitable reference miRNAs for future miRNA expression experiments using CRC- and PC FFPE tissue samples. Formalin fixation decreased miRNA expression considerably, while the effect of increasing sample age was estimated to be negligible in a clinical setting.
Annual variation in the atmospheric radon concentration in Japan.
Kobayashi, Yuka; Yasuoka, Yumi; Omori, Yasutaka; Nagahama, Hiroyuki; Sanada, Tetsuya; Muto, Jun; Suzuki, Toshiyuki; Homma, Yoshimi; Ihara, Hayato; Kubota, Kazuhito; Mukai, Takahiro
2015-08-01
Anomalous atmospheric variations in radon related to earthquakes have been observed in hourly exhaust-monitoring data from radioisotope institutes in Japan. The extraction of seismic anomalous radon variations would be greatly aided by understanding the normal pattern of variation in radon concentrations. Using atmospheric daily minimum radon concentration data from five sampling sites, we show that a sinusoidal regression curve can be fitted to the data. In addition, we identify areas where the atmospheric radon variation is significantly affected by the variation in atmospheric turbulence and the onshore-offshore pattern of Asian monsoons. Furthermore, by comparing the sinusoidal regression curve for the normal annual (seasonal) variations at the five sites to the sinusoidal regression curve for a previously published dataset of radon values at the five Japanese prefectures, we can estimate the normal annual variation pattern. By fitting sinusoidal regression curves to the previously published dataset containing sites in all Japanese prefectures, we find that 72% of the Japanese prefectures satisfy the requirements of the sinusoidal regression curve pattern. Using the normal annual variation pattern of atmospheric daily minimum radon concentration data, these prefectures are suitable areas for obtaining anomalous radon variations related to earthquakes. Copyright © 2015 Elsevier Ltd. All rights reserved.
Use of AMMI and linear regression models to analyze genotype-environment interaction in durum wheat.
Nachit, M M; Nachit, G; Ketata, H; Gauch, H G; Zobel, R W
1992-03-01
The joint durum wheat (Triticum turgidum L var 'durum') breeding program of the International Maize and Wheat Improvement Center (CIMMYT) and the International Center for Agricultural Research in the Dry Areas (ICARDA) for the Mediterranean region employs extensive multilocation testing. Multilocation testing produces significant genotype-environment (GE) interaction that reduces the accuracy for estimating yield and selecting appropriate germ plasm. The sum of squares (SS) of GE interaction was partitioned by linear regression techniques into joint, genotypic, and environmental regressions, and by Additive Main effects and the Multiplicative Interactions (AMMI) model into five significant Interaction Principal Component Axes (IPCA). The AMMI model was more effective in partitioning the interaction SS than the linear regression technique. The SS contained in the AMMI model was 6 times higher than the SS for all three regressions. Postdictive assessment recommended the use of the first five IPCA axes, while predictive assessment AMMI1 (main effects plus IPCA1). After elimination of random variation, AMMI1 estimates for genotypic yields within sites were more precise than unadjusted means. This increased precision was equivalent to increasing the number of replications by a factor of 3.7.
Lorenzo-Seva, Urbano; Ferrando, Pere J
2011-03-01
We provide an SPSS program that implements currently recommended techniques and recent developments for selecting variables in multiple linear regression analysis via the relative importance of predictors. The approach consists of: (1) optimally splitting the data for cross-validation, (2) selecting the final set of predictors to be retained in the equation regression, and (3) assessing the behavior of the chosen model using standard indices and procedures. The SPSS syntax, a short manual, and data files related to this article are available as supplemental materials from brm.psychonomic-journals.org/content/supplemental.
NASA Astrophysics Data System (ADS)
Gusriani, N.; Firdaniza
2018-03-01
The existence of outliers on multiple linear regression analysis causes the Gaussian assumption to be unfulfilled. If the Least Square method is forcedly used on these data, it will produce a model that cannot represent most data. For that, we need a robust regression method against outliers. This paper will compare the Minimum Covariance Determinant (MCD) method and the TELBS method on secondary data on the productivity of phytoplankton, which contains outliers. Based on the robust determinant coefficient value, MCD method produces a better model compared to TELBS method.
Orthogonal Projection in Teaching Regression and Financial Mathematics
ERIC Educational Resources Information Center
Kachapova, Farida; Kachapov, Ilias
2010-01-01
Two improvements in teaching linear regression are suggested. The first is to include the population regression model at the beginning of the topic. The second is to use a geometric approach: to interpret the regression estimate as an orthogonal projection and the estimation error as the distance (which is minimized by the projection). Linear…
Logistic models--an odd(s) kind of regression.
Jupiter, Daniel C
2013-01-01
The logistic regression model bears some similarity to the multivariable linear regression with which we are familiar. However, the differences are great enough to warrant a discussion of the need for and interpretation of logistic regression. Copyright © 2013 American College of Foot and Ankle Surgeons. Published by Elsevier Inc. All rights reserved.
Polynomial compensation, inversion, and approximation of discrete time linear systems
NASA Technical Reports Server (NTRS)
Baram, Yoram
1987-01-01
The least-squares transformation of a discrete-time multivariable linear system into a desired one by convolving the first with a polynomial system yields optimal polynomial solutions to the problems of system compensation, inversion, and approximation. The polynomial coefficients are obtained from the solution to a so-called normal linear matrix equation, whose coefficients are shown to be the weighting patterns of certain linear systems. These, in turn, can be used in the recursive solution of the normal equation.
Use of an acoustic helium analyzer for measuring lung volumes.
Krumpe, P E; MacDannald, H J; Finley, T N; Schear, H E; Hall, J; Cribbs, D
1981-01-01
We have evaluated the use of an acoustic gas analyzer (AGA) for the measurement of total lung capacity (TLC) by single-breath helium dilution. The AGA has a rapid response time (0-90% response = 160 ms for 10% He), is linear for helium concentration of 0.1-10%, is stable over a wide range of ambient temperatures, and is small and portable. We plotted the output of the AGA vs. expired lung volume after a vital capacity breath of 10% He. However, since the AGA is sensitive to changes in speed of sound relative to air, the AGA output signal also reports an artifact due to alveolar gases. We corrected for this artifact by replotting a single-breath expiration after a vital capacity breath of room air. Mean alveolar helium concentration (HeA) was then measured by planimetry, using this alveolar gas curve as the base line. TLC was calculated using the HeA from the corrected AGA output and compared with TLC calculated from HeA simultaneously measured using a mass spectrometer (MS). In 12 normal subjects and 9 patients with chronic obstructive pulmonary disease (COPD) TLC-AGA and TLC-MS were compared by linear regression analysis; correlation coefficient (r) was 0.973 for normals and 0.968 for COPD patients (P less than 0.001). This single-breath; estimation of TLC using the corrected signal of the AGA vs. Expired volume seems ideally suited for the measurement of subdivisions of lung volume in field studies.