Wieczorek, Michael E.
2014-01-01
This digital data release consists of seven data files of soil attributes for the United States and the District of Columbia. The files are derived from National Resources Conservations Service’s (NRCS) Soil Survey Geographic database (SSURGO). The data files can be linked to the raster datasets of soil mapping unit identifiers (MUKEY) available through the NRCS’s Gridded Soil Survey Geographic (gSSURGO) database (http://www.nrcs.usda.gov/wps/portal/nrcs/detail/soils/survey/geo/?cid=nrcs142p2_053628). The associated files, named DRAINAGECLASS, HYDRATING, HYDGRP, HYDRICCONDITION, LAYER, TEXT, and WTDEP are area- and depth-weighted average values for selected soil characteristics from the SSURGO database for the conterminous United States and the District of Columbia. The SSURGO tables were acquired from the NRCS on March 5, 2014. The soil characteristics in the DRAINAGE table are drainage class (DRNCLASS), which identifies the natural drainage conditions of the soil and refers to the frequency and duration of wet periods. The soil characteristics in the HYDRATING table are hydric rating (HYDRATE), a yes/no field that indicates whether or not a map unit component is classified as a "hydric soil". The soil characteristics in the HYDGRP table are the percentages for each hydrologic group per MUKEY. The soil characteristics in the HYDRICCONDITION table are hydric condition (HYDCON), which describes the natural condition of the soil component. The soil characteristics in the LAYER table are available water capacity (AVG_AWC), bulk density (AVG_BD), saturated hydraulic conductivity (AVG_KSAT), vertical saturated hydraulic conductivity (AVG_KV), soil erodibility factor (AVG_KFACT), porosity (AVG_POR), field capacity (AVG_FC), the soil fraction passing a number 4 sieve (AVG_NO4), the soil fraction passing a number 10 sieve (AVG_NO10), the soil fraction passing a number 200 sieve (AVG_NO200), and organic matter (AVG_OM). The soil characteristics in the TEXT table are percent sand, silt, and clay (AVG_SAND, AVG_SILT, and AVG_CLAY). The soil characteristics in the WTDEP table are the annual minimum water table depth (WTDEP_MIN), available water storage in the 0-25 cm soil horizon (AWS025), the minimum water table depth for the months April, May and June (WTDEPAMJ), the available water storage in the first 25 centimeters of the soil horizon (AWS25), the dominant drainage class (DRCLSD), the wettest drainage class (DRCLSWET), and the hydric classification (HYDCLASS), which is an indication of the proportion of the map unit, expressed as a class, that is "hydric", based on the hydric classification of a given MUKEY. (See Entity_Description for more detail). The tables were created with a set of arc macro language (aml) and awk (awk was created at Bell Labsin the 1970s and its name is derived from the first letters of the last names of its authors – Alfred Aho, Peter Weinberger, and Brian Kernighan) scripts. Send an email to mewieczo@usgs.gov to obtain copies of the computer code (See Process_Description.) The methods used are outlined in NRCS's "SSURGO Data Packaging and Use" (NRCS, 2011). The tables can be related or joined to the gSSURGO rasters of MUKEYs by the item 'MUKEY.' Joining or relating the tables to a MUKEY grid allows the creation of grids of area- and depth-weighted soil characteristics. A 90-meter raster of MUKEYs is provided which can be used to produce rasters of soil attributes. More detailed resolution rasters are available through NRCS via the link above.
Average Weighted Receiving Time of Weighted Tetrahedron Koch Networks
NASA Astrophysics Data System (ADS)
Dai, Meifeng; Zhang, Danping; Ye, Dandan; Zhang, Cheng; Li, Lei
2015-07-01
We introduce weighted tetrahedron Koch networks with infinite weight factors, which are generalization of finite ones. The term of weighted time is firstly defined in this literature. The mean weighted first-passing time (MWFPT) and the average weighted receiving time (AWRT) are defined by weighted time accordingly. We study the AWRT with weight-dependent walk. Results show that the AWRT for a nontrivial weight factor sequence grows sublinearly with the network order. To investigate the reason of sublinearity, the average receiving time (ART) for four cases are discussed.
Asymmetric network connectivity using weighted harmonic averages
NASA Astrophysics Data System (ADS)
Morrison, Greg; Mahadevan, L.
2011-02-01
We propose a non-metric measure of the "closeness" felt between two nodes in an undirected, weighted graph using a simple weighted harmonic average of connectivity, that is a real-valued Generalized Erds Number (GEN). While our measure is developed with a collaborative network in mind, the approach can be of use in a variety of artificial and real-world networks. We are able to distinguish between network topologies that standard distance metrics view as identical, and use our measure to study some simple analytically tractable networks. We show how this might be used to look at asymmetry in authorship networks such as those that inspired the integer Erds numbers in mathematical coauthorships. We also show the utility of our approach to devise a ratings scheme that we apply to the data from the NetFlix prize, and find a significant improvement using our method over a baseline.
NASA Astrophysics Data System (ADS)
Sun, Yu; Dai, Meifeng; Xi, Lifeng
Recent work on the networks has focused on the weighted hierarchical networks that are significantly different from the un-weighted hierarchical networks. In this paper we study a family of weighted hierarchical networks which are recursively defined from an initial uncompleted graph, in which the weights of edges have been assigned to different values with certain scale. Firstly, we study analytically the average weighted shortest path (AWSP) on the weighted hierarchical networks. Using a recursive method, we determine explicitly the AWSP. The obtained rigorous solution shows that the networks grow unbounded but with the logarithm of the network size, while the weighted shortest paths stay bounded. Then, depending on a biased random walk, we research the mean first-passage time (MFPT) between a hub node and any peripheral node. Finally, we deduce the analytical expression of the average of MFPTs for a random walker originating from any node to first visit to a hub node, which is named as the average receiving time (ART). The obtained result shows that ART is bounded or grows sublinearly with the network order relating to the number of initial nodes and the weighted factor or grows quadratically with the iteration.
NASA Astrophysics Data System (ADS)
Wu, Zikai; Hou, Baoyu; Zhang, Hongjuan; Jin, Feng
2014-04-01
Deterministic network models have been attractive media for discussing dynamical processes' dependence on network structural features. On the other hand, the heterogeneity of weights affect dynamical processes taking place on networks. In this paper, we present a family of weighted expanded Koch networks based on Koch networks. They originate from a r-polygon, and each node of current generation produces m r-polygons including the node and whose weighted edges are scaled by factor w in subsequent evolutionary step. We derive closed-form expressions for average weighted shortest path length (AWSP). In large network, AWSP stays bounded with network order growing (0 < w < 1). Then, we focus on a special random walks and trapping issue on the networks. In more detail, we calculate exactly the average receiving time (ART). ART exhibits a sub-linear dependence on network order (0 < w < 1), which implies that nontrivial weighted expanded Koch networks are more efficient than un-weighted expanded Koch networks in receiving information. Besides, efficiency of receiving information at hub nodes is also dependent on parameters m and r. These findings may pave the way for controlling information transportation on general weighted networks.
Describing Average- and Longtime-Behavior by Weighted MSO Logics
NASA Astrophysics Data System (ADS)
Droste, Manfred; Meinecke, Ingmar
Weighted automata model quantitative aspects of systems like memory or power consumption. Recently, Chatterjee, Doyen, and Henzinger introduced a new kind of weighted automata which compute objectives like the average cost or the longtime peak power consumption. In these automata, operations like average, limit superior, limit inferior, limit average, or discounting are used to assign values to finite or infinite words. In general, these weighted automata are not semiring weighted anymore. Here, we establish a connection between such new kinds of weighted automata and weighted logics. We show that suitable weighted MSO logics and these new weighted automata are expressively equivalent, both for finite and infinite words. The constructions employed are effective, leading to decidability results for the weighted logic formulas considered.
Global Rotation Estimation Using Weighted Iterative Lie Algebraic Averaging
NASA Astrophysics Data System (ADS)
Reich, M.; Heipke, C.
2015-08-01
In this paper we present an approach for a weighted rotation averaging to estimate absolute rotations from relative rotations between two images for a set of multiple overlapping images. The solution does not depend on initial values for the unknown parameters and is robust against outliers. Our approach is one part of a solution for a global image orientation. Often relative rotations are not free from outliers, thus we use the redundancy in available pairwise relative rotations and present a novel graph-based algorithm to detect and eliminate inconsistent rotations. The remaining relative rotations are input to a weighted least squares adjustment performed in the Lie algebra of the rotation manifold SO(3) to obtain absolute orientation parameters for each image. Weights are determined using the prior information we derived from the estimation of the relative rotations. Because we use the Lie algebra of SO(3) for averaging no subsequent adaptation of the results has to be performed but the lossless projection to the manifold. We evaluate our approach on synthetic and real data. Our approach often is able to detect and eliminate all outliers from the relative rotations even if very high outlier rates are present. We show that we improve the quality of the estimated absolute rotations by introducing individual weights for the relative rotations based on various indicators. In comparison with the state-of-the-art in recent publications to global image orientation we achieve best results in the examined datasets.
Weighted Average Consensus-Based Unscented Kalman Filtering.
Li, Wangyan; Wei, Guoliang; Han, Fei; Liu, Yurong
2016-02-01
In this paper, we are devoted to investigate the consensus-based distributed state estimation problems for a class of sensor networks within the unscented Kalman filter (UKF) framework. The communication status among sensors is represented by a connected undirected graph. Moreover, a weighted average consensus-based UKF algorithm is developed for the purpose of estimating the true state of interest, and its estimation error is bounded in mean square which has been proven in the following section. Finally, the effectiveness of the proposed consensus-based UKF algorithm is validated through a simulation example. PMID:26168453
Modified box dimension and average weighted receiving time on the weighted fractal networks
Dai, Meifeng; Sun, Yanqiu; Shao, Shuxiang; Xi, Lifeng; Su, Weiyi
2015-01-01
In this paper a family of weighted fractal networks, in which the weights of edges have been assigned to different values with certain scale, are studied. For the case of the weighted fractal networks the definition of modified box dimension is introduced, and a rigorous proof for its existence is given. Then, the modified box dimension depending on the weighted factor and the number of copies is deduced. Assuming that the walker, at each step, starting from its current node, moves uniformly to any of its nearest neighbors. The weighted time for two adjacency nodes is the weight connecting the two nodes. Then the average weighted receiving time (AWRT) is a corresponding definition. The obtained remarkable result displays that in the large network, when the weight factor is larger than the number of copies, the AWRT grows as a power law function of the network order with the exponent, being the reciprocal of modified box dimension. This result shows that the efficiency of the trapping process depends on the modified box dimension: the larger the value of modified box dimension, the more efficient the trapping process is. PMID:26666355
Modified box dimension and average weighted receiving time on the weighted fractal networks.
Dai, Meifeng; Sun, Yanqiu; Shao, Shuxiang; Xi, Lifeng; Su, Weiyi
2015-01-01
In this paper a family of weighted fractal networks, in which the weights of edges have been assigned to different values with certain scale, are studied. For the case of the weighted fractal networks the definition of modified box dimension is introduced, and a rigorous proof for its existence is given. Then, the modified box dimension depending on the weighted factor and the number of copies is deduced. Assuming that the walker, at each step, starting from its current node, moves uniformly to any of its nearest neighbors. The weighted time for two adjacency nodes is the weight connecting the two nodes. Then the average weighted receiving time (AWRT) is a corresponding definition. The obtained remarkable result displays that in the large network, when the weight factor is larger than the number of copies, the AWRT grows as a power law function of the network order with the exponent, being the reciprocal of modified box dimension. This result shows that the efficiency of the trapping process depends on the modified box dimension: the larger the value of modified box dimension, the more efficient the trapping process is. PMID:26666355
Modified box dimension and average weighted receiving time on the weighted fractal networks
NASA Astrophysics Data System (ADS)
Dai, Meifeng; Sun, Yanqiu; Shao, Shuxiang; Xi, Lifeng; Su, Weiyi
2015-12-01
In this paper a family of weighted fractal networks, in which the weights of edges have been assigned to different values with certain scale, are studied. For the case of the weighted fractal networks the definition of modified box dimension is introduced, and a rigorous proof for its existence is given. Then, the modified box dimension depending on the weighted factor and the number of copies is deduced. Assuming that the walker, at each step, starting from its current node, moves uniformly to any of its nearest neighbors. The weighted time for two adjacency nodes is the weight connecting the two nodes. Then the average weighted receiving time (AWRT) is a corresponding definition. The obtained remarkable result displays that in the large network, when the weight factor is larger than the number of copies, the AWRT grows as a power law function of the network order with the exponent, being the reciprocal of modified box dimension. This result shows that the efficiency of the trapping process depends on the modified box dimension: the larger the value of modified box dimension, the more efficient the trapping process is.
26 CFR 1.989(b)-1 - Definition of weighted average exchange rate.
Code of Federal Regulations, 2011 CFR
2011-04-01
... 26 Internal Revenue 10 2011-04-01 2011-04-01 false Definition of weighted average exchange rate. 1... of weighted average exchange rate. For purposes of section 989(b)(3) and (4), the term weighted average exchange rate means the simple average of the daily exchange rates (determined by reference to...
26 CFR 1.989(b)-1 - Definition of weighted average exchange rate.
Code of Federal Regulations, 2010 CFR
2010-04-01
... 26 Internal Revenue 10 2010-04-01 2010-04-01 false Definition of weighted average exchange rate. 1.989(b)-1 Section 1.989(b)-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES Export Trade Corporations 1.989(b)-1 Definition of weighted average exchange rate. For purposes...
47 CFR 65.305 - Calculation of the weighted average cost of capital.
Code of Federal Regulations, 2010 CFR
2010-10-01
..., each weighted by its proportion in the capital structure of the telephone companies. (b) Unless the... capital. 65.305 Section 65.305 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON... Carriers 65.305 Calculation of the weighted average cost of capital. (a) The composite weighted...
Cohen's Linearly Weighted Kappa Is a Weighted Average of 2 x 2 Kappas
ERIC Educational Resources Information Center
Warrens, Matthijs J.
2011-01-01
An agreement table with [n as an element of N is greater than or equal to] 3 ordered categories can be collapsed into n - 1 distinct 2 x 2 tables by combining adjacent categories. Vanbelle and Albert ("Stat. Methodol." 6:157-163, 2009c) showed that the components of Cohen's weighted kappa with linear weights can be obtained from these n - 1
76 FR 13580 - Bus Testing; Calculation of Average Passenger Weight and Test Vehicle Weight
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-14
... Federal Register (74 FR 51083) that incorporated brake performance and emissions tests into FTA's bus... Weight Per Person (See, ``Passenger Weight and Inspected Vessel Stability Requirements: Final Rule, 75 FR... Transportation (44 FR 11032). Executive Order 12866 requires agencies to regulate in the ``most...
Latent-variable approaches to the Jamesian model of importance-weighted averages.
Scalas, L Francesca; Marsh, Herbert W; Nagengast, Benjamin; Morin, Alexandre J S
2013-01-01
The individually importance-weighted average (IIWA) model posits that the contribution of specific areas of self-concept to global self-esteem varies systematically with the individual importance placed on each specific component. Although intuitively appealing, this model has weak empirical support; thus, within the framework of a substantive-methodological synergy, we propose a multiple-item latent approach to the IIWA model as applied to a range of self-concept domains (physical, academic, spiritual self-concepts) and subdomains (appearance, math, verbal self-concepts) in young adolescents from two countries. Tests considering simultaneously the effects of self-concept domains on trait self-esteem did not support the IIWA model. On the contrary, support for a normative group importance model was found, in which importance varied as a function of domains but not individuals. Individuals differentially weight the various components of self-concept; however, the weights are largely determined by normative processes, so that little additional information is gained from individual weightings. PMID:23150198
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-19
...The Federal Transit Administration (FTA) is hosting a conference call to address issues concerning its notice of proposed rulemaking (NPRM) regarding the calculation of average passenger weights and test vehicle weights. The purpose of this conference call is to allow the public to ask questions and to clarify any misunderstandings regarding the NPRM published on March 14, 2011 (76 FR 13580).......
77 FR 74452 - Bus Testing: Calculation of Average Passenger Weight and Test Vehicle Weight
Federal Register 2010, 2011, 2012, 2013, 2014
2012-12-14
... (GVWR) (74 FR 51083, October 5, 2009). The testing procedure simulated a 150 lb. weight for each seated... square feet (76 FR 13850, March 14, 2011). Subsequent to the NPRM, on July 6, 2012, Congress passed the..., Executive Order 13563, the Regulatory Flexibility Act, or the DOT Regulatory Policies and Procedures (44...
Sensitivity Analysis of Ordered Weighted Averaging Operator in Earthquake Vulnerability Assessment
NASA Astrophysics Data System (ADS)
Moradi, M.; Delavar, M. R.; Moshiri, B.
2013-09-01
The main objective of this research is to find the extent to which the minimal variability Ordered Weighted Averaging (OWA) model of seismic vulnerability assessment is sensitive to variation of optimism degree. There are a variety of models proposed for seismic vulnerability assessment. In order to examine the efficiency of seismic vulnerability assessment models, the stability of results could be analysed. Seismic vulnerability assessment is done to estimate the probable losses in the future earthquake. Multi-Criteria Decision Making (MCDM) methods have been applied by a number of researchers to estimate the human, physical and financial losses in urban areas. The study area of this research is Tehran Metropolitan Area (TMA) which has more than eight million inhabitants. In addition, this paper assumes that North Tehran Fault (NTF) is activated and caused an earthquake in TMA. 1996 census data is used to extract the attribute values for six effective criteria in seismic vulnerability assessment. The results demonstrate that minimal variability OWA model of Seismic Loss Estimation (SLE) is more stable where the aggregated seismic vulnerability degree has a lower value. Moreover, minimal variability OWA is very sensitive to optimism degree in northern areas of Tehran. A number of statistical units in southern areas of the city also indicate considerable sensitivity to optimism degree due to numerous non-standard buildings. In addition, the change of seismic vulnerability degree caused by variation of optimism degree does not exceed 25 % of the original value which means that the overall accuracy of the model is acceptable.
12 CFR 702.105 - Weighted-average life of investments.
Code of Federal Regulations, 2011 CFR
2011-01-01
..., as defined in 17 CFR 270.2a-7, and collective investment funds operated in accordance with short-term investment fund rules set forth in 12 CFR 9.18(b)(4)(ii)(B)(1)-(3), the weighted-average life is defined as... unions, as defined in 12 CFR 704.2, the weighted-average life is defined as greater than one (1)...
SIMPLE AND WEIGHTED AVERAGING APPROACHES TO SCALING: WHEN CAN SPATIAL CONTEXT BE IGNORED?
Technology Transfer Automated Retrieval System (TEKTRAN)
Scaling from plots to landscapes, landscapes to regions, and regions to the globe based on simple or weighted averaging techniques can be accurate when applied to the appropriate problems. Simple averaging approaches work well when conditions are homogeneous spatially and temporally. For example, ...
ERIC Educational Resources Information Center
Krijnen, Wim P.
1994-01-01
To assess association between rows of proximity matrices, H. de Vries (1993) introduces weighted average and row-wise average variants for Pearson's product-moment correlation, Spearman's rank correlation, and Kendall's rank correlation. For all three, the absolute value of the first variant is greater than or equal to the second. (SLD)
Predicting annual average particulate concentration in urban areas.
Progiou, Athena G; Ziomas, Ioannis C
2015-11-01
Particulate matter concentrations are in most cities a major environmental problem. This is also the case in Greece where, despite the various measures taken in the past, the problem still persists. In this aspect, a cost efficient, comprehensive method was developed in order to help decision makers to take the most appropriate measures towards particulates pollution abatement. The method is based on the source apportionment estimation from the application of 3D meteorological and dispersion modeling and is validated with the use of 10 years (2002-2012) PM10 monitoring data, in Athens, Greece, as well as using PM10 emission data for the same area and time period. It appears that the methodology can be used for estimating yearly average PM10 concentrations in a quite realistic manner, giving thus the decision makers the possibility to evaluate ex ante the effectiveness of specific abatement measures. PMID:26081738
NASA Technical Reports Server (NTRS)
Scruggs, T.; Moraguez, M.; Patankar, K.; Fitz-Coy, N.; Liou, J.-C.; Sorge, M.; Huynh, T.
2016-01-01
Debris fragments from the hypervelocity impact testing of DebriSat are being collected and characterized for use in updating existing satellite breakup models. One of the key parameters utilized in these models is the ballistic coefficient of the fragment which is directly related to its area-to-mass ratio. However, since the attitude of fragments varies during their orbital lifetime, it is customary to use the average cross-sectional area in the calculation of the area-to-mass ratio. The average cross-sectional area is defined as the average of the projected surface areas perpendicular to the direction of motion and has been shown to be equal to one-fourth of the total surface area of a convex object. Unfortunately, numerous fragments obtained from the DebriSat experiment show significant concavity (i.e., shadowing) and thus we have explored alternate methods for computing the average cross-sectional area of the fragments. An imaging system based on the volumetric reconstruction of a 3D object from multiple 2D photographs of the object was developed for use in determining the size characteristic (i.e., characteristics length) of the DebriSat fragments. For each fragment, the imaging system generates N number of images from varied azimuth and elevation angles and processes them using a space-carving algorithm to construct a 3D point cloud of the fragment. This paper describes two approaches for calculating the average cross-sectional area of debris fragments based on the 3D imager. Approach A utilizes the constructed 3D object to generate equally distributed cross-sectional area projections and then averages them to determine the average cross-sectional area. Approach B utilizes a weighted average of the area of the 2D photographs to directly compute the average cross-sectional area. A comparison of the accuracy and computational needs of each approach is described as well as preliminary results of an analysis to determine the "optimal" number of images needed for the 3D imager to accurately measure the average cross sectional area of objects with known dimensions.
Lu, Dan; Ye, Ming; Meyer, Philip D.; Curtis, Gary P.; Shi, Xiaoqing; Niu, Xu-Feng; Yabusaki, Steve B.
2013-01-01
When conducting model averaging for assessing groundwater conceptual model uncertainty, the averaging weights are often evaluated using model selection criteria such as AIC, AICc, BIC, and KIC (Akaike Information Criterion, Corrected Akaike Information Criterion, Bayesian Information Criterion, and Kashyap Information Criterion, respectively). However, this method often leads to an unrealistic situation in which the best model receives overwhelmingly large averaging weight (close to 100%), which cannot be justified by available data and knowledge. It was found in this study that this problem was caused by using the covariance matrix, CE, of measurement errors for estimating the negative log likelihood function common to all the model selection criteria. This problem can be resolved by using the covariance matrix, Cek, of total errors (including model errors and measurement errors) to account for the correlation between the total errors. An iterative two-stage method was developed in the context of maximum likelihood inverse modeling to iteratively infer the unknown Cek from the residuals during model calibration. The inferred Cek was then used in the evaluation of model selection criteria and model averaging weights. While this method was limited to serial data using time series techniques in this study, it can be extended to spatial data using geostatistical techniques. The method was first evaluated in a synthetic study and then applied to an experimental study, in which alternative surface complexation models were developed to simulate column experiments of uranium reactive transport. It was found that the total errors of the alternative models were temporally correlated due to the model errors. The iterative two-stage method using Cekresolved the problem that the best model receives 100% model averaging weight, and the resulting model averaging weights were supported by the calibration results and physical understanding of the alternative models. Using Cek obtained from the iterative two-stage method also improved predictive performance of the individual models and model averaging in both synthetic and experimental studies.
Lu, Dan; Ye, Ming; Meyer, Philip D.; Curtis, Gary P.; Shi, Xiaoqing; Niu, Xu-Feng; Yabusaki, Steven B.
2013-07-23
When conducting model averaging for assessing groundwater conceptual model uncertainty, the averaging weights are often evaluated using model selection criteria such as AIC, AICc, BIC, and KIC (Akaike Information Criterion, Corrected Akaike Information Criterion, Bayesian Information Criterion, and Kashyap Information Criterion, respectively). However, this method often leads to an unrealistic situation in which the best model receives overwhelmingly large averaging weight (close to 100%), which cannot be justified by available data and knowledge. It was found in this study that this problem was caused by using the covariance matrix, CE, of measurement errors for estimating the negative log likelihood function common to all the model selection criteria. This problem can be resolved by using the covariance matrix, Cek, of total errors (including model errors and measurement errors) to account for the correlation between the total errors. An iterative two-stage method was developed in the context of maximum likelihood inverse modeling to iteratively infer the unknown Cek from the residuals during model calibration. The inferred Cek was then used in the evaluation of model selection criteria and model averaging weights. While this method was limited to serial data using time series techniques in this study, it can be extended to spatial data using geostatistical techniques. The method was first evaluated in a synthetic study and then applied to an experimental study, in which alternative surface complexation models were developed to simulate column experiments of uranium reactive transport. It was found that the total errors of the alternative models were temporally correlated due to the model errors. The iterative two-stage method using Cek resolved the problem that the best model receives 100% model averaging weight, and the resulting model averaging weights were supported by the calibration results and physical understanding of the alternative models. Using Cek obtained from the iterative two-stage method also improved predictive performance of the individual models and model averaging in both synthetic and experimental studies.
12 CFR 702.105 - Weighted-average life of investments.
Code of Federal Regulations, 2014 CFR
2014-01-01
..., as defined in 17 CFR 270.2a-7, and collective investment funds operated in accordance with short-term investment fund rules set forth in 12 CFR 9.18(b)(4)(ii)(B)(1)-(3), the weighted-average life is defined as... and nonperpetual capital in corporate credit unions, as defined in 12 CFR 704.2, the...
12 CFR 702.105 - Weighted-average life of investments.
Code of Federal Regulations, 2013 CFR
2013-01-01
..., as defined in 17 CFR 270.2a-7, and collective investment funds operated in accordance with short-term investment fund rules set forth in 12 CFR 9.18(b)(4)(ii)(B)(1)-(3), the weighted-average life is defined as... and nonperpetual capital in corporate credit unions, as defined in 12 CFR 704.2, the...
12 CFR 702.105 - Weighted-average life of investments.
Code of Federal Regulations, 2012 CFR
2012-01-01
..., as defined in 17 CFR 270.2a-7, and collective investment funds operated in accordance with short-term investment fund rules set forth in 12 CFR 9.18(b)(4)(ii)(B)(1)-(3), the weighted-average life is defined as... and nonperpetual capital in corporate credit unions, as defined in 12 CFR 704.2, the...
[Weighted-averaging multi-planar reconstruction method for multi-detector row computed tomography].
Aizawa, Mitsuhiro; Nishikawa, Keiichi; Sasaki, Keita; Kobayashi, Norio; Yama, Mitsuru; Sano, Tsukasa; Murakami, Shin-ichi
2012-01-01
Development of multi-detector row computed tomography (MDCT) has enabled three-dimensions (3D) scanning with minute voxels. Minute voxels improve spatial resolution of CT images. At the same time, however, they increase image noise. Multi-planar reconstruction (MPR) is one of effective 3D-image processing techniques. The conventional MPR technique can adjust slice thickness of MPR images. When a thick slice is used, the image noise is decreased. In this case, however, spatial resolution is deteriorated. In order to deal with this trade-off problem, we have developed the weighted-averaging multi-planar reconstruction (W-MPR) technique to control the balance between the spatial resolution and noise. The weighted-average is determined by the Gaussian-type weighting function. In this study, we compared the performance of W-MPR with that of conventional simple-addition-averaging MPR. As a result, we could confirm that W-MPR can decrease the image noise without significant deterioration of spatial resolution. W-MPR can adjust freely the weight for each slice by changing the shape of the weighting function. Therefore, W-MPR can allow us to select a proper balance of spatial resolution and noise and at the same time produce suitable MPR images for observation of targeted anatomical structures. PMID:22277813
Modeling daily average stream temperature from air temperature and watershed area
NASA Astrophysics Data System (ADS)
Butler, N. L.; Hunt, J. R.
2012-12-01
Habitat restoration efforts within watersheds require spatial and temporal estimates of water temperature for aquatic species especially species that migrate within watersheds at different life stages. Monitoring programs are not able to fully sample all aquatic environments within watersheds under the extreme conditions that determine long-term habitat viability. Under these circumstances a combination of selective monitoring and modeling are required for predicting future geospatial and temporal conditions. This study describes a model that is broadly applicable to different watersheds while using readily available regional air temperature data. Daily water temperature data from thirty-eight gauges with drainage areas from 2 km2 to 2000 km2 in the Sonoma Valley, Napa Valley, and Russian River Valley in California were used to develop, calibrate, and test a stream temperature model. Air temperature data from seven NOAA gauges provided the daily maximum and minimum air temperatures. The model was developed and calibrated using five years of data from the Sonoma Valley at ten water temperature gauges and a NOAA air temperature gauge. The daily average stream temperatures within this watershed were bounded by the preceding maximum and minimum air temperatures with smaller upstream watersheds being more dependent on the minimum air temperature than maximum air temperature. The model assumed a linear dependence on maximum and minimum air temperature with a weighting factor dependent on upstream area determined by error minimization using observed data. Fitted minimum air temperature weighting factors were consistent over all five years of data for each gauge, and they ranged from 0.75 for upstream drainage areas less than 2 km2 to 0.45 for upstream drainage areas greater than 100 km2. For the calibration data sets within the Sonoma Valley, the average error between the model estimated daily water temperature and the observed water temperature data ranged from 0.7 C to 1.5 C for the different gauges. To test the model, the average water temperature was estimated at the six locations within the Sonoma Valley not used in the calibration. For each water temperature record, the prior area dependent weighting factor was used. Regional maximum and minimum air temperature data were then used to estimate the average stream water temperature over the period of recorded water temperature. The average error between model-estimated and observed water temperature for the additional locations in the Sonoma Valley ranged from 0.7 C to 3.5 C. The model estimated water temperature for gauges with upstream drainage area less than 50 km2 had average error between estimated and observed water temperature less than 1.7 C. When upstream drainage area was greater than 50 km2, the average error increased up to 3.5C for some gauge locations. The model could also estimate water temperature in streams in other basins using the same area-dependent weighting factor. For eighteen gauges in the Napa Valley to the east , the average error between estimated and observed water temperature ranged from 0.7 C to 1.9 C, while for four gauges in the Russian River Valley to the northwest, the average error ranged from 1.2 C to 3.2 C. We speculate the area-dependent weighting factor reflects the temperature of groundwater contributions to stream flow.
12 CFR 702.105 - Weighted-average life of investments.
Code of Federal Regulations, 2010 CFR
2010-01-01
..., as defined in 17 CFR 270.2a-7, and collective investment funds operated in accordance with short-term investment fund rules set forth in 12 CFR 9.18(b)(4)(ii)(B)(1)-(3), the weighted-average life is defined as...-in capital and membership capital in corporate credit unions, as defined in 12 CFR 704.2,...
Girshick, Ahna R.; Banks, Martin S.
2010-01-01
Depth perception involves combining multiple, possibly conflicting, sensory measurements to estimate the 3D structure of the viewed scene. Previous work has shown that the perceptual system combines measurements using a statistically optimal weighted average. However, the system should only combine measurements when they come from the same source. We asked whether the brain avoids combining measurements when they differ from one another: that is, whether the system is robust to outliers. To do this, we investigated how two slant cuesbinocular disparity and texture gradientsinfluence perceived slant as a function of the size of the conflict between the cues. When the conflict was small, we observed weighted averaging. When the conflict was large, we observed robust behavior: perceived slant was dictated solely by one cue, the other being rejected. Interestingly, the rejected cue was either disparity or texture, and was not necessarily the more variable cue. We modeled the data in a probabilistic framework, and showed that weighted averaging and robustness are predicted if the underlying likelihoods have heavier tails than Gaussians. We also asked whether observers had conscious access to the single-cue estimates when they exhibited robustness and found they did not, i.e. they completely fused despite the robust percepts. PMID:19761341
Real diffusion-weighted MRI enabling true signal averaging and increased diffusion contrast.
Eichner, Cornelius; Cauley, Stephen F; Cohen-Adad, Julien; Mller, Harald E; Turner, Robert; Setsompop, Kawin; Wald, Lawrence L
2015-11-15
This project aims to characterize the impact of underlying noise distributions on diffusion-weighted imaging. The noise floor is a well-known problem for traditional magnitude-based diffusion-weighted MRI (dMRI) data, leading to biased diffusion model fits and inaccurate signal averaging. Here, we introduce a total-variation-based algorithm to eliminate shot-to-shot phase variations of complex-valued diffusion data with the intention to extract real-valued dMRI datasets. The obtained real-valued diffusion data are no longer superimposed by a noise floor but instead by a zero-mean Gaussian noise distribution, yielding dMRI data without signal bias. We acquired high-resolution dMRI data with strong diffusion weighting and, thus, low signal-to-noise ratio. Both the extracted real-valued and traditional magnitude data were compared regarding signal averaging, diffusion model fitting and accuracy in resolving crossing fibers. Our results clearly indicate that real-valued diffusion data enables idealized conditions for signal averaging. Furthermore, the proposed method enables unbiased use of widely employed linear least squares estimators for model fitting and demonstrates an increased sensitivity to detect secondary fiber directions with reduced angular error. The use of phase-corrected, real-valued data for dMRI will therefore help to clear the way for more detailed and accurate studies of white matter microstructure and structural connectivity on a fine scale. PMID:26241680
Exponentially Weighted Moving Average Change Detection Around the Country (and the World)
NASA Astrophysics Data System (ADS)
Brooks, E.; Wynne, R. H.; Thomas, V. A.; Blinn, C. E.; Coulston, J.
2014-12-01
With continuous, freely available moderate-resolution imagery of the Earth's surface available, and with the promise of more imagery to come, change detection based on continuous process models continues to be a major area of research. One such method, exponentially weighted moving average change detection (EWMACD), is based on a mixture of harmonic regression (HR) and statistical quality control, a branch of statistics commonly used to detect aberrations in industrial and medical processes. By using HR to approximate per-pixel seasonal curves, the resulting residuals characterize information about the pixels which stands outside of the periodic structure imposed by HR. Under stable pixels, these residuals behave as might be expected, but in the presence of changes (growth, stress, removal), the residuals clearly show these changes when they are used as inputs into an EWMA chart. In prior work in Alabama, USA, EWMACD yielded an overall accuracy of 85% on a random sample of known thinned stands, in some cases detecting thinnings as sparse as 25% removal. It was also shown to correctly identify the timing of the thinning activity, typically within a single image date of the change. The net result of the algorithm was to produce date-by-date maps of afforestation and deforestation on a variable scale of severity. In other research, EWMACD has also been applied to detect land use and land cover changes in central Java, Indonesia, despite the heavy incidence of clouds and a monsoonal climate. Preliminary results show that EWMACD accurately identifies land use conversion (agricultural to residential, for example) and also identifies neighborhoods where the building density has increased, removing neighborhood vegetation. In both cases, initial results indicate the potential utility of EWMACD to detect both gross and subtle ecosystem disturbance, but further testing across a range of ecosystems and disturbances is clearly warranted.
Rong, Y; Sillick, M; Gregson, C M
2009-01-01
Dextrose equivalent (DE) value is the most common parameter used to characterize the molecular weight of maltodextrins. Its theoretical value is inversely proportional to number average molecular weight (M(n)), providing a theoretical basis for correlations with physical properties important to food manufacturing, such as: hygroscopicity, the glass transition temperature, and colligative properties. The use of freezing point osmometry to measure DE and M(n) was assessed. Measurements were made on a homologous series of malto-oligomers as well as a variety of commercially available maltodextrin products with DE values ranging from 5 to 18. Results on malto-oligomer samples confirmed that freezing point osmometry provided a linear response with number average molecular weight. However, noncarbohydrate species in some commercial maltodextrin products were found to be in high enough concentration to interfere appreciably with DE measurement. Energy dispersive spectroscopy showed that sodium and chloride were the major ions present in most commercial samples. Osmolality was successfully corrected using conductivity measurements to estimate ion concentrations. The conductivity correction factor appeared to be dependent on the concentration of maltodextrin. Equations were developed to calculate corrected values of DE and M(n) based on measurements of osmolality, conductivity, and maltodextrin concentration. This study builds upon previously reported results through the identification of the major interfering ions and provides an osmolality correction factor that successfully accounts for the influence of maltodextrin concentration on the conductivity measurement. The resulting technique was found to be rapid, robust, and required no reagents. PMID:19200083
A new state reconstructor for digital controls systems using weighted-average measurements
NASA Technical Reports Server (NTRS)
Polites, Michael E.
1989-01-01
A state reconstructor is presented for a linear continuous-time plant driven by a zero-order-hold. It takes a continuous-time output vector from the plant and convolutes it with a weighting-function matrix whose elements are time dependent. This result is integrated over T second intervals to generate weighted-averaged measurements, every T seconds, that are used in the state reconstruction process. If the plant is noise-free and can be modeled precisely, the output of this state reconstructor exactly equals the true state of the plant and accomplishes this without knowledge of the plant's initial state. If noise or modeling errors are a problem, it can be catenated with a state observer or a Kalman filter for a synergistic effect.
Calculation of weighted averages approach for the estimation of ping tolerance values
Silalom, S.; Carter, J.L.; Chantaramongkol, P.
2010-01-01
A biotic index was created and proposed as a tool to assess water quality in the Upper Mae Ping sub-watersheds. The Ping biotic index was calculated by utilizing Ping tolerance values. This paper presents the calculation of Ping tolerance values of the collected macroinvertebrates. Ping tolerance values were estimated by a weighted averages approach based on the abundance of macroinvertebrates and six chemical constituents that include conductivity, dissolved oxygen, biochemical oxygen demand, ammonia nitrogen, nitrate nitrogen and orthophosphate. Ping tolerance values range from 0 to 10. Macroinvertebrates assigned a 0 are very sensitive to organic pollution while macroinvertebrates assigned 10 are highly tolerant to pollution.
Correlation between weighted spectral distribution and average path length in evolving networks
NASA Astrophysics Data System (ADS)
Jiao, Bo; Shi, Jianmai; Wu, Xiaoqun; Nie, Yuanping; Huang, Chengdong; Du, Jing; Zhou, Ying; Guo, Ronghua; Tao, Yerong
2016-02-01
The weighted spectral distribution (WSD) is a metric defined on the normalized Laplacian spectrum. In this study, synchronic random graphs are first used to rigorously analyze the metric's scaling feature, which indicates that the metric grows sublinearly as the network size increases, and the metric's scaling feature is demonstrated to be common in networks with Gaussian, exponential, and power-law degree distributions. Furthermore, a deterministic model of diachronic graphs is developed to illustrate the correlation between the slope coefficient of the metric's asymptotic line and the average path length, and the similarities and differences between synchronic and diachronic random graphs are investigated to better understand the correlation. Finally, numerical analysis is presented based on simulated and real-world data of evolving networks, which shows that the ratio of the WSD to the network size is a good indicator of the average path length.
Pedersen, T.A.; LaVelle, J.M.
1997-12-31
Estimating average soil contaminant concentrations is a routine site investigation, risk assessment and remediation verification activity. Some techniques however fail to adequately consider the spatial distribution of contaminants and may result in erroneous estimates. Kriging may provides the best linear unbiased estimator of soil contaminant levels for sites when concentrations are spatially correlated. When variogram models show limited relationships, or when sampling density is limited, the application of geostatistical techniques may, not be appropriate. In these situations the geographic distribution of contaminants is often discounted when developing estimates of soil contaminant concentrations. In this poster presentation we describe the use of a Voronoi tessellation weighted averaging approach for estimating soil lead and polyaromatic hydrocarbons (PAHs) that in turn were used in performing risk assessment computations. The exposure point concentrations estimated using this approach provided a more realistic assessment of the risks actually posed by soil contaminants at this site.
Correlation between weighted spectral distribution and average path length in evolving networks.
Jiao, Bo; Shi, Jianmai; Wu, Xiaoqun; Nie, Yuanping; Huang, Chengdong; Du, Jing; Zhou, Ying; Guo, Ronghua; Tao, Yerong
2016-02-01
The weighted spectral distribution (WSD) is a metric defined on the normalized Laplacian spectrum. In this study, synchronic random graphs are first used to rigorously analyze the metric's scaling feature, which indicates that the metric grows sublinearly as the network size increases, and the metric's scaling feature is demonstrated to be common in networks with Gaussian, exponential, and power-law degree distributions. Furthermore, a deterministic model of diachronic graphs is developed to illustrate the correlation between the slope coefficient of the metric's asymptotic line and the average path length, and the similarities and differences between synchronic and diachronic random graphs are investigated to better understand the correlation. Finally, numerical analysis is presented based on simulated and real-world data of evolving networks, which shows that the ratio of the WSD to the network size is a good indicator of the average path length. PMID:26931591
Fuzzy weighted average based on left and right scores in Malaysia tourism industry
NASA Astrophysics Data System (ADS)
Kamis, Nor Hanimah; Abdullah, Kamilah; Zulkifli, Muhammad Hazim; Sahlan, Shahrazali; Mohd Yunus, Syaizzal
2013-04-01
Tourism is known as an important sector to the Malaysian economy including economic generator, creating business and job offers. It is reported to bring in almost RM30 billion of the national income, thanks to intense worldwide promotion by Tourism Malaysia. One of the well-known attractions in Malaysia is our beautiful islands. The islands continue to be developed into tourist spots and attracting a continuous number of tourists. Chalets, luxury bungalows and resorts quickly develop along the coastlines of popular islands like Tioman, Redang, Pangkor, Perhentian, Sibu and so many others. In this study, we applied Fuzzy Weighted Average (FWA) method based on left and right scores in order to determine the criteria weights and to select the best island in Malaysia. Cost, safety, attractive activities, accommodation and scenery are five main criteria to be considered and five selected islands in Malaysia are taken into accounts as alternatives. The most important criteria that have been considered by the tourist are defined based on criteria weights ranking order and the best island in Malaysia is then determined in terms of FWA values. This pilot study can be used as a reference to evaluate performances or solving any selection problems, where more criteria, alternatives and decision makers will be considered in the future.
Code of Federal Regulations, 2010 CFR
2010-04-01
... 19 Customs Duties 3 2010-04-01 2010-04-01 false De minimis net countervailable subsidies and... minimis net countervailable subsidies and weighted-average dumping margins disregarded. (a) Introduction... practice of disregarding net countervailable subsidies or weighted-average dumping margins that were...
47 CFR 54.1309 - National and study area average unseparated loop costs.
Code of Federal Regulations, 2014 CFR
2014-10-01
... 47 Telecommunication 3 2014-10-01 2014-10-01 false National and study area average unseparated... 54.1309 National and study area average unseparated loop costs. (a) National average unseparated loop... the Loop Costs for each study area in the country as calculated pursuant to 54.1308(a) divided...
NASA Astrophysics Data System (ADS)
Gader, Paul D.; Grandhi, Roopnath; Lee, Wen-Hsiung; Wilson, Joseph N.; Ho, Dominic K. C.
2004-09-01
An automated methodology for combining Ground Penetrating Radar features from different depths is presented and analyzed. GPR data from the NIITEK system are processed by a depth-dependent, adaptive whitening algorithm. Shape and contrast features, including compactness, solidity, eccentricity, and relative area are computed at the different depths. These features must be combined to make a decision as to the presence of a landmine at a specific location. Since many of the depths contain no useful information and the depths of the mines are unknown, a strategy based on sorting is used. In a previous work, sorted features were combined via a rule-based system. In the current paper, an automated algorithm that builds a decision rule from sets of Ordered Weighted Average (OWA) operators is described. The OWA operator sorts the feature values, weights them, and performs a weighted sum of the sorted values, resulting in a nonlinear combination of the feature values. The weights of the OWA operators are trained off-line in combination with those of a decision-making network and held fixed during testing. The combination of OWA operators and decision-making network is called a FOWA network. The FOWA network is compared to the rule-based method on real data taken from multiple collections at two outdoor test sites.
Excessive weight gain after pregnancy in urban areas: one important area to prevent diabetes.
Bhattarai, Madhur Dev; Singh, Dhruba Lall
2005-12-01
Diabetes has been found to be common in people, including the women of child bearing age, in urban Nepal. There is a custom of feeding the women with foods rich in calories and fats, as much as possible, after delivery. Weights of 200 women of urban Kathmandu before the first pregnancy and 6 months and 1 year after delivery were studied. The mean + SD of weights before pregnancy, six months after delivery and one year after delivery were 51.3 + 4.9, 67.2 + 6.3 and 63.4 + 6.4 respectively. Similarly, the mean + SD of body mass index (BMI) before pregnancy, six months after delivery and one year after delivery were 21.3 + 1.8, 27.9 + 2.5 and 26.7 + 2.8 respectively. The average increases in weight and BMI six months after delivery were 15.9 kg and 6.6; the increases from the basal values were statistically significant (p<0.01). Six months later the average increased weight and BMI decreased by 3.8 kg and 1.2 respectively. Thus, the average increases in weight and BMI one year after first pregnancy were 12.1 kg and 5.4; the increases from the basal values were statistically significant (p<0.01). The women seemed to have difficulty in losing weight gained. In view of the problem of overweight, the custom of excess feeding after delivery seems required to be discouraged in urban areas to prevent subsequent gestational and type 2 diabetes in women and the population. PMID:16519070
Merigó, José M.
2014-01-01
Linguistic variables are very useful to evaluate alternatives in decision making problems because they provide a vocabulary in natural language rather than numbers. Some aggregation operators for linguistic variables force the use of a symmetric and uniformly distributed set of terms. The need to relax these conditions has recently been posited. This paper presents the induced unbalanced linguistic ordered weighted average (IULOWA) operator. This operator can deal with a set of unbalanced linguistic terms that are represented using fuzzy sets. We propose a new order-inducing criterion based on the specificity and fuzziness of the linguistic terms. Different relevancies are given to the fuzzy values according to their uncertainty degree. To illustrate the behaviour of the precision-based IULOWA operator, we present an environmental assessment case study in which a multiperson multicriteria decision making model is applied. PMID:25136677
Effects of average molecular weight and concentration of polymer additive on friction and wear
Yoshida, Kazuo )
1990-04-01
Tribological behavior with oils containing polymethacrylates (PMAs) differing in average molecular weight, Mw, is examined in sliding concentrated contacts. At low loads, low PMA Mw or higher concentrations of PMA have a beneficial effect on wear, but at high loads, PMAs are detrimental. The beneficial effect is attributed to elastohydrodynamic film formation. The most important parameter is the number of PMA molecules per unit oil volume. The prowear action can be explained by the fact that PMA molecules may accumulate in the inlet region of the contact. The polymer accumulation may block the base oil entering the contact leading to oil starvation which in turn leads to severe contacts and increases in wear. This anomalous behavior may result from the competition between the prowear action and oil film formation. 13 refs.
Weighted averages of magnetization from magnetic field measurements: A fast interpretation tool
NASA Astrophysics Data System (ADS)
Fedi, Maurizio
2003-08-01
Magnetic anomalies may be interpreted in terms of weighted averages of magnetization (WAM) by a simple transformation. The WAM transformation consists of dividing at each measurement point the experimental magnetic field by a normalizing field, computed from a source volume with a homogeneous unit-magnetization. The transformation yields a straightforward link among source and field position vectors. A main WAM outcome is that sources at different depths appear well discriminated. Due to the symmetry of the problem, the higher the considered field altitude, the deeper the sources outlined by the transformation. This is shown for single and multi-source synthetic cases as well as for real data. We analyze the real case of Mt. Vulture volcano (Southern Italy), where the related anomaly strongly interferes with that from deep intrusive sources. The volcanic edifice is well identified. The deep source is estimated at about 9 km depth, in agreement with other results.
Marin, Lucas; Valls, Aida; Isern, David; Moreno, Antonio; Merigó, José M
2014-01-01
Linguistic variables are very useful to evaluate alternatives in decision making problems because they provide a vocabulary in natural language rather than numbers. Some aggregation operators for linguistic variables force the use of a symmetric and uniformly distributed set of terms. The need to relax these conditions has recently been posited. This paper presents the induced unbalanced linguistic ordered weighted average (IULOWA) operator. This operator can deal with a set of unbalanced linguistic terms that are represented using fuzzy sets. We propose a new order-inducing criterion based on the specificity and fuzziness of the linguistic terms. Different relevancies are given to the fuzzy values according to their uncertainty degree. To illustrate the behaviour of the precision-based IULOWA operator, we present an environmental assessment case study in which a multiperson multicriteria decision making model is applied. PMID:25136677
A Weight-Averaged Interpolation Method for Coupling Time-Accurate Rarefied and Continuum Flows
NASA Astrophysics Data System (ADS)
Diaz, Steven William
A novel approach to coupling rarefied and continuum flow regimes as a single, hybrid model is introduced. The method borrows from techniques used in the simulation of spray flows to interpolate Lagrangian point-particles onto an Eulerian grid in a weight-averaged sense. A brief overview of traditional methods for modeling both rarefied and continuum domains is given, and a review of the literature regarding rarefied/continuum flow coupling is presented. Details of the theoretical development of the method of weighted interpolation are then described. The method evaluates macroscopic properties at the nodes of a CFD grid via the weighted interpolation of all simulated molecules in a set surrounding the node. The weight factor applied to each simulated molecule is the inverse of the linear distance between it and the given node. During development, the method was applied to several preliminary cases, including supersonic flow over an airfoil, subsonic flow over tandem airfoils, and supersonic flow over a backward facing step; all at low Knudsen numbers. The main thrust of the research centered on the time-accurate expansion of a rocket plume into a near-vacuum. The method proves flexible enough to be used with various flow solvers, demonstrated by the use of Fluent as the continuum solver for the preliminary cases and a NASA-developed Large Eddy Simulation research code, WRLES, for the full lunar model. The method is applicable to a wide range of Mach numbers and is completely grid independent, allowing the rarefied and continuum solvers to be optimized for their respective domains without consideration of the other. The work presented demonstrates the validity, and flexibility of the method of weighted interpolation as a novel concept in the field of hybrid flow coupling. The method marks a significant divergence from current practices in the coupling of rarefied and continuum flow domains and offers a kernel on which to base an ongoing field of research. It has the potential to significantly increase the flexibility of hybrid rarefied/continuum flow analyses.
NASA Astrophysics Data System (ADS)
Nadi, S.; Delavar, M. R.
2011-06-01
This paper presents a generic model for using different decision strategies in multi-criteria, personalized route planning. Some researchers have considered user preferences in navigation systems. However, these prior studies typically employed a high tradeoff decision strategy, which used a weighted linear aggregation rule, and neglected other decision strategies. The proposed model integrates a pairwise comparison method and quantifier-guided ordered weighted averaging (OWA) aggregation operators to form a personalized route planning method that incorporates different decision strategies. The model can be used to calculate the impedance of each link regarding user preferences in terms of the route criteria, criteria importance and the selected decision strategy. Regarding the decision strategy, the calculated impedance lies between aggregations that use a logical "and" (which requires all the criteria to be satisfied) and a logical "or" (which requires at least one criterion to be satisfied). The calculated impedance also includes taking the average of the criteria scores. The model results in multiple alternative routes, which apply different decision strategies and provide users with the flexibility to select one of them en-route based on the real world situation. The model also defines the robust personalized route under different decision strategies. The influence of different decision strategies on the results are investigated in an illustrative example. This model is implemented in a web-based geographical information system (GIS) for Isfahan in Iran and verified in a tourist routing scenario. The results demonstrated, in real world situations, the validity of the route planning carried out in the model.
Robust HLLC Riemann solver with weighted average flux scheme for strong shock
NASA Astrophysics Data System (ADS)
Kim, Sung Don; Lee, Bok Jik; Lee, Hyoung Jin; Jeung, In-Seuck
2009-11-01
Many researchers have reported failures of the approximate Riemann solvers in the presence of strong shock. This is believed to be due to perturbation transfer in the transverse direction of shock waves. We propose a simple and clear method to prevent such problems for the Harten-Lax-van Leer contact (HLLC) scheme. By defining a sensing function in the transverse direction of strong shock, the HLLC flux is switched to the Harten-Lax-van Leer (HLL) flux in that direction locally, and the magnitude of the additional dissipation is automatically determined using the HLL scheme. We combine the HLLC and HLL schemes in a single framework using a switching function. High-order accuracy is achieved using a weighted average flux (WAF) scheme, and a method for v-shear treatment is presented. The modified HLLC scheme is named HLLC-HLL. It is tested against a steady normal shock instability problem and Quirk's test problems, and spurious solutions in the strong shock regions are successfully controlled.
Time-weighted average SPME analysis for in planta determination of cVOCs.
Sheehan, Emily M; Limmer, Matt A; Mayer, Philipp; Karlson, Ulrich Gosewinkel; Burken, Joel G
2012-03-20
The potential of phytoscreening for plume delineation at contaminated sites has promoted interest in innovative, sensitive contaminant sampling techniques. Solid-phase microextraction (SPME) methods have been developed, offering quick, undemanding, noninvasive sampling without the use of solvents. In this study, time-weighted average SPME (TWA-SPME) sampling was evaluated for in planta quantification of chlorinated solvents. TWA-SPME was found to have increased sensitivity over headspace and equilibrium SPME sampling. Using a variety of chlorinated solvents and a polydimethylsiloxane/carboxen (PDMS/CAR) SPME fiber, most compounds exhibited near linear or linear uptake over the sampling period. Smaller, less hydrophobic compounds exhibited more nonlinearity than larger, more hydrophobic molecules. Using a specifically designed in planta sampler, field sampling was conducted at a site contaminated with chlorinated solvents. Sampling with TWA-SPME produced instrument responses ranging from 5 to over 200 times higher than headspace tree core sampling. This work demonstrates that TWA-SPME can be used for in planta detection of a broad range of chlorinated solvents and methods can likely be applied to other volatile and semivolatile organic compounds. PMID:22332592
Wingard, G.L.; Hudley, J.W.
2012-01-01
A molluscan analogue dataset is presented in conjunction with a weighted-averaging technique as a tool for estimating past salinity patterns in south Florida’s estuaries and developing targets for restoration based on these reconstructions. The method, here referred to as cumulative weighted percent (CWP), was tested using modern surficial samples collected in Florida Bay from sites located near fixed water monitoring stations that record salinity. The results were calibrated using species weighting factors derived from examining species occurrence patterns. A comparison of the resulting calibrated species-weighted CWP (SW-CWP) to the observed salinity at the water monitoring stations averaged over a 3-year time period indicates, on average, the SW-CWP comes within less than two salinity units of estimating the observed salinity. The SW-CWP reconstructions were conducted on a core from near the mouth of Taylor Slough to illustrate the application of the method.
Area-averaged profiles over the mock urban setting test array
Nelson, M. A.; Brown, M. J.; Pardyjak, E. R.; Klewicki, J. C.
2004-01-01
Urban areas have a large effect on the local climate and meteorology. Efforts have been made to incorporate the bulk dynamic and thermodynamic effects of urban areas into mesoscale models (e.g., Chin et al., 2000; Holt et al., 2002; Lacser and Otte, 2002). At this scale buildings cannot be resolved individually, but parameterizations have been developed to capture their aggregate effect. These urban canopy parameterizations have been designed to account for the area-average drag, turbulent kinetic energy (TKE) production, and surface energy balance modifications due to buildings (e.g., Sorbjan and Uliasz, 1982; Ca, 1999; Brown, 2000; Martilli et al., 2002). These models compute an area-averaged mean profile that is representative of the bulk flow characteristics over the entire mesoscale grid cell. One difficulty has been testing of these parameterizations due to lack of area-averaged data. In this paper, area-averaged velocity and turbulent kinetic energy profiles are derived from data collected at the Mock Urban Setting Test (MUST). The MUST experiment was designed to be a near full-scale model of an idealized urban area imbedded in the Atmospheric Surface Layer (ASL). It's purpose was to study airflow and plume transport in urban areas and to provide a test case for model validation. A large number of velocity measurements were taken at the test site so that it was possible to derive area-averaged velocity and TKE profiles.
Full-custom design of split-set data weighted averaging with output register for jitter suppression
NASA Astrophysics Data System (ADS)
Jubay, M. C.; Gerasta, O. J.
2015-06-01
A full-custom design of an element selection algorithm, named as Split-set Data Weighted Averaging (SDWA) is implemented in 90nm CMOS Technology Synopsys Library. SDWA is applied in seven unit elements (3-bit) using a thermometer-coded input. Split-set DWA is an improved DWA algorithm which caters the requirement for randomization along with long-term equal element usage. Randomization and equal element-usage improve the spectral response of the unit elements due to higher Spurious-free dynamic range (SFDR) and without significantly degrading signal-to-noise ratio (SNR). Since a full-custom, the design is brought to transistor-level and the chip custom layout is also provided, having a total area of 0.3mm2, a power consumption of 0.566 mW, and simulated at 50MHz clock frequency. On this implementation, SDWA is successfully derived and improved by introducing a register at the output that suppresses the jitter introduced at the final stage due to switching loops and successive delays.
47 CFR 36.622 - National and study area average unseparated loop costs.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 47 Telecommunication 2 2010-10-01 2010-10-01 false National and study area average unseparated... Universal Service Fund Calculation of Loop Costs for Expense Adjustment 36.622 National and study area... provided in paragraph (c) of this section, this is equal to the sum of the Loop Costs for each study...
47 CFR 36.622 - National and study area average unseparated loop costs.
Code of Federal Regulations, 2012 CFR
2012-10-01
... 47 Telecommunication 2 2012-10-01 2012-10-01 false National and study area average unseparated...-Cost Loop Support Calculation of Loop Costs for Expense Adjustment 36.622 National and study area... provided in paragraph (c) of this section, this is equal to the sum of the Loop Costs for each study...
47 CFR 36.622 - National and study area average unseparated loop costs.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 47 Telecommunication 2 2011-10-01 2011-10-01 false National and study area average unseparated... Universal Service Fund Calculation of Loop Costs for Expense Adjustment 36.622 National and study area... provided in paragraph (c) of this section, this is equal to the sum of the Loop Costs for each study...
47 CFR 36.622 - National and study area average unseparated loop costs.
Code of Federal Regulations, 2013 CFR
2013-10-01
... 47 Telecommunication 2 2013-10-01 2013-10-01 false National and study area average unseparated...-Cost Loop Support Calculation of Loop Costs for Expense Adjustment 36.622 National and study area... provided in paragraph (c) of this section, this is equal to the sum of the Loop Costs for each study...
Prediction of oil palm production using the weighted average of fuzzy sets concept approach
NASA Astrophysics Data System (ADS)
Nugraha, R. F.; Setiyowati, Susi; Mukhaiyar, Utriweni; Yuliawati, Apriliani
2015-12-01
Proper planning becomes crucial for decision making in a company. For oil palm producer companies, the prediction of future products realizations is useful and considered in making company's strategies. It is mean that to do the best in predicting is absolute. Until now, to predict the next monthly oil palm productions, the company use simple mean statistics of the latest five-year observations. Lately, imprecision in estimates of oil palm production (overestimate) becomes a problem and the focus of attention in a company. Here we proposed weighted mean approach by using fuzzy concept approach to do estimation and prediction. We obtain that the prediction using fuzzy concept almost always give underestimate of realizations than the simple mean.
Appiani, Elena; Page, Sarah E; McNeill, Kristopher
2014-10-21
Dissolved organic matter (DOM) is involved in numerous environmental processes, and its molecular size is important in many of these processes, such as DOM bioavailability, DOM sorptive capacity, and the formation of disinfection byproducts during water treatment. The size and size distribution of the molecules composing DOM remains an open question. In this contribution, an indirect method to assess the average size of DOM is described, which is based on the reaction of hydroxyl radical (HO()) quenching by DOM. HO() is often assumed to be relatively unselective, reacting with nearly all organic molecules with similar rate constants. Literature values for HO() reaction with organic molecules were surveyed to assess the unselectivity of DOM and to determine a representative quenching rate constant (k(rep) = 5.6 10(9) M(-1) s(-1)). This value was used to assess the average molecular weight of various humic and fulvic acid isolates as model DOM, using literature HO() quenching constants, kC,DOM. The results obtained by this method were compared with previous estimates of average molecular weight. The average molecular weight (Mn) values obtained with this approach are lower than the Mn measured by other techniques such as size exclusion chromatography (SEC), vapor pressure osmometry (VPO), and flow field fractionation (FFF). This suggests that DOM is an especially good quencher for HO(), reacting at rates close to the diffusion-control limit. It was further observed that humic acids generally react faster than fulvic acids. The high reactivity of humic acids toward HO() is in line with the antioxidant properties of DOM. The benefit of this method is that it provides a firm upper bound on the average molecular weight of DOM, based on the kinetic limits of the HO() reaction. The results indicate low average molecular weight values, which is most consistent with the recent understanding of DOM. A possible DOM size distribution is discussed to reconcile the small nature of DOM with the large-molecule behavior observed in other studies. PMID:25222517
On the theory relating changes in area-average and pan evaporation (Invited)
NASA Astrophysics Data System (ADS)
Shuttleworth, W.; Serrat-Capdevila, A.; Roderick, M. L.; Scott, R.
2009-12-01
Theory relating changes in area-average evaporation with changes in the evaporation from pans or open water is developed. Such changes can arise by Type (a) processes related to large-scale changes in atmospheric concentrations and circulation that modify surface evaporation rates in the same direction, and Type (b) processes related to coupling between the surface and atmospheric boundary layer (ABL) at the landscape scale that usually modify area-average evaporation and pan evaporation in different directions. The interrelationship between evaporation rates in response to Type (a) changes is derived. They have the same sign and broadly similar magnitude but the change in area-average evaporation is modified by surface resistance. As an alternative to assuming the complementary evaporation hypothesis, the results of previous modeling studies that investigated surface-atmosphere coupling are parameterized and used to develop a theoretical description of Type (b) coupling via vapor pressure deficit (VPD) in the ABL. The interrelationship between appropriately normalized pan and area-average evaporation rates is shown to vary with temperature and wind speed but, on average, the Type (b) changes are approximately equal and opposite. Long-term Australian pan evaporation data are analyzed to demonstrate the simultaneous presence of Type (a) and (b) processes, and observations from three field sites in southwestern USA show support for the theory describing Type (b) coupling via VPD. England's victory over Australia in 2009 Ashes cricket test match series will not be mentioned.
NASA Astrophysics Data System (ADS)
Ishihara, Takemi
2015-12-01
The author has developed a new leveling method for use with magnetic survey data, which consists of adjusting each measurement using the weighted spatial average of its neighboring data and subsequent temporal filtering. There are two key parameters in the method: the `weight distance' represents the characteristic distance of the weight function and the `filtering width' represents the full width of the Gaussian filtering function on the time series. This new method was applied to three examples of actual marine survey data. Leveling using optimum values of these two parameters for each example was found to significantly reduce the standard deviations of crossover differences by one third to one fifth of the values before leveling. The obtained time series of correction values for each example had a good correlation with the magnetic observatory data obtained relatively close to the survey areas, thus validating this new leveling method.
Shih, H C; Tsai, S W; Kuo, C H
2012-01-01
A solid-phase microextraction (SPME) device was used as a diffusive sampler for airborne propylene glycol ethers (PGEs), including propylene glycol monomethyl ether (PGME), propylene glycol monomethyl ether acetate (PGMEA), and dipropylene glycol monomethyl ether (DPGME). Carboxen-polydimethylsiloxane (CAR/PDMS) SPME fiber was selected for this study. A polytetrafluoroethylene (PTFE) tubing was used as the holder, and the SPME fiber assembly was inserted into the tubing as a diffusive sampler. The diffusion path length and area of the sampler were 0.3 cm and 0.00086 cm(2), respectively. The theoretical sampling constants at 30C and 1 atm for PGME, PGMEA, and DPGME were 1.50 10(-2), 1.23 10(-2) and 1.14 10(-2) cm(3) min(-1), respectively. For evaluations, known concentrations of PGEs around the threshold limit values/time-weighted average with specific relative humidities (10% and 80%) were generated both by the air bag method and the dynamic generation system, while 15, 30, 60, 120, and 240 min were selected as the time periods for vapor exposures. Comparisons of the SPME diffusive sampling method to Occupational Safety and Health Administration (OSHA) organic Method 99 were performed side-by-side in an exposure chamber at 30C for PGME. A gas chromatography/flame ionization detector (GC/FID) was used for sample analysis. The experimental sampling constants of the sampler at 30C were (6.93 0.12) 10(-1), (4.72 0.03) 10(-1), and (3.29 0.20) 10(-1) cm(3) min(-1) for PGME, PGMEA, and DPGME, respectively. The adsorption of chemicals on the stainless steel needle of the SPME fiber was suspected to be one of the reasons why significant differences between theoretical and experimental sampling rates were observed. Correlations between the results for PGME from both SPME device and OSHA organic Method 99 were linear (r = 0.9984) and consistent (slope = 0.97 0.03). Face velocity (0-0.18 m/s) also proved to have no effects on the sampler. However, the effects of temperature and humidity have been observed. Therefore, adjustments of experimental sampling constants at different environmental conditions will be necessary. PMID:22651222
NASA Technical Reports Server (NTRS)
Shin, Kyung-Sup; Riba, Phil E.; North, Gerald R.
1990-01-01
This paper presents a new simple retrieval algorithm for estimating area-time averaged rain rates over tropical oceans by using single channel microwave measurements from satellites. The algorithm was tested by using the Nimbus-5 Electrically Scanning Microwave Radiometer and a simple microwave radiative transfer model to retrieve seasonal 5-deg x 5-deg area averaged rainrate over the tropical Atlantic and Pacific from December 1973 to November 1974. The brightness temperatures were collected and analyzed into histograms for each season and in each grid box from December 1973 to November 1974. The histograms suggest a normal distribution of background noise plus a skewed rain distribution at the higher brightness temperatures. By using a statistical estimation procedure based upon normally distributed background noise, the rain distribution was separated from the raw histogram. The radiative transfer model was applied to the rain-only distribution to retrieve area-time averaged rainrates throughout the tropics. Despite limitations of single channel information, the retrieved seasonal rain rates agree well in the open ocean with expectations based upon previous estimates of tropical rainfall over the oceans.
High surface area, low weight composite nickel fiber electrodes
NASA Technical Reports Server (NTRS)
Johnson, Bradley A.; Ferro, Richard E.; Swain, Greg M.; Tatarchuk, Bruce J.
1993-01-01
The energy density and power density of light weight aerospace batteries utilizing the nickel oxide electrode are often limited by the microstructures of both the collector and the resulting active deposit in/on the collector. Heretofore, these two microstructures were intimately linked to one another by the materials used to prepare the collector grid as well as the methods and conditions used to deposit the active material. Significant weight and performance advantages were demonstrated by Britton and Reid at NASA-LeRC using FIBREX nickel mats of ca. 28-32 microns diameter. Work in our laboratory investigated the potential performance advantages offered by nickel fiber composite electrodes containing a mixture of fibers as small as 2 microns diameter (Available from Memtec America Corporation). These electrode collectors possess in excess of an order of magnitude more surface area per gram of collector than FIBREX nickel. The increase in surface area of the collector roughly translates into an order of magnitude thinner layer of active material. Performance data and advantages of these thin layer structures are presented. Attributes and limitations of their electrode microstructure to independently control void volume, pore structure of the Ni(OH)2 deposition, and resulting electrical properties are discussed.
An analysis of the threshold method for measuring area-average rainfall
NASA Technical Reports Server (NTRS)
Kedem, Benjamin; Chiu, Long S.; Karni, Zvi
1990-01-01
Experimental evidence shows that the area-average rain rate and the fractional area covered by rain rate exceeding a fixed threshold are highly correlated; that is, are highly linearly related. A precise theoretical explanation of this fact is given. The explanation is based on the observation that rain rate has a mixed distribution, one that is a mixture of a discrete distribution and a continuous distribution. Under a homogeneity assumption, the slope of the linear relationship depends only on the continuous part of the distribution and as such is found to be markedly immune to parameter changes. This is illustrated by certain slope surfaces obtained from three specific distributions. The threshold level can be chosen in an optimal way by minimizing a certain distance function defined over the threshold range. In general, the threshold level should be not too far from the mean rain rate conditional on rain. The so-called threshold method advocates measuring rainfall from fractional area exploiting the observed linear relationship of the later with the area average rain rate. The method is potentially useful for the estimation of rainfall from space via satellites.
Estimation of the Area of a Reverberant Plate Using Average Reverberation Properties
NASA Astrophysics Data System (ADS)
Achdjian, Hossep; Moulin, Emmanuel; Benmeddour, Farouk; Assaad, Jamal
This paper aims to present an original method for the estimation of the area of thin plates of arbitrary geometrical shapes. This method relies on the acquisition and ensemble processing of reverberated elastic signals on few sensors. The acoustical Green's function in a reverberant solid medium is modeled by a nonstationary random process based on the image-sources method. In that way, mathematical expectations of the signal envelopes can be analytically related to reverberation properties and structural parameters such as plate area, group velocity, or source-receiver distance. Then, a simple curve fitting applied to an ensemble average over N realizations of the late envelopes allows to estimate a global term involving the values of structural parameters. From simple statistical modal arguments, it is shown that the obtained relation depends on the plate area and not on the plate shape. Finally, by considering an additional relation obtained from the early characteristics (treated in a deterministic way) of the reverberation signals, it is possible to deduce the area value. This estimation is performed without geometrical measurements and requires an access to only a small portion of the plate. Furthermore, this method does not require any time measurement nor trigger synchronization between the input channels of instrumentation (between measured signals), thus implying low hardware constraints. Experimental results obtained on metallic plates with free boundary conditions and embedded window glasses will be presented. Areas of up to several meter-squares are correctly estimated with a relative error of a few percents.
Coombes, Brandon; Basu, Saonli; Guha, Sharmistha; Schork, Nicholas
2015-01-01
Multi-locus effect modeling is a powerful approach for detection of genes influencing a complex disease. Especially for rare variants, we need to analyze multiple variants together to achieve adequate power for detection. In this paper, we propose several parsimonious branching model techniques to assess the joint effect of a group of rare variants in a case-control study. These models implement a data reduction strategy within a likelihood framework and use a weighted score test to assess the statistical significance of the effect of the group of variants on the disease. The primary advantage of the proposed approach is that it performs model-averaging over a substantially smaller set of models supported by the data and thus gains power to detect multi-locus effects. We illustrate these proposed approaches on simulated and real data and study their performance compared to several existing rare variant detection approaches. The primary goal of this paper is to assess if there is any gain in power to detect association by averaging over a number of models instead of selecting the best model. Extensive simulations and real data application demonstrate the advantage the proposed approach in presence of causal variants with opposite directional effects along with a moderate number of null variants in linkage disequilibrium. PMID:26436424
Area preserving map models of gyro-averaged ExB chaotic transport
NASA Astrophysics Data System (ADS)
Fonseca, J.; Del-Castillo-Negrete, D.
2013-10-01
Area preserving maps have been extensively used to model 2-dimensional chaotic transport in plasmas. Here we focus on drift-wave maps describing chaotic transport in E B zonal flows perturbed by electrostatic drift waves. Going beyond previous studies, we include finite Larmor radius (FLR) effects by gyro-averaging the corresponding Hamiltonian of the map. In the limit of zero Larmor radius, the gyro-averaged map reduces to the standard Chirikov map in the case of monotonic E B shear flows, and to the standard non-twist map in the case of non-monotonic E B shear flows. Like in the case of continuous E B drift wave models, we show that in the gyro-averaged maps, FLR effects lead to chaos suppression, bifurcation of the shearless curve, and a complex phase space topology. Dynamical systems methods are used to quantify the dependence on the Larmor radius of the threshold for the destruction of transport barriers, and the transport properties of an ensemble of test particles with a Maxwellian distribution. J.F. acknowledges support from the Brazilian government agency FAPESP. D. dCN acknowledges support from the U.S. Department of Energy.
Quantum black hole wave packet: Average area entropy and temperature dependent width
NASA Astrophysics Data System (ADS)
Davidson, Aharon; Yellin, Ben
2014-09-01
A quantum Schwarzschild black hole is described, at the mini super spacetime level, by a non-singular wave packet composed of plane wave eigenstates of the momentum Dirac-conjugate to the mass operator. The entropy of the mass spectrum acquires then independent contributions from the average mass and the width. Hence, Bekenstein's area entropy is formulated using the
Heo, Seo Weon; Kim, Hyungsuk
2010-05-01
An estimation of ultrasound attenuation in soft tissues is critical in the quantitative ultrasound analysis since it is not only related to the estimations of other ultrasound parameters, such as speed of sound, integrated scatterers, or scatterer size, but also provides pathological information of the scanned tissue. However, estimation performances of ultrasound attenuation are intimately tied to the accurate extraction of spectral information from the backscattered radiofrequency (RF) signals. In this paper, we propose two novel techniques for calculating a block power spectrum from the backscattered ultrasound signals. These are based on the phase-compensation of each RF segment using the normalized cross-correlation to minimize estimation errors due to phase variations, and the weighted averaging technique to maximize the signal-to-noise ratio (SNR). The simulation results with uniform numerical phantoms demonstrate that the proposed method estimates local attenuation coefficients within 1.57% of the actual values while the conventional methods estimate those within 2.96%. The proposed method is especially effective when we deal with the signal reflected from the deeper depth where the SNR level is lower or when the gated window contains a small number of signal samples. Experimental results, performed at 5MHz, were obtained with a one-dimensional 128 elements array, using the tissue-mimicking phantoms also show that the proposed method provides better estimation results (within 3.04% of the actual value) with smaller estimation variances compared to the conventional methods (within 5.93%) for all cases considered. PMID:20083291
Chi, Chang-Feng; Cao, Zi-Hao; Wang, Bin; Hu, Fa-Yuan; Li, Zhong-Rui; Zhang, Bin
2014-01-01
In the current study, the relationships between functional properties and average molecular weight (AMW) of collagen hydrolysates from Spanish mackerel (Scomberomorous niphonius) skin were researched. Seven hydrolysate fractions (5.04 ? AMW ? 47.82 kDa) from collagen of Spanish mackerel skin were obtained through the processes of acid extraction, proteolysis, and fractionation using gel filtration chromatography. The physicochemical properties of the collagen hydrolysate fractions were studied by sodium dodecyl sulfate polyacrylamide gel electrophoresis (SDS-PAGE), gel filtration chromatography, scanning electron microscope (SEM) and Fourier transform infrared spectroscopy (FTIR). The results indicated that there was an inverse relationship between the antioxidant activities and the logarithm of the AMW of the hydrolysate fractions in the tested AMW range. However, the reduction of AMW significantly enhanced the solubility of the hydrolysate fractions, and a similar AMW decrease of the hydrolysate fractions negatively affected the emulsifying and foaming capacities. This presented as a positive correlation between the logarithm of AMW and emulsion stability index, emulsifying activity index, foam stability, and foam capacity. Therefore, these collagen hydrolysates with excellent antioxidant activities or good functionalities as emulsifiers could be obtained by controlling the effect of the digestion process on the AMW of the resultant hydrolysates. PMID:25090114
NASA Astrophysics Data System (ADS)
Davies, G. R.; Chaplin, W. J.; Elsworth, Y.; Hale, S. J.
2014-07-01
The Birmingham Solar Oscillations Network (BiSON) has provided high-quality high-cadence observations from as far back in time as 1978. These data must be calibrated from the raw observations into radial velocity and the quality of the calibration has a large impact on the signal-to-noise ratio of the final time series. The aim of this work is to maximize the potential science that can be performed with the BiSON data set by optimizing the calibration procedure. To achieve better levels of signal-to-noise ratio, we perform two key steps in the calibration process: we attempt a correction for terrestrial atmospheric differential extinction; and the resulting improvement in the calibration allows us to perform weighted averaging of contemporaneous data from different BiSON stations. The improvements listed produce significant improvement in the signal-to-noise ratio of the BiSON frequency-power spectrum across all frequency ranges. The reduction of noise in the power spectrum will allow future work to provide greater constraint on changes in the oscillation spectrum with solar activity. In addition, the analysis of the low-frequency region suggests that we have achieved a noise level that may allow us to improve estimates of the upper limit of g-mode amplitudes.
NASA Astrophysics Data System (ADS)
Gasser, Guy; Pankratov, Irena; Elhanany, Sara; Glazman, Hillel; Lev, Ovadia
2014-05-01
A methodology used to estimate the percentage of wastewater effluent in an otherwise pristine water site is proposed on the basis of the weighted mean of the level of a consortium of indicator pollutants. This method considers the levels of uncertainty in the evaluation of each of the indicators in the site, potential effluent sources, and uncontaminated surroundings. A detailed demonstrative study was conducted on a site that is potentially subject to wastewater leakage. The research concentrated on several perched springs that are influenced to an unknown extent by agricultural communities. A comparison was made to a heavily contaminated site receiving wastewater effluent and surface water runoff. We investigated six springs in two nearby ridges where fecal contamination was detected in the past; the major sources of pollution in the area have since been diverted to a wastewater treatment system. We used chloride, acesulfame, and carbamazepine as domestic pollution tracers. Good correlation (R2 > 0.86) was observed between the mixing ratio predictions based on the two organic tracers (the slope of the linear regression was 1.05), whereas the chloride predictions differed considerably. This methodology is potentially useful, particularly for cases in which detailed hydrological modeling is unavailable but in which quantification of wastewater penetration is required. We demonstrate that the use of more than one tracer for estimation of the mixing ratio reduces the combined uncertainty level associated with the estimate and can also help to disqualify biased tracers.
NASA Astrophysics Data System (ADS)
Zhang, Leiming; Brook, Jeffrey R.
A method for deriving the site-specific and subgrid area wind speed and friction velocity from regional model output and detailed land type information is developed. The "subgrid velocity scale" is introduced to account for generation of turbulent fluxes by subgrid motions. The grid vector averaged wind speed is adjusted by adding the subgrid velocity scale. This is to account for the fact that the spatial average of the local wind speed is usually larger than the absolute value of the vector averaged velocity ( | limitV?| ), especially when there are different land or surface types within the spatial averaging area and when limitV? is small. The assumption of uu*=constant is then applied within a model grid area to obtain wind speed and friction velocity for specific sites and subgrid areas. Using this method, the site-specific and subgrid area wind speed and friction velocity can be estimated from grid-averaged model output. In addition, more realistic air pollutant dry deposition velocities for specific locations and subgrid areas can be calculated. Grid-averaged deposition velocity values calculated using this approach tend to be about 30% different (either larger or smaller) for HNO 3 and sulphate and about 10% different for SO 2 and O 3 compared to values calculated by assuming a constant wind speed over the whole model grid area. These differences are found to be even larger at specific sites or over some subgrid areas. This method can be applied to determine a more realistic wind speed, friction velocity and pollutant dry deposition velocity at specific locations using gridded meteorological data.
NASA Technical Reports Server (NTRS)
Markley, F. Landis; Cheng, Yang; Crassidis, John L.; Oshman, Yaakov
2007-01-01
Many applications require an algorithm that averages quaternions in an optimal manner. For example, when combining the quaternion outputs of multiple star trackers having this output capability, it is desirable to properly average the quaternions without recomputing the attitude from the the raw star tracker data. Other applications requiring some sort of optimal quaternion averaging include particle filtering and multiple-model adaptive estimation, where weighted quaternions are used to determine the quaternion estimate. For spacecraft attitude estimation applications, derives an optimal averaging scheme to compute the average of a set of weighted attitude matrices using the singular value decomposition method. Focusing on a 4-dimensional quaternion Gaussian distribution on the unit hypersphere, provides an approach to computing the average quaternion by minimizing a quaternion cost function that is equivalent to the attitude matrix cost function Motivated by and extending its results, this Note derives an algorithm that deterniines an optimal average quaternion from a set of scalar- or matrix-weighted quaternions. Rirthermore, a sufficient condition for the uniqueness of the average quaternion, and the equivalence of the mininiization problem, stated herein, to maximum likelihood estimation, are shown.
Code of Federal Regulations, 2012 CFR
2012-07-01
... weighted average in Equation 2 of § 63.2840 to determine the compliance ratio. (b) To determine the volume... determine chemical properties of the solvent and the volume percentage of all HAP components present in the... by the total volume of all deliveries as expressed in Equation 1 of this section. Record the...
Kassianov, Evgueni I.; Barnard, James C.; Flynn, Connor J.; Riihimaki, Laura D.; Marinovici, Maria C.
2015-10-15
Areal-averaged albedos are particularly difficult to measure in coastal regions, because the surface is not homogenous, consisting of a sharp demarcation between land and water. With this difficulty in mind, we evaluate a simple retrieval of areal-averaged surface albedo using ground-based measurements of atmospheric transmission alone under fully overcast conditions. To illustrate the performance of our retrieval, we find the areal-averaged albedo using measurements from the Multi-Filter Rotating Shadowband Radiometer (MFRSR) at five wavelengths (415, 500, 615, 673, and 870 nm). These MFRSR data are collected at a coastal site in Graciosa Island, Azores supported by the U.S. Department of Energy’s (DOE’s) Atmospheric Radiation Measurement (ARM) Program. The areal-averaged albedos obtained from the MFRSR are compared with collocated and coincident Moderate Resolution Imaging Spectroradiometer (MODIS) white-sky albedo at four nominal wavelengths (470, 560, 670 and 860 nm). These comparisons are made during a 19-month period (June 2009 - December 2010). We also calculate composite-based spectral values of surface albedo by a weighted-average approach using estimated fractions of major surface types observed in an area surrounding this coastal site. Taken as a whole, these three methods of finding albedo show spectral and temporal similarities, and suggest that our simple, transmission-based technique holds promise, but with estimated errors of about ±0.03. Additional work is needed to reduce this uncertainty in areas with inhomogeneous surfaces.
NASA Astrophysics Data System (ADS)
Kassianov, Evgueni; Barnard, James; Flynn, Connor; Riihimaki, Laura; Marinovici, Cristina
2015-10-01
Areal-averaged albedos are particularly difficult to measure in coastal regions, because the surface is not homogenous, consisting of a sharp demarcation between land and water. With this difficulty in mind, we evaluate a simple retrieval of areal-averaged surface albedo using ground-based measurements of atmospheric transmission alone under fully overcast conditions. To illustrate the performance of our retrieval, we find the areal-averaged albedo using measurements from the Multi-Filter Rotating Shadowband Radiometer (MFRSR) at five wavelengths (415, 500, 615, 673, and 870 nm). These MFRSR data are collected at a coastal site in Graciosa Island, Azores supported by the U.S. Department of Energy's (DOE's) Atmospheric Radiation Measurement (ARM) Program. The areal-averaged albedos obtained from the MFRSR are compared with collocated and coincident Moderate Resolution Imaging Spectroradiometer (MODIS) whitesky albedo at four nominal wavelengths (470, 560, 670 and 860 nm). These comparisons are made during a 19-month period (June 2009 - December 2010). We also calculate composite-based spectral values of surface albedo by a weighted-average approach using estimated fractions of major surface types observed in an area surrounding this coastal site. Taken as a whole, these three methods of finding albedo show spectral and temporal similarities, and suggest that our simple, transmission-based technique holds promise, but with estimated errors of about 0.03. Additional work is needed to reduce this uncertainty in areas with inhomogeneous surfaces.
ERIC Educational Resources Information Center
Warne, Russell T.; Nagaishi, Chanel; Slade, Michael K.; Hermesmeyer, Paul; Peck, Elizabeth Kimberli
2014-01-01
While research has shown the statistical significance of high school grade point averages (HSGPAs) in predicting future academic outcomes, the systems with which HSGPAs are calculated vary drastically across schools. Some schools employ unweighted grades that carry the same point value regardless of the course in which they are earned; other…
Thompson, Amanda L; Adair, Linda; Bentley, Margaret E
2014-01-01
Biomedical researchers have raised concerns that mothers’ inability to recognize infant and toddler overweight poses a barrier to stemming increasing rates of overweight and obesity, particularly among low-income or minority mothers. Little anthropological research has examined the sociocultural, economic or structural factors shaping maternal perceptions of infant and toddler size or addressed biomedical depictions of maternal misperception as a “socio-cultural problem.” We use qualitative and quantitative data from 237 low-income, African-American mothers to explore how they define ‘normal’ infant growth and infant overweight. Our quantitative results document that mothers’ perceptions of infant size change with infant age, are sensitive to the size of other infants in the community, and are associated with concerns over health and appetite. Qualitative analysis documents that mothers are concerned with their children’s weight status and assess size in relation to their infants’ cues, local and societal norms of appropriate size, interactions with biomedicine, and concerns about infant health and sufficiency. These findings suggest that mothers use multiple models to interpret and respond to child weight. An anthropological focus on the complex social and structural factors shaping what is considered ‘normal’ and ‘abnormal’ infant weight is critical for shaping appropriate and successful interventions. PMID:25684782
Sether, Bradley A.; Berkas, Wayne R.; Vecchia, Aldo V.
2004-01-01
Data were collected at 11 water-quality sampling sites in the upper Red River of the North (Red River) Basin from May 1997 through September 1999 to describe the water-quality characteristics of the upper Red River and to estimate constituent loads and flow-weighted average concentrations for major tributaries of the Red River upstream from the bridge crossing the Red River at Perley, Minn. Samples collected from the sites were analyzed for 5-day biochemical oxygen demand, bacteria, dissolved solids, nutrients, and suspended sediment. Concentration data indicated the median concentrations for most constituents and sampling sites during the study period were less than existing North Dakota and Minnesota standards or guidelines. However, more than 25 percent of the samples for the Red River at Perley, Minn., site had fecal coliform concentrations that were greater than 200 colonies per 100 milliliters, indicating an abundance of pathogens in the upper Red River Basin. Although total nitrite plus nitrate concentrations generally increased in a downstream direction, the median concentrations for all sites were less than the North Dakota suggested guideline of 1.0 milligram per liter. Total and dissolved phosphorus concentrations also generally increased in a downstream direction, but, for those constituents, the median concentrations for most sampling sites exceeded the North Dakota suggested guideline of 0.1 milligram per liter. For dissolved solids, nutrients, and suspended sediments, a relation between constituent concentration and streamflow was determined using the data collected during the study period. The relation was determined by a multiple regression model in which concentration was the dependent variable and streamflow was the primary explanatory variable. The regression model was used to compute unbiased estimates of annual loads for each constituent and for each of eight primary water-quality sampling sites and to compute the degree of uncertainty associated with each estimated annual load. The estimated annual loads for the eight primary sites then were used to estimate annual loads for five intervening reaches in the study area. Results were used as a screening tool to identify which subbasins contributed a disproportionate amount of pollutants to the Red River. To compare the relative water quality of the different subbasins, an estimated flow-weighted average (FWA) concentration was computed from the estimated average annual load and the average annual streamflow for each subbasin. The 5-day biochemical oxygen demands in the upper Red River Basin were fairly small, and medians ranged from 1 to 3 milligrams per liter. The largest estimated FWA concentration for dissolved solids (about 630 milligrams per liter) was for the Bois de Sioux River near Doran, Minn., site. The Otter Tail River above Breckenridge, Minn., site had the smallest estimated FWA concentration (about 240 milligrams per liter). The estimated FWA concentrations for dissolved solids for the main-stem sites ranged from about 300 to 500 milligrams per liter and generally increased in a downstream direction. The estimated FWA concentrations for total nitrite plus nitrate for the main-stem sites increased from about 0.2 milligram per liter for the Red River below Wahpeton, N. Dak., site to about 0.9 milligram per liter for the Red River at Perley, Minn., site. Much of the increase probably resulted from flows from the tributary sites and intervening reaches, excluding the Otter Tail River above Breckenridge, Minn., site. However, uncertainty in the estimated concentrations prevented any reliable conclusions regarding which sites or reaches contributed most to the increase. The estimated FWA concentrations for total ammonia for the main-stem sites increased from about 0.05 milligram per liter for the Red River above Fargo, N. Dak., site to about 0.15 milligram per liter for the Red River near Harwood, N. Dak., site. T
Survey of one-year average Rn levels in Pittsburgh area homes.
Cohen, B L
1985-12-01
Radon concentrations averaged over 1 yr were measured in 169 homes selected without regard to factors known to influence Rn levels. The average level was 6.3 pCi/l in basements, 2.4 pCi/l on first floors, and 2.0 pCi/l on second floors; medians were considerably smaller, 2.6, 1.20, and 0.95 pCi/l, respectively. Distributions deviate considerably from log normal in having an excess of very large values and a deficiency of very low values. Levels in upper floors were somewhat lower where the basement was isolated by closed doors than where it was not. Drafty houses had lower Rn levels than less drafty ones. Houses exposed to wind had higher levels than those well sheltered from wind. There was little correlation between Rn concentration and age of the house, indicating that Rn exposures may not have changed appreciably during the past 50 yr; this is important for estimating risks of Rn-induced lung cancer. Houses with forced air heating systems seemed to have higher Rn levels in the upper floors than those with steam or hot water heat. No strong correlations were found between Rn levels and cooking with natural gas, or river vs well-water supplies. PMID:4077511
Area-to-point parameter estimation with geographically weighted regression
NASA Astrophysics Data System (ADS)
Murakami, Daisuke; Tsutsumi, Morito
2015-07-01
The modifiable areal unit problem (MAUP) is a problem by which aggregated units of data influence the results of spatial data analysis. Standard GWR, which ignores aggregation mechanisms, cannot be considered to serve as an efficient countermeasure of MAUP. Accordingly, this study proposes a type of GWR with aggregation mechanisms, termed area-to-point (ATP) GWR herein. ATP GWR, which is closely related to geostatistical approaches, estimates the disaggregate-level local trend parameters by using aggregated variables. We examine the effectiveness of ATP GWR for mitigating MAUP through a simulation study and an empirical study. The simulation study indicates that the method proposed herein is robust to the MAUP when the spatial scales of aggregation are not too global compared with the scale of the underlying spatial variations. The empirical studies demonstrate that the method provides intuitively consistent estimates.
Numerous urban canopy schemes have recently been developed for mesoscale models in order to approximate the drag and turbulent production effects of a city on the air flow. However, little data exists by which to evaluate the efficacy of the schemes since "area-averaged&quo...
NASA Astrophysics Data System (ADS)
Pelgrum, H.; Bastiaanssen, W. G. M.
1996-04-01
A knowledge of the area-averaged latent heat flux <λE> is necessary to validate large-scale model predictions of heat fluxes over heterogeneous land surfaces. This paper describes different procedures to obtain <λE> as a weighted average of ground-based observations. The weighting coefficients are obtained from remote sensing measurements. The remote sensing data used in this study consist of a Landsat thematic mapper image of the European Field Experiment in a Desertification-Threatened Area (EFEDA) grid box in central Spain, acquired on June 12, 1991. A newly developed remote sensing algorithm, the surface energy balance for land algorithm (SEBAL), solves the energy budget on a pixel-by-pixel basis. From the resulting frequency distribution of the latent heat flux, the area-averaged latent heat flux was calculated as <λE> = 164 W m-2. This method was validated with field measurements of latent heat flux, sensible heat flux, and soil moisture. In general, the SEBAL-derived output compared well with field measurements. Two other methods for retrieval of weighting coefficients were tested against SEBAL. The second method combines satellite images of surface temperature, surface albedo, and normalized difference vegetation index (NDVI) into an index on a pixel-by-pixel basis. After inclusion of ground-based measurements of the latent heat flux, a linear relationship between the index and the latent heat flux was established. This relationship was used to map the latent heat flux on a pixel-by-pixel basis, resulting in <λE> = 194 W m-2. The third method makes use of a supervised classification of the thematic mapper image into eight land use classes. An average latent heat flux was assigned to each class by using field measurements of the latent heat flux. According to the percentage of occurrence of each class in the image, <λE> was calculated as 110 W m-2. A weighting scheme was produced to make an estimation of <λE> possible from in situ observations. The weighting scheme contained a multiplication factor for each measurement site in order to compensate for the relative contribution of that site to <λE>. It was shown that <λE> derived as the arithmetic mean of 13 individual in situ observations leads to a difference of 34% (<λE> = 104 W m-2), which emphasizes the need for improved weighting procedures.
NASA Astrophysics Data System (ADS)
Elmore, A. J.; Guinn, S. M.
2009-12-01
Land surface phenology (LSP) is the seasonal pattern of vegetation dynamics that occur each spring and fall. Multiple drivers of spatial variation in LSP and its variation over time have been analyzed using satellite remote sensing. Until recently, these observations have been restricted to moderate- and low-resolution data, as it is only at these spatial resolutions for which temporally continuous data is available. However, understanding small scale variation in LSP over space and time may be key to linking pattern to process, and in particular, could be used to understand how ecological processes at the stand level scale to landscapes and continents. Through utilization of the large, and now free, Landsat record, recent research has led to the development of robust methods for calculating average phenological patterns at 30-m resolution by stacking two decades worth of data by acquisition day of year (DOY). Here we have extended these techniques to calculate the deviation from the average LSP for any given acquisition DOY-year combination. We model the average LSP as two sigmoid functions, one increasing in spring and a second decreasing in fall, connected by a sloped line representing gradual summer leaf area changes (see Figure). Deviation from the average LSP is considered here to take two forms: (1) residual vegetation cover in mid- to late-summer represent locations in which disturbance, drought, or (alternatively) better than average growing conditions have resulted a separation (either negative or positive) from the average vegetation cover for that DOY, and (2) climate conditions that result in an earlier or later onset of greenness, exhibited as a separation from the average spring onset of greenness curve in the DOY direction (either early or late.) Our study system for this work is the deciduous forests of the mid-Atlantic, USA, where we show that late summer vegetation cover is tied to edaphic properties governing the site specific soil moisture balance. Additionally, we show that climatic factors (mostly related to topography) strongly influence the average start of spring. Annual deviations in the start of spring do not always scale linearly suggesting a spatially complex relationship between climate and the onset of spring. Model fit for a single pixel of mid-Atlantic deciduous forest. Shades of gray represent the weight each datum has on the model fit (increasing, white to black). Data weights account for variable atmospheric conditions between acquisitions.
Tsodikov, Oleg V; Record, M Thomas; Sergeev, Yuri V
2002-04-30
New computer programs, SurfRace and FastSurf, perform fast calculations of the solvent accessible and molecular (solvent excluded) surface areas of macromolecules. Program SurfRace also calculates the areas of cavities inaccessible from the outside. We introduce the definition of average curvature of molecular surface and calculate average molecular surface curvatures for each atom in a structure. All surface area and curvature calculations are analytic and therefore yield exact values of these quantities. High calculation speed of this software is achieved primarily by avoiding computationally expensive mathematical procedures wherever possible and by efficient handling of surface data structures. The programs are written initially in the language C for PCs running Windows 2000/98/NT, but their code is portable to other platforms with only minor changes in input-output procedures. The algorithm is robust and does not ignore either multiplicity or degeneracy of atomic overlaps. Fast, memory-efficient and robust execution make this software attractive for applications both in computationally expensive energy minimization algorithms, such as docking or molecular dynamics simulations, and in stand-alone surface area and curvature calculations. PMID:11939594
Gorsevski, Pece V; Donevska, Katerina R; Mitrovski, Cvetko D; Frizado, Joseph P
2012-02-01
This paper presents a GIS-based multi-criteria decision analysis approach for evaluating the suitability for landfill site selection in the Polog Region, Macedonia. The multi-criteria decision framework considers environmental and economic factors which are standardized by fuzzy membership functions and combined by integration of analytical hierarchy process (AHP) and ordered weighted average (OWA) techniques. The AHP is used for the elicitation of attribute weights while the OWA operator function is used to generate a wide range of decision alternatives for addressing uncertainty associated with interaction between multiple criteria. The usefulness of the approach is illustrated by different OWA scenarios that report landfill suitability on a scale between 0 and 1. The OWA scenarios are intended to quantify the level of risk taking (i.e., optimistic, pessimistic, and neutral) and to facilitate a better understanding of patterns that emerge from decision alternatives involved in the decision making process. PMID:22030279
Residence in coal-mining areas and low-birth-weight outcomes.
Ahern, Melissa; Mullett, Martha; Mackay, Katherine; Hamilton, Candice
2011-10-01
The objective of this study was to estimate the association between residence in coal mining environments and low birth weight. We conducted a cross-sectional, retrospective analysis of the association between low birth weight and mother's residence in coal mining areas in West Virginia. Birth data were obtained from the West Virginia Birthscore Dataset, 2005-2007 (n = 42,770). Data on coal mining were from the US Department of Energy. Covariates regarding mothers' demographics, behaviors, and insurance coverage were included. We used nested logistic regression (SUDAAN Proc Multilog) to conduct the study. Mothers who were older, unmarried, less educated, smoked, did not receive prenatal care, were on Medicaid, and had recorded medical risks had a greater risk of low birth weight. After controlling for covariates, residence in coal mining areas of West Virginia posed an independent risk of low birth weight. Odds ratios for both unadjusted and adjusted findings suggest a dose-response effect. Adjusted findings show that living in areas with high levels of coal mining elevates the odds of a low-birth-weight infant by 16%, and by 14% in areas with lower mining levels, relative to counties with no coal mining. After covariate adjustment, the persistence of a mining effect on low-birth-weight outcomes suggests an environmental effect resulting from pollution from mining activities. Air and water quality assessments have been largely missing from mining communities, but the need for them is indicated by these findings. PMID:20091110
Mean skin temperature weighted by skin area, heat transfer coefficients and thermal sensitivity
NASA Astrophysics Data System (ADS)
Mochida, T.
1983-07-01
Formulas for calculating the mean skin temperature are described by a general form - the sum total of the product of both the regional skin temperature and the weighting factor concerned with the region. The weighting factor in these formulas was classified into five groups from the point of the content and the concrete values were compared. Based on the heat equilibrium between man and his environment, a mean skin temperature formula weighted by skin areas and the heat transfer coefficients was derived. With reference to the thermal sensitivity coefficients given, a new formula, which is weighted by three important factors - the skin area, the heat transfer coefficients and the thermal sensitivity, was proposed. As the result of a comparison run against formulas reported previously, the weighting factors of the skin area heat transfer coefficient formula are similar to those of the Hardy - DuBois formula, and the weighting factors of the formula by the skin area, the heat transfer coefficients and the thermal sensitivity are similar to those of the formula by Nadel et al.
NASA Technical Reports Server (NTRS)
Schols, J. L.; Eloranta, E. W.
1992-01-01
Area-averaged horizontal wind measurements are derived from the motion of spatial inhomogeneities in aerosol backscattering observed with a volume-imaging lidar. Spatial averaging provides high precision, reducing sample variations of wind measurements well below the level of turbulent fluctuations, even under conditions of very light mean winds and strong convection or under the difficult conditions represented by roll convection. Wind velocities are measured using the two-dimensional spatial cross correlation computed between successive horizontal plane maps of aerosol backscattering, assembled from three-dimensional lidar scans. Prior to calculation of the correlation function, three crucial steps are performed: (1) the scans are corrected for image distortion by the wind during a finite scan time; (2) a temporal high pass median filtering is applied to eliminate structures that do not move with the wind; and (3) a histogram equalization is employed to reduce biases to the brightest features.
Boosting with Averaged Weight Vectors
NASA Technical Reports Server (NTRS)
Oza, Nikunj C.; Clancy, Daniel (Technical Monitor)
2002-01-01
AdaBoost is a well-known ensemble learning algorithm that constructs its constituent or base models in sequence. A key step in AdaBoost is constructing a distribution over the training examples to create each base model. This distribution, represented as a vector, is constructed to be orthogonal to the vector of mistakes made by the previous base model in the sequence. The idea is to make the next base model's errors uncorrelated with those of the previous model. Some researchers have pointed out the intuition that it is probably better to construct a distribution that is orthogonal to the mistake vectors of all the previous base models, but that this is not always possible. We present an algorithm that attempts to come as close as possible to this goal in an efficient manner. We present experimental results demonstrating significant improvement over AdaBoost and the Totally Corrective boosting algorithm, which also attempts to satisfy this goal.
Larsen, Inge; Hjulsager, Charlotte Kristiane; Holm, Anders; Olsen, John Elmerdahl; Nielsen, Sren Saxmose; Nielsen, Jens Peter
2016-01-01
Oral treatment with antimicrobials is widely used in pig production for the control of gastrointestinal infections. Lawsonia intracellularis (LI) causes enteritis in pigs older than six weeks of age and is commonly treated with antimicrobials. The objective of this study was to evaluate the efficacy of three oral dosage regimens (5, 10 and 20mg/kg body weight) of oxytetracycline (OTC) in drinking water over a five-day period on diarrhoea, faecal shedding of LI and average daily weight gain (ADG). A randomised clinical trial was carried out in four Danish pig herds. In total, 539 animals from 37 batches of nursery pigs were included in the study. The dosage regimens were randomly allocated to each batch and initiated at presence of assumed LI-related diarrhoea. In general, all OTC doses used for the treatment of LI infection resulted in reduced diarrhoea and LI shedding after treatment. Treatment with a low dose of 5mg/kg OTC per kg body weight, however, tended to cause more watery faeces and resulted in higher odds of pigs shedding LI above detection level when compared to medium and high doses (with odds ratios of 5.5 and 8.4, respectively). No association was found between the dose of OTC and the ADG. In conclusion, a dose of 5mg OTC per kg body weight was adequate for reducing the high-level LI shedding associated with enteropathy, but a dose of 10mg OTC per kg body weight was necessary to obtain a maximum reduction in LI shedding. PMID:26718056
NASA Astrophysics Data System (ADS)
Ghiglieri, Giorgio; Carletti, Alberto; Pittalis, Daniele
2014-11-01
Runoff estimation and water budget in ungauged basins is a challenge for hydrological researchers and planners. The principal aim of this study was the application and validation of the Kennessey method, which is a physiography-based indirect process for determining the average annual runoff coefficient and the basin-scale water balance. The coefficient can be calculated using specific physiographic characteristics (slope, permeability and vegetation cover) and a parameter that defines climatic conditions and does not require instrumental data. One of the main purposes of this study was to compare the average annual runoff coefficient obtained using the Kennessey method with the coefficients calculated using data from 30 instrumented drainage basins in Sardinia (Italy) over 71 years (from 1922 to 1992). These measurements represent an important and complete historical dataset from the study area. Using the runoff coefficient map, the method was also applied to assess the effective annual recharge rate of the aquifers of the Calich hydrogeological basin in the Nurra Plain (Alghero, NW Sardinia-Italy). The groundwater recharge rate was compared with rates calculated using the standard water balance method. The implementation of the method at the regional and basin scales was supported by GIS analyses. The results of the method are promising but show some discrepancies with other methodologies due to the higher weights given to the physiographic parameters than to the meteorological parameters. However, even though the weights assigned to the parameters require improvements, the Kennessey method is a useful tool for evaluating hydrologic processes, particularly for water management in areas where instrumental data are not available.
Inducing Conservation of Number, Weight, Volume, Area, and Mass in Pre-School Children.
ERIC Educational Resources Information Center
Young, Beverly S.
The major question this study attempted to answer was, "Can conservation of number, area, weight, mass, and volume to be induced and retained by 3- and 4-year-old children by structured instruction with a multivariate approach? Three nursery schools in Iowa City supplied subjects for this study. The Institute of Child Behavior and Development…
Rieder, Anne; Knutsen, Svein Halvor; Ulset, Ann-Sissel T; Christensen, Bjrn E; Andersson, Roger; Mikkelson, Atte; Tuomainen, Pivi; Maina, Ndegwa; Ballance, Simon
2015-06-25
Even though size exclusion chromatography (SEC) with post column addition of calcofluor (SEC-calcofluor) has been used for the determination of cereal ?-glucan molar mass in foods for many years, there is a lack of systematic evaluation of the method. To address this issue a set of suitable ?-glucan standards were generated by preparative SEC and their molar mass characteristics were determined by analytical multi-detection SEC (refractive index (RI), light scattering). Each standard was then analysed by SEC-calcofluor at three different labs. As a direct comparison, the analyses were repeated with a RI detector. For SEC-calcofluor accurate measurements of weight average molar mass (Mw) can be made for ?-glucan populations within 10-50010(3)g/mol. Above this molar mass threshold there is an increasing tendency for underestimation of Mw. Precipitation of some ?-glucan-calcofluor complexes may have delayed their transport into the detector. PMID:25839819
NASA Astrophysics Data System (ADS)
Fonseca, Julio; Del-Castillo-Negrete, Diego; Caldas, Ibere
2014-10-01
Area preserving maps have been extensively used to model 2-dimensional chaotic transport in plasmas and fluids. Here we focus on three types of area preserving maps describing ExB chaotic transport in magnetized plasmas with zonal flows perturbed by electrostatic drift waves. We include finite Larmor radius (FLR) effects by gyro-averaging the corresponding Hamiltonians of the maps. The Hamiltonians have frequencies with monotonic and non-monotonic profiles. In the limit of zero Larmor radius, the monotonic frequency map reduces to the standard Chirikov-Taylor map, and, in the case of non-monotonic frequency, the map reduces to the standard nontwist map. We show that FLR leads to chaos suppression, modifies the stability of fixed points, and changes the robustness of transport barriers. FLR effects also modify the phase space topology and give rise to bifurcations of the zonal flow ExB velocity profile. Dynamical systems methods based on recurrence time statistics are used to quantify the dependence on the Larmor radius of the threshold for the destruction of transport barriers.
The weight of spectacle frames and the area of their nose pads.
Walsh, Glyn
2010-07-01
The current international standard for spectacle frames recommends that frames weighing up to 25 g should have a minimum nose-pad contact area of 200 mm(2) and that those weighing over 25 g should have a minimum contact area of 250 mm(2). It is shown that these recommendations are being almost universally ignored for frames with separate pads on arms. The information on frame materials in manufacturers' literature is woefully inadequate but, from the information available, there was little difference in weight between stainless steel frames and those from unnamed materials. PMID:20629962
NASA Astrophysics Data System (ADS)
Naik, Haladhara; Kim, Guinyun; Kim, Kwangsoo; Zaman, Muhammad; Goswami, Ashok; Lee, Man Woo; Yang, Sung-Chul; Lee, Young-Ouk; Shin, Sung-Gyun; Cho, Moo-Hyun
2016-04-01
Photo-neutron cross sections of 197Au were experimentally determined for the bremsstrahlung end-point energies of 50, 60, and 70 MeV, by utilizing activation and off-line γ-ray spectrometric technique, using the 100 MeV electron linac at the Pohang Accelerator Laboratory (PAL), Pohang, Korea. The 197Au(γ, xn; x = 1- 6) reaction cross sections were calculated as a function of the bombarding photon energy by using the TALYS 1.6 computer code with default parameters. The flux-weighted average cross sections were obtained from the literature data and the theoretical values of TALYS 1.6 and TENDL-2014, for mono-energetic photons, and are found to be in good agreement with the present data. Isomeric yield ratios of 196m2,gAu from the 197Au(γ, n) reaction were also determined for the bremsstrahlung end-point energies of 50, 60, and 70 MeV, from the reaction cross sections of m2- and g-states, based on the present experimental data, and are found to be in good agreement with the theoretical values based on TALYS 1.6 and TENDL-2014.
Mazzella, Nicolas; Debenest, Timothée; Delmas, François
2008-09-01
Polar organic chemical integrative samplers (POCIS) were exposed for 9 days in two different microcosms that contained river waters spiked with deethylterbuthylazine, terbuthylazine and isoproturon. The experiment was performed with natural light and strong turbulence (flow velocities of about 15-50cms(-1)) for reproducing natural conditions. The concentrations were kept relatively constant in the first microcosm (2.6-3.6microgl(-1)) and were variable in the second microcosm (peak concentrations ranged from 15 to 24microgl(-1) during the 3 day pulse phase). The time-weighted average (TWA) concentrations were determined with both POCIS and repetitive grab sampling followed by solid-phase extraction. The results showed a systematic and significant overestimation of the TWA concentrations with the POCIS most probably due to the use of sampling rates derived under low flow scenario. The results showed also that peak concentrations of pollutants are fully integrated by this passive sampler. Even if the POCIS should not provide very accurate concentration estimates without the application of adequate sampling rate values or the use of performance reference compounds, it can be a really useful tool for detecting episodic or short-term pollution events (e.g. increased herbicide concentrations during a flood), which may be missed with classical and low frequency grab sampling. PMID:18649919
NASA Astrophysics Data System (ADS)
Yu, C.; Zinniker, D. A.; Moldowan, J.
2010-12-01
Urban air pollution is an ongoing and complicated problem for both residents and policy makers. This study aims to provide a better understanding of the geographic source and fate of organic pollutants in a dynamic urban environment. Natural and artificial hydrophobic substrates were employed for the passive monitoring and mapping of ground-level organic pollutants in the San Francisco Bay area. We focused specifically on volatile and semi-volatile polycyclic aromatic hydrocarbons (PAHs). These compounds are proxies for a broad range of combustion related air pollutants derived from local, regional, and global combustion sources. PAHs include several well-studied carcinogens and can be measured easily and accurately across a broad range of concentrations. Estimates of time-integrated vapor phase and particle deposition were made from measuring accumulated PAHs in the leaves of several widely distributed tree species (including the Quercus agrifolia and Sequoia sempervirens) and an artificial wax film. Samples were designed to represent pollutant exposure over a period of one to several months. The selective sampling and analysis of hydrophobic substrates providess insight into the average geographic distribution of ground-level air pollutants in a simple and inexpensive way. However, accumulated organics do not directly correlated with human exposure and the source signature of PAHs may be obscured by transport, deposition, and flux processes. We attempted to address some of these complications by studying 1) PAH accumulation rates within substrates in a controlled microcosm, 2) differences in PAH abundance in different substrate types at the same locality, and 3) samples near long-term high volume air sampling stations. We also set out to create a map of PAH concentrations based on our measurements. This map can be directly compared with interpolated data from high-volume sampling stations and used to address questions concerning atmospheric heterogeneity of these pollutants (i.e. due to both source localization and dominant wind patterns). Our initial results indicate that exposure to PAHs in the bay area is geographically heterogeneous and individual exposure may vary by more than two orders of magnitude. The signatures of PAH contamination also varies considerably, indicating different sources and differing transportation mechanisms may be important at different sites and times.
Shmool, Jessie L C; Bobb, Jennifer F; Ito, Kazuhiko; Elston, Beth; Savitz, David A; Ross, Zev; Matte, Thomas D; Johnson, Sarah; Dominici, Francesca; Clougherty, Jane E
2015-10-01
Numerous studies have linked air pollution with adverse birth outcomes, but relatively few have examined differential associations across the socioeconomic gradient. To evaluate interaction effects of gestational nitrogen dioxide (NO2) and area-level socioeconomic deprivation on fetal growth, we used: (1) highly spatially-resolved air pollution data from the New York City Community Air Survey (NYCCAS); and (2) spatially-stratified principle component analysis of census variables previously associated with birth outcomes to define area-level deprivation. New York City (NYC) hospital birth records for years 2008-2010 were restricted to full-term, singleton births to non-smoking mothers (n=243,853). We used generalized additive mixed models to examine the potentially non-linear interaction of nitrogen dioxide (NO2) and deprivation categories on birth weight (and estimated linear associations, for comparison), adjusting for individual-level socio-demographic characteristics and sensitivity testing adjustment for co-pollutant exposures. Estimated NO2 exposures were highest, and most varying, among mothers residing in the most-affluent census tracts, and lowest among mothers residing in mid-range deprivation tracts. In non-linear models, we found an inverse association between NO2 and birth weight in the least-deprived and most-deprived areas (p-values<0.001 and 0.05, respectively) but no association in the mid-range of deprivation (p=0.8). Likewise, in linear models, a 10 ppb increase in NO2 was associated with a decrease in birth weight among mothers in the least-deprived and most-deprived areas of -16.2g (95% CI: -21.9 to -10.5) and -11.0 g (95% CI: -22.8 to 0.9), respectively, and a non-significant change in the mid-range areas [?=0.5 g (95% CI: -7.7 to 8.7)]. Linear slopes in the most- and least-deprived quartiles differed from the mid-range (reference group) (p-values<0.001 and 0.09, respectively). The complex patterning in air pollution exposure and deprivation in NYC, however, precludes simple interpretation of interactive effects on birth weight, and highlights the importance of considering differential distributions of air pollution concentrations, and potential differences in susceptibility, across deprivation levels. PMID:26318257
Baimatova, Nassiba; Koziel, Jacek A; Kenessov, Bulat
2015-05-11
A new and simple method for benzene, toluene, ethylbenzene and o-xylene (BTEX) quantification in vehicle exhaust was developed based on diffusion-controlled extraction onto a retracted solid-phase microextraction (SPME) fiber coating. The rationale was to develop a method based on existing and proven SPME technology that is feasible for field adaptation in developing countries. Passive sampling with SPME fiber retracted into the needle extracted nearly two orders of magnitude less mass (n) compared with exposed fiber (outside of needle) and sampling was in a time weighted-averaging (TWA) mode. Both the sampling time (t) and fiber retraction depth (Z) were adjusted to quantify a wider range of Cgas. Extraction and quantification is conducted in a non-equilibrium mode. Effects of Cgas, t, Z and T were tested. In addition, contribution of n extracted by metallic surfaces of needle assembly without SPME coating was studied. Effects of sample storage time on n loss was studied. Retracted TWA-SPME extractions followed the theoretical model. Extracted n of BTEX was proportional to Cgas, t, Dg, T and inversely proportional to Z. Method detection limits were 1.8, 2.7, 2.1 and 5.2 mg m(-3) (0.51, 0.83, 0.66 and 1.62 ppm) for BTEX, respectively. The contribution of extraction onto metallic surfaces was reproducible and influenced by Cgas and t and less so by T and by the Z. The new method was applied to measure BTEX in the exhaust gas of a Ford Crown Victoria 1995 and compared with a whole gas and direct injection method. PMID:25911428
MPWide: a light-weight library for efficient message passing over wide area networks
NASA Astrophysics Data System (ADS)
Groen, D.; Rieder, S.; Portegies Zwart, S.
2013-12-01
We present MPWide, a light weight communication library which allows efficient message passing over a distributed network. MPWide has been designed to connect application running on distributed (super)computing resources, and to maximize the communication performance on wide area networks for those without administrative privileges. It can be used to provide message-passing between application, move files, and make very fast connections in client-server environments. MPWide has already been applied to enable distributed cosmological simulations across up to four supercomputers on two continents, and to couple two different bloodflow simulations to form a multiscale simulation.
A comparison of spatial smoothing methods for small area estimation with sampling weights
Mercer, Laina; Wakefield, Jon; Chen, Cici; Lumley, Thomas
2014-01-01
Small area estimation (SAE) is an important endeavor in many fields and is used for resource allocation by both public health and government organizations. Often, complex surveys are carried out within areas, in which case it is common for the data to consist only of the response of interest and an associated sampling weight, reflecting the design. While it is appealing to use spatial smoothing models, and many approaches have been suggested for this endeavor, it is rare for spatial models to incorporate the weighting scheme, leaving the analysis potentially subject to bias. To examine the properties of various approaches to estimation we carry out a simulation study, looking at bias due to both non-response and non-random sampling. We also carry out SAE of smoking prevalence in Washington State, at the zip code level, using data from the 2006 Behavioral Risk Factor Surveillance System. The computation times for the methods we compare are short, and all approaches are implemented in R using currently available packages. PMID:24959396
White, R R; Capper, J L
2013-12-01
The objective of this study was to assess environmental impact, economic viability, and social acceptability of 3 beef production systems with differing levels of efficiency. A deterministic model of U.S. beef production was used to predict the number of animals required to produce 1 × 10(9) kg HCW beef. Three production treatments were compared, 1 representing average U.S. production (control), 1 with a 15% increase in ADG, and 1 with a 15% increase in finishing weight (FW). For each treatment, various socioeconomic scenarios were compared to account for uncertainty in producer and consumer behavior. Environmental impact metrics included feed consumption, land use, water use, greenhouse gas emissions (GHGe), and N and P excretion. Feed cost, animal purchase cost, animal sales revenue, and income over costs (IOVC) were used as metrics of economic viability. Willingness to pay (WTP) was used to identify improvements or reductions in social acceptability. When ADG improved, feedstuff consumption, land use, and water use decreased by 6.4%, 3.2%, and 12.3%, respectively, compared with the control. Carbon footprint decreased 11.7% and N and P excretion were reduced by 4% and 13.8%, respectively. When FW improved, decreases were seen in feedstuff consumption (12.1%), water use (9.2%). and land use (15.5%); total GHGe decreased 14.7%; and N and P excretion decreased by 10.1% and 17.2%, compared with the control. Changes in IOVC were dependent on socioeconomic scenario. When the ADG scenario was compared with the control, changes in sector profitability ranged from 51 to 117% (cow-calf), -38 to 157% (stocker), and 37 to 134% (feedlot). When improved FW was compared, changes in cow-calf profit ranged from 67% to 143%, stocker profit ranged from -41% to 155% and feedlot profit ranged from 37% to 136%. When WTP was based on marketing beef being more efficiently produced, WTP improved by 10%; thus, social acceptability increased. When marketing was based on production efficiency and consumer knowledge of growth-enhancing technology use, WTP decreased by 12%-leading to a decrease in social acceptability. Results demonstrated that improved efficiency also improved environmental impact, but impacts on economic viability and social acceptability are highly dependent on consumer and producer behavioral responses to efficiency improvements. PMID:24146151
Krpálková, L; Cabrera, V E; Kvapilík, J; Burdych, J; Crump, P
2014-10-01
The objective of this study was to evaluate the associations of variable intensity in rearing dairy heifers on 33 commercial dairy herds, including 23,008 cows and 18,139 heifers, with age at first calving (AFC), average daily weight gain (ADG), and milk yield (MY) level on reproduction traits and profitability. Milk yield during the production period was analyzed relative to reproduction and economic parameters. Data were collected during a 1-yr period (2011). The farms were located in 12 regions in the Czech Republic. The results show that those herds with more intensive rearing periods had lower conception rates among heifers at first and overall services. The differences in those conception rates between the group with the greatest ADG (≥0.800 kg/d) and the group with the least ADG (≤0.699 kg/d) were approximately 10 percentage points in favor of the least ADG. All the evaluated reproduction traits differed between AFC groups. Conception at first and overall services (cows) was greatest in herds with AFC ≥800 d. The shortest days open (105 d) and calving interval (396 d) were found in the middle AFC group (799 to 750 d). The highest number of completed lactations (2.67) was observed in the group with latest AFC (≥800 d). The earliest AFC group (≤749 d) was characterized by the highest depreciation costs per cow at 8,275 Czech crowns (US$414), and the highest culling rate for cows of 41%. The most profitable rearing approach was reflected in the middle AFC (799 to 750 d) and middle ADG (0.799 to 0.700 kg) groups. The highest MY (≥8,500 kg) occurred with the earliest AFC of 780 d. Higher MY led to lower conception rates in cows, but the highest MY group also had the shortest days open (106 d) and a calving interval of 386 d. The same MY group had the highest cow depreciation costs, net profit, and profitability without subsidies of 2.67%. We conclude that achieving low AFC will not always be the most profitable approach, which will depend upon farm-specific herd management. The MY is a very important factor for dairy farm profitability. The group of farms having the highest MY achieved the highest net profit despite having greater fertility problems. PMID:25064657
Relationship of T2-Weighted MRI Myocardial Hyperintensity and the Ischemic Area-At-Risk
Kim, Han W.; Van Assche, Lowie; Jennings, Robert B.; Wince, W. Benjamin; Jensen, Christoph J.; Rehwald, Wolfgang G.; Wendell, David C.; Bhatti, Lubna; Spatz, Deneen M.; Parker, Michele A.; Jenista, Elizabeth R.; Klem, Igor; Crowley, Anna Lisa C.; Chen, Enn-Ling; Judd, Robert M.
2015-01-01
Rationale: After acute myocardial infarction (MI), delineating the area-at-risk (AAR) is crucial for measuring how much, if any, ischemic myocardium has been salvaged. T2-weighted MRI is promoted as an excellent method to delineate the AAR. However, the evidence supporting the validity of this method to measure the AAR is indirect, and it has never been validated with direct anatomic measurements. Objective: To determine whether T2-weighted MRI delineates the AAR. Methods and Results: Twenty-one canines and 24 patients with acute MI were studied. We compared bright-blood and black-blood T2-weighted MRI with images of the AAR and MI by histopathology in canines and with MI by in vivo delayed-enhancement MRI in canines and patients. Abnormal regions on MRI and pathology were compared by (a) quantitative measurement of the transmural-extent of the abnormality and (b) picture matching of contours. We found no relationship between the transmural-extent of T2-hyperintense regions and that of the AAR (bright-blood-T2: r=0.06, P=0.69; black-blood-T2: r=0.01, P=0.97). Instead, there was a strong correlation with that of infarction (bright-blood-T2: r=0.94, P<0.0001; black-blood-T2: r=0.95, P<0.0001). Additionally, contour analysis demonstrated a fingerprint match of T2-hyperintense regions with the intricate contour of infarcted regions by delayed-enhancement MRI. Similarly, in patients there was a close correspondence between contours of T2-hyperintense and infarcted regions, and the transmural-extent of these regions were highly correlated (bright-blood-T2: r=0.82, P<0.0001; black-blood-T2: r=0.83, P<0.0001). Conclusion: T2-weighted MRI does not depict the AAR. Accordingly, T2-weighted MRI should not be used to measure myocardial salvage, either to inform patient management decisions or to evaluate novel therapies for acute MI. PMID:25972514
NASA Astrophysics Data System (ADS)
Hu, X.; Waller, L.; Liu, Y.
2010-12-01
Using remote sensing data to study the characteristics of PM2.5 (particles smaller than 2.5µm in size) especially in areas not covered by ground monitoring networks has attracted much interest due to multiple health outcomes related to its exposure. To accurately predict PM2.5 exposure, successfully modeling the relationship between PM2.5 concentration and aerosol optical thickness (AOT), as well as other environmental parameters, is crucial. Most of currently reported models are global methods without considering local variations, which might introduce significant errors into prediction results. In this paper, a geographically weighted regression model (GWR) was developed to model the relationship among PM2.5, AOT, and meteorological parameters such as mixing height, surface air temperature, relative humidity, and surface wind speed. GWR is capable of estimating local parameters instead of global parameters in terms of the geographical location, and all coefficients vary geographically to indicate the spatial variation. The study area is centered around Atlanta Metro area, and the data from 2001 to 2007 was collected from various sources. After developing the model, cross-validation techniques were implemented to assess the accuracy of our model. The results indicated that GWR, due to its ability of explaining local variations, has the potential to generate a better fit and can provide a promising alternative in PM2.5 exposure estimation.
Motz, R; Schumacher, M; Nrnberg, J; Viemann, M; Grafmller, S; Fiedler, K; Claus, M; Kronberg, K
2014-12-01
Looking after children means caring for very small infants up to adult-sized adolescents, with weights ranging from 500 g to more than 100 kg and heights ranging from 25 to more than 200 cm. The available echocardiographic reference data were drawn from a small sample, which did not include preterm infants. Most authors have used body weight or body surface area to predict left ventricular dimensions. The current authors had the impression that body length would be a better surrogate parameter than body weight or body surface area. They analyzed their echocardiographic database retrospectively. The analysis included all available echocardiographic data from 6 June 2001 to 15 December 2011 from their echocardiographic database. The authors included 12,086 of 26,325 subjects documented as patients with normal hearts in their analysis by the examining the pediatric cardiologist. For their analysis, they selected body weight, length, age, and aortic and pulmonary valve diameter in two-dimensional echocardiography and left ventricular dimension in M-mode. They found good correlation between echocardiographic dimensions and body surface area, body weight, and body length. The analysis showed a complex relationship between echocardiographic measurements and body weight and body surface area, whereas body length showed a linear relationship. This makes prediction of echo parameters more reliable. According to this retrospective analysis, body length is a better parameter for evaluating echocardiographic measurements than body weight or body surface area and should therefore be used in daily practice. PMID:24894894
Code of Federal Regulations, 2011 CFR
2011-07-01
... Extraction for Vegetable Oil Production Compliance Requirements § 63.2854 How do I determine the weighted... received for use in your vegetable oil production process. By the end of each calendar month following an... the solvent in each delivery of solvent, including solvent recovered from off-site oil. To...
Code of Federal Regulations, 2010 CFR
2010-07-01
... Extraction for Vegetable Oil Production Compliance Requirements § 63.2854 How do I determine the weighted... received for use in your vegetable oil production process. By the end of each calendar month following an... the solvent in each delivery of solvent, including solvent recovered from off-site oil. To...
Roberts, Graham J; McDonald, Fraser; Neil, Monica; Lucas, Victoria S
2014-08-01
The mathematical principle of weighting averages to determine the most appropriate numerical outcome is well established in economic and social studies. It has seen little application in forensic dentistry. This study re-evaluated the data from a previous study of age assessment at the 10 year threshold. A semiautomatic process of weighting averages by n-td, x-tds, sd-tds, se-tds, 1/sd-tds, 1/se-tds was prepared in an Excel worksheet and the different weighted mean values reported. In addition the Fixed Effects and Random Effects models for Meta-Analysis were used and applied to the same data sets. In conclusion it has been shown that the most accurate age estimation method is to use the Random Effects Model for the mathematical procedures. PMID:25066175
NASA Astrophysics Data System (ADS)
Scherf, A.; Roth, R.
1996-12-01
During the field campaign of EFEDA II several aircraft measurements were performed in order to evaluate area mean values of turbulent energy fluxes over a relatively flat terrain in a desertification threatened area in Spain. Since earlier field experiments indicated differences between airborne measurements and surface observations, we tried to close the gap by carefully analysing the turbulence measurements. In order to evaluate the influence of the temporal variation of the convective boundary layer, the rise of the inversion, derived from simultaneously performed radiosonde ascents, was taken into account. By estimating the linear approximated fields of the meteorological parameters, it was possible to calculate the mean values of these quantities as well as the temporal and spatial derivatives, which are necessary for the evaluation of the advective terms of the energy budget. In this way is possible to examine the terms of the conservation equations in a supplementary way.
NASA Technical Reports Server (NTRS)
Huff, Edward M.; Mosher, Marianne; Barszcz, Eric
2002-01-01
Recent research using NASA Ames AH-1 and OH-58C helicopters, and NASA Glenn test rigs, has shown that in-flight vibration data are typically non-stationary [l-4]. The nature and extent of this non-stationarity is most likely produced by several factors operating simultaneously. The aerodynamic flight environment and pilot commands provide continuously changing inputs, with a complex dynamic response that includes automatic feedback control from the engine regulator. It would appear that the combined effects operate primarily through an induced torque profile, which causes concomitant stress modulation at the individual internal gear meshes in the transmission. This notion is supported by several analyses, which show that upwards of 93% of the vibration signal s variance can be explained by knowledge of torque alone. That this relationship is stronger in an AH-1 than an OH-58, where measured non-stationarity is greater, suggests that the overall mass of the vehicle is an important consideration. In the lighter aircraft, the unsteady aerodynamic influences transmit relatively greater unsteady dynamic forces on the mechanical components, quite possibly contributing to its greater non-stationarity . In a recent paper using OH-58C pinion data [5], the authors have shown that in computing a time synchronous average (TSA) for various single-value metric computations, an effective trade-off can be obtained between sample size and measured stationarity by using data from only a single mesh cycle. A mesh cycle, which is defined as the number of rotations required for the gear teeth to return to their original mating position, has the property of representing all of the discrete phase angles of the opposing gears exactly once in the average. Measured stationarity is probably maximized because a single mesh cycle of the pinion gear occurs over a very short span of time, during which time-dependent non-stationary effects are kept to a minimum. Clearly, the advantage of local stationarity diminishes as the temporal duration of the cycle increases. This is most evident for a planetary mesh cycle, which can take several minutes to complete.
Ruiz, J M; Busnel, J P; Benoît, J P
1990-09-01
The phase separation of fractionated poly(DL-lactic acid-co-glycolic acid) copolymers 50/50 was determined by silicone oil addition. Polymer fractionation by preparative size exclusion chromatography afforded five different microsphere batches. Average molecular weight determined the existence, width, and displacement of the "stability window" inside the phase diagrams, and also microsphere characteristics such as core loading and amount released over 6 hr. Further, the gyration and hydrodynamic radii were measured by light scattering. It is concluded that the polymer-solvent affinity is largely modified by the variation of average molecular weights owing to different levels of solubility. The lower the average molecular weight is, the better methylene chloride serves as a solvent for the coating material. However, a paradoxical effect due to an increase in free carboxyl and hydroxyl groups is noticed for polymers of 18,130 and 31,030 SEC (size exclusion chromatography) Mw. For microencapsulation, polymers having an intermediate molecular weight (47,250) were the most appropriate in terms of core loading and release purposes. PMID:2235892
NASA Astrophysics Data System (ADS)
Hernndez, Leonor; Juli, J. Enrique; Paranjape, Sidharth; Hibiki, Takashi; Ishii, Mamoru
2010-11-01
In this work, the use of the area-averaged void fraction and bubble chord length entropies is introduced as flow regime indicators in two-phase flow systems. The entropy provides quantitative information about the disorder in the area-averaged void fraction or bubble chord length distributions. The CPDF (cumulative probability distribution function) of void fractions and bubble chord lengths obtained by means of impedance meters and conductivity probes are used to calculate both entropies. Entropy values for 242 flow conditions in upward two-phase flows in 25.4 and 50.8-mm pipes have been calculated. The measured conditions cover ranges from 0.13 to 5 m/s in the superficial liquid velocity j f and ranges from 0.01 to 25 m/s in the superficial gas velocity j g. The physical meaning of both entropies has been interpreted using the visual flow regime map information. The area-averaged void fraction and bubble chord length entropies capability as flow regime indicators have been checked with other statistical parameters and also with different input signals durations. The area-averaged void fraction and the bubble chord length entropies provide better or at least similar results than those obtained with other indicators that include more than one parameter. The entropy is capable to reduce the relevant information of the flow regimes in only one significant and useful parameter. In addition, the entropy computation time is shorter than the majority of the other indicators. The use of one parameter as input also represents faster predictions.
Ito, Tadashi; Sakai, Yoshihito; Nakamura, Eishi; Yamazaki, Kazunori; Yamada, Ayaka; Sato, Noritaka; Morita, Yoshifumi
2015-01-01
[Purpose] The purpose of this study was to examine the relationship between the paraspinal muscle cross-sectional area and the relative proprioceptive weighting ratio during local vibratory stimulation of older persons with lumbar spondylosis in an upright position. [Subjects] In all, 74 older persons hospitalized for lumbar spondylosis were included. [Methods] We measured the relative proprioceptive weighting ratio of postural sway using a Wii board while vibratory stimulations of 30, 60, or 240?Hz were applied to the subjects paraspinal or gastrocnemius muscles. Back strength, abdominal muscle strength, and erector spinae muscle (L1/L2, L4/L5) and lumbar multifidus (L1/L2, L4/L5) cross-sectional areas were evaluated. [Results] The erector spinae muscle (L1/L2) cross-sectional area was associated with the relative proprioceptive weighting ratio during 60Hz stimulation. [Conclusion] These findings show that the relative proprioceptive weighting ratio compared to the erector spinae muscle (L1/L2) cross-sectional area under 60Hz proprioceptive stimulation might be a good indicator of trunk proprioceptive sensitivity. PMID:26311962
Noguchi, Y; Miura, H; Takasugi, S; Iwamoto, Y
1999-01-01
To clarify the intra-articular pathology of dysplastic osteoarthritis, we conducted 120 hip arthroscopies. Hips were classified radiologically into four stages: 41 prearthritic stage, 42 early-stage, 34 advanced-stage, and 3 end-stage. As the radiological stage progressed, articular cartilage degeneration became more extensive and severe. In the prearthritic and early stages, cartilage degeneration generally originated in the anterosuperior portion of the weight-bearing area of the femoral head and acetabulum. Cartilage degeneration of the acetabulum usually preceded that of the femoral head. Most detaching tears of the acetabular labrum originated in the anterosuperior area, too. In conclusion, cartilage and labrum degeneration in the dysplastic hip generally originates in the anterosuperior portion of weight-bearing area. PMID:10424553
Tan, Xiang-Ling; Zhou, Yan; Cai, Ke-Fu; Shao, Yi-Xiang; Zhu, Shun-Xing; Lu, L; Williams, R W; Gu, Xiao-Song
2004-08-01
Two inbred strains of mice, A/J, C57BL/6J and F2 intercross progenies,were used for QTL mapping for weight and cross-sectional area on cervical enlargement of spinal cord in mice. 13 QTLs located on Chromosome 2, 4, 8, 14, 15, 17, 18, 19 and X, respectively, for these two traits were found. Six QTLs were responsible for the cord weight, four for the cross-sectional area and three for both. Among 13 QTLs, three QTLs (P < 0.01) termed SC1 (located near D15Mit158) ,SC2 (DXMit140) and SC3 (DXMit64) accounted for 24%, 19% and 15% of the total variance in weight phenotype, and -3.78, 3.41 and 2.06 mg additive effect, respectively. The P value of other QTLs is between 0.01 and 0.05. SC1 is only one QTL that responsible for both weight and cross-sectional area in three QTLs above. This study revealed the location of major QTLs related size of spinal cord in mice, and may be helpful in fine mapping and ultimate identification of candidate genes. PMID:15481534
2014-01-01
Mesoporous ZnO nanoparticles have been synthesized with tremendous increase in specific surface area of up to 578m2/g which was 5.54m2/g in previous reports (J. Phys. Chem. C 113:14676-14680, 2009). Different mesoporous ZnO nanoparticles with average pore sizes ranging from 7.22 to 13.43nm and specific surface area ranging from 50.41 to 578m2/g were prepared through the sol-gel method via a simple evaporation-induced self-assembly process. The hydrolysis rate of zinc acetate was varied using different concentrations of sodium hydroxide. Morphology, crystallinity, porosity, and J-V characteristics of the materials have been studied using transmission electron microscopy (TEM), X-ray diffraction (XRD), BET nitrogen adsorption/desorption, and Keithley instruments. PMID:25339855
NASA Astrophysics Data System (ADS)
Obata, Kenta; Huete, Alfredo R.
2014-01-01
This study investigated the mechanisms underlying the scaling effects that apply to a fraction of vegetation cover (FVC) estimates derived using two-band spectral vegetation index (VI) isoline-based linear mixture models (VI isoline-based LMM). The VIs included the normalized difference vegetation index, a soil-adjusted vegetation index, and a two-band enhanced vegetation index (EVI2). This study focused in part on the monotonicity of an area-averaged FVC estimate as a function of spatial resolution. The proof of monotonicity yielded measures of the intrinsic area-averaged FVC uncertainties due to scaling effects. The derived results demonstrate that a factor ξ, which was defined as a function of "true" and "estimated" endmember spectra of the vegetated and nonvegetated surfaces, was responsible for conveying monotonicity or nonmonotonicity. The monotonic FVC values displayed a uniform increasing or decreasing trend that was independent of the choice of the two-band VI. Conditions under which scaling effects were eliminated from the FVC were identified. Numerical simulations verifying the monotonicity and the practical utility of the scaling theory were evaluated using numerical experiments applied to Landsat7-Enhanced Thematic Mapper Plus (ETM+) data. The findings contribute to developing scale-invariant FVC estimation algorithms for multisensor and data continuity.
NASA Astrophysics Data System (ADS)
Nagaoka, Tomoaki; Watanabe, Soichi; Sakurai, Kiyoko; Kunieda, Etsuo; Watanabe, Satoshi; Taki, Masao; Yamanaka, Yukio
2004-01-01
With advances in computer performance, the use of high-resolution voxel models of the entire human body has become more frequent in numerical dosimetries of electromagnetic waves. Using magnetic resonance imaging, we have developed realistic high-resolution whole-body voxel models for Japanese adult males and females of average height and weight. The developed models consist of cubic voxels of 2 mm on each side; the models are segmented into 51 anatomic regions. The adult female model is the first of its kind in the world and both are the first Asian voxel models (representing average Japanese) that enable numerical evaluation of electromagnetic dosimetry at high frequencies of up to 3 GHz. In this paper, we will also describe the basic SAR characteristics of the developed models for the VHF/UHF bands, calculated using the finite-difference time-domain method.
NASA Astrophysics Data System (ADS)
Liu, X.; Zhang, J. X.; Zhao, Z.; Ma, A. D.
2015-06-01
Synthetic aperture radar in the application of remote sensing technology is becoming more and more widely because of its all-time and all-weather operation, feature extraction research in high resolution SAR image has become a hot topic of concern. In particular, with the continuous improvement of airborne SAR image resolution, image texture information become more abundant. It's of great significance to classification and extraction. In this paper, a novel method for built-up areas extraction using both statistical and structural features is proposed according to the built-up texture features. First of all, statistical texture features and structural features are respectively extracted by classical method of gray level co-occurrence matrix and method of variogram function, and the direction information is considered in this process. Next, feature weights are calculated innovatively according to the Bhattacharyya distance. Then, all features are weighted fusion. At last, the fused image is classified with K-means classification method and the built-up areas are extracted after post classification process. The proposed method has been tested by domestic airborne P band polarization SAR images, at the same time, two groups of experiments based on the method of statistical texture and the method of structural texture were carried out respectively. On the basis of qualitative analysis, quantitative analysis based on the built-up area selected artificially is enforced, in the relatively simple experimentation area, detection rate is more than 90%, in the relatively complex experimentation area, detection rate is also higher than the other two methods. In the study-area, the results show that this method can effectively and accurately extract built-up areas in high resolution airborne SAR imagery.
Ghosh, Debarchana; Manson, Steven M.
2013-01-01
In this paper, we present a hybrid approach, robust principal component geographically weighted regression (RPCGWR), in examining urbanization as a function of both extant urban land use and the effect of social and environmental factors in the Twin Cities Metropolitan Area (TCMA) of Minnesota. We used remotely sensed data to treat urbanization via the proxy of impervious surface. We then integrated two different methods, robust principal component analysis (RPCA) and geographically weighted regression (GWR) to create an innovative approach to model urbanization. The RPCGWR results show significant spatial heterogeneity in the relationships between proportion of impervious surface and the explanatory factors in the TCMA. We link this heterogeneity to the sprawling nature of urban land use that has moved outward from the core Twin Cities through to their suburbs and exurbs. PMID:23814454
Mitchell, Nia S; Nassel, Ariann F; Thomas, Deborah
2015-12-01
Obesity rates are higher for ethnic minority, low-income, and rural communities. Programs are needed to support these communities with weight management. We determined the reach of a low-cost, nationally-available weight loss program in Health Resources and Services Administration medically underserved areas (MUAs) and described the demographics of the communities with program locations. This is a cross-sectional analysis of Take Off Pounds Sensibly (TOPS) chapter locations. Geographic information systems technology was used to combine information about TOPS chapter locations, the geographic boundaries of MUAs, and socioeconomic data from the Decennial 2010 Census. TOPS is available in 30 % of MUAs. The typical TOPS chapter is in a Census Tract that is predominantly white, urban, with a median annual income between $25,000 and $50,000. However, there are TOPS chapters in Census Tracts that can be classified as predominantly black or predominantly Hispanic; predominantly rural; and as low or high income. TOPS provides weight management services in MUAs and across many types of communities. TOPS can help treat obesity in the medically underserved. Future research should determine the differential effectiveness among chapters in different types of communities. PMID:26072259
Kimbro, Rachel Tolbert; Brooks-Gunn, Jeanne; McLanahan, Sara
2011-01-01
Although research consistently demonstrates a link between residential context and physical activity for adults and adolescents, less is known about young childrens physical activity. Using data from the U.S. Fragile Families and Child Wellbeing Study (N=1822, 51% male), we explored whether outdoor play and television watching were associated with childrens body mass indexes (BMIs) at age five using OLS regression models, controlling for a wide array of potential confounders, including maternal BMI. We also tested whether subjective and objective neighborhood measures - socioeconomic status (from U.S. Census tract data), type of dwelling, perceived collective efficacy, and interviewer-assessed physical disorder of the immediate environment outside the home -were associated with childrens activities, using negative binomial regression models. Overall, 19% of the sample were overweight (between the 85th and 95th percentiles), and 16% were obese (?95th percentile). Hours of outdoor play were negatively associated with BMI, and hours of television were positively associated with BMI. Moreover, a ratio of outdoor play to television time was a significant predictor of BMI. Higher maternal perceptions of neighborhood collective efficacy were associated with more hours of outdoor play, fewer hours of television viewing, and more trips to a park or playground. In addition, we found that neighborhood physical disorder was associated with both more outdoor play and more television watching. Finally, contrary to expectations, we found that children living in public housing had significantly more hours of outdoor play and watched more television, than other children. We hypothesize that poorer children may have more unstructured time, which they fill with television time but also with outdoor play time; and that children in public housing may be likely to have access to play areas on the grounds of their housing facilities. PMID:21324574
Estimating Average Domain Scores.
ERIC Educational Resources Information Center
Pommerich, Mary; Nicewander, W. Alan
A simulation study was performed to determine whether a group's average percent correct in a content domain could be accurately estimated for groups taking a single test form and not the entire domain of items. Six Item Response Theory (IRT) -based domain score estimation methods were evaluated, under conditions of few items per content area per
NASA Technical Reports Server (NTRS)
Kovich, G.; Moore, R. D.; Urasek, D. C.
1973-01-01
The overall and blade-element performance are presented for an air compressor stage designed to study the effect of weight flow per unit annulus area on efficiency and flow range. At the design speed of 424.8 m/sec the peak efficiency of 0.81 occurred at the design weight flow and a total pressure ratio of 1.56. Design pressure ratio and weight flow were 1.57 and 29.5 kg/sec (65.0 lb/sec), respectively. Stall margin at design speed was 19 percent based on the weight flow and pressure ratio at peak efficiency and at stall.
NASA Astrophysics Data System (ADS)
Franz, Trenton E.; Zreda, M.; Ferre, T. P. A.; Rosolem, R.
2013-10-01
The cosmic-ray neutron probe measures soil moisture over tens of hectares, thus averaging spatially variable soil moisture fields. A previous paper described how variable soil moisture profiles affect the integrated cosmic-ray neutron signal from which depth-average soil moisture is computed. Here, we investigate the effect of horizontal heterogeneity on the relationship between neutron counts and average soil moisture. Observations from a distributed sensor network at a site in southern Arizona indicate that the horizontal component of the total variance of the soil moisture field is less variably in time than the vertical component. Using results from neutron particle transport simulations we show that 1-D binary distributions of soil moisture may affect both the mean and variance of neutron counts of a cosmic-ray neutron detector placed arbitrarily in a soil moisture field, potentially giving rise to an underestimate of the footprint average soil moisture. Similar simulations that used 1-D and 2-D Gaussian soil moisture fields indicate consistent mean and variances of a randomly placed detector if the correlation length scales are short (less than 30 m) and/or the soil moisture field variance is small (<0.032 m6 m-6). Taken together, these soil moisture observations and neutron transport simulations show that horizontal heterogeneity likely has a small effect on the relationship between mean neutron counts and average soil moisture for soils under natural conditions.
van der Pals, Jesper; Hammer-Hansen, Sophia; Nielles-Vallespin, Sonia; Kellman, Peter; Taylor, Joni; Kozlov, Shawn; Hsu, Li-Yueh; Chen, Marcus Y.; Arai, Andrew E.
2015-01-01
Aims Cardiovascular magnetic resonance (CMR) imaging can measure the myocardial area at risk (AAR), but the technique has received criticism for inadequate validation. CMR commonly depicts an AAR that is wider than the infarct, which in turn would require a lateral perfusion gradient within the AAR. We investigated the presence of a lateral perfusion gradient within the AAR and validated CMR measures of AAR against three independent reference standards of high quality. Methods and results Computed tomography (CT) perfusion imaging, microsphere blood flow analysis, T1-weighted 3T CMR and fluorescent microparticle pathology were used to investigate the AAR in a canine model (n = 10) of ischaemia and reperfusion. AAR size by CMR correlated well with CT (R2 = 0.80), microsphere blood flow (R2 = 0.80), and pathology (R2 = 0.74) with good limits of agreement [?0.79 4.02% of the left ventricular mass (LVM) vs. CT; ?1.49 4.04% LVM vs. blood flow and ?1.01 4.18% LVM vs. pathology]. The lateral portion of the AAR had higher perfusion than the core of the AAR by CT perfusion imaging (40.7 11.8 vs. 25.2 17.7 Hounsfield units, P = 0.0008) and microsphere blood flow (0.11 0.04 vs. 0.05 0.02 mL/g/min, lateral vs. core, P = 0.001). The transmural extent of MI was lower in the lateral portion of the AAR than the core (28.2 10.2 vs. 17.4 8.4% of the wall, P = 0.001). Conclusion T1-weighted CMR accurately quantifies size of the AAR with excellent agreement compared with three independent reference standards. A lateral perfusion gradient results in lower transmural extent of infarction at the edges of the AAR compared with the core. PMID:25881901
NASA Astrophysics Data System (ADS)
Rung-Arunwan, Tawat; Siripunvaraporn, Weerachai; Utada, Hisashi
2016-04-01
Through a large number of magnetotelluric (MT) observations conducted in a study area, one can obtain regional one-dimensional (1-D) features of the subsurface electrical conductivity structure simply by taking the geometric average of determinant invariants of observed impedances. This method was proposed by Berdichevsky and coworkers, which is based on the expectation that distortion effects due to near-surface electrical heterogeneities will be statistically smoothed out. A good estimation of a regional mean 1-D model is useful, especially in recent years, to be used as a priori (or a starting) model in 3-D inversion. However, the original theory was derived before the establishment of the present knowledge on galvanic distortion. This paper, therefore, reexamines the meaning of the Berdichevsky average by using the conventional formulation of galvanic distortion. A simple derivation shows that the determinant invariant of distorted impedance and its Berdichevsky average is always downward biased by the distortion parameters of shear and splitting. This means that the regional mean 1-D model obtained from the Berdichevsky average tends to be more conductive. As an alternative rotational invariant, the sum of the squared elements (ssq) invariant is found to be less affected by bias from distortion parameters; thus, we conclude that its geometric average would be more suitable for estimating the regional structure. We find that the combination of determinant and ssq invariants provides parameters useful in dealing with a set of distorted MT impedances.
Kalauzi, Aleksandar; Nikoli?, Ljiljana; Savi?, Danijela; Radoti?, Ksenija
2015-02-01
The physiological and biochemical factors that lead to cell death have not been recognized completely. To our knowledge, there are no data on the bioelectric parameters that characterize early period of cell death, as well as on the appearance of related membrane current frequencies. We studied early parameters of glutaraldehyde (GA)-induced cell death, by examining the membrane properties of mouse microglia using the whole-cell patch-clamp technique. In addition, we investigated the GA-induced changes in the membrane current frequency, to see if characteristic frequencies would appear in dying cell. For data analysis, we applied a new approach, an improved multiple moving window length analysis and interval weighted spectra averaging (IWSA). We chose GA for its ability to induce almost instantaneous cell death. The 0.6% GA did not induce changes in the bioelectric membrane properties of microglia. However, the 3% GA caused significant decrease of membrane capacitance and resistance accompanied by the prominent increase in the membrane currents and nearly ohmic current response of microglial cells. These data indicate that 3% GA caused complete loss of the membrane function consequently inducing instantaneous cell death. The membrane function loss was characterized by appearance of the 1.26-4.62 Hz frequency peak in the IWSA spectra, while no significant increase of amplitudes could be observed for cells treated with 0.6% GA. To our knowledge, this is the first record of a frequency associated with complete loss of the membrane function and thus can be considered as an early indicator of cell death. PMID:25367146
NASA Astrophysics Data System (ADS)
Shi, Y.; Long, Y.; Wi, X. L.
2014-04-01
When tourists visiting multiple tourist scenic spots, the travel line is usually the most effective road network according to the actual tour process, and maybe the travel line is different from planned travel line. For in the field of navigation, a proposed travel line is normally generated automatically by path planning algorithm, considering the scenic spots' positions and road networks. But when a scenic spot have a certain area and have multiple entrances or exits, the traditional described mechanism of single point coordinates is difficult to reflect these own structural features. In order to solve this problem, this paper focuses on the influence on the process of path planning caused by scenic spots' own structural features such as multiple entrances or exits, and then proposes a doubleweighted Graph Model, for the weight of both vertexes and edges of proposed Model can be selected dynamically. And then discusses the model building method, and the optimal path planning algorithm based on Dijkstra algorithm and Prim algorithm. Experimental results show that the optimal planned travel line derived from the proposed model and algorithm is more reasonable, and the travelling order and distance would be further optimized.
NASA Technical Reports Server (NTRS)
Urasek, D. C.; Kovich, G.; Moore, R. D.
1973-01-01
Performance was obtained for a 50-cm-diameter compressor designed for a high weight flow per unit annulus area of 208 (kg/sec)/sq m. Peak efficiency values of 0.83 and 0.79 were obtained for the rotor and stage, respectively. The stall margin for the stage was 23 percent, based on equivalent weight flow and total-pressure ratio at peak efficiency and stall.
Wu, Jihuai; Xiao, Yaoming; Tang, Qunwei; Yue, Gentian; Lin, Jianming; Huang, Miaoliang; Huang, Yunfang; Fan, Leqing; Lan, Zhang; Yin, Shu; Sato, Tsugio
2012-04-10
Light-weight PEDOT-Pt/Ti mesh and Ti/TiO(2) foil electrodes are prepared. Owing to the PEDOT-Pt/Ti photocathode's high transparency, good electrocatalytic activity, and low resistance; the Ti/TiO(2) anode's large specific area and high conductivity, a light-weight backside illuminated large-area (100 cm(2) ) dye-sensitized solar cell achieves an energy conversion efficiency of 6.69% under an outdoors sunlight irradiation of 55 mW cm(-2) . PMID:22407518
NASA Astrophysics Data System (ADS)
Kuchment, L.; Romanov, P.; Gelfan, A.; Demidov, V.; Tarpley, D.
2007-12-01
Improvement of long-range forecasts of snowmelt flood volume is one of key hydrological problems in Northern Russia. Accurate quantitative characterization of snow cover properties required in snowmelt runoff models is challenging in this region since the existing network of hydrometeorological stations is sparse. Application of satellite data for snow monitoring is hampered by large areas of coniferous forests masking the snow pack and by persistent cloudiness in the fall and winter season. In order to enhance quantitative characterization of snowpack properties we have developed a new technique where satellite data are coupled with a snow cover model. The physically-based snowpack model uses interpolated data from ground-based meteorological stations and incorporates a number of products derived from Moderate Resolution Imaging Spectroradiometer (MODIS) onboard Terra and Aqua satellites. The input satellite data include albedo, land surface temperature, leaf area index and the canopy coverage. The outputs of the model are the snow depth, snow density, ice and liquid water content of snow and the snow grain size. The model was tested over a region with a size of ~240 000 km2 (56°N to 60°N, and 48°E to 54°E) located within the NEESPI area. This region includes the Vyatka River basin with the catchment area of about 120 000 km2. Snow pack simulations were conducted for 1 x 1 km grid cells for the spring season of 2002 and 2003. Spatial correlation between the modeled snow extent and the MODIS-derived snow cover distribution over the study area ranged from 0.9-1.0 in the beginning and in the end of the melt season to 0.5-0.6 during the period of intensive snow melt. The analysis of MODIS snow retrievals over the study area demonstrated their good agreement with surface observations. Satellite information on snow cover was not used in the current version of the model, however high accuracy of satellite snow retrievals makes their incorporation in the next version of the model very attractive. In the presentation we will discuss ways to incorporate satellite snow retrievals in the snowpack model and advantages of the use of improved estimates of SWE in runoff hydrograph calculations.
NASA Astrophysics Data System (ADS)
Wang, Gongwen; Chen, Jianping; Li, Qing; Ding, Huoping
2007-06-01
This paper aims to monitor desertification evolution of different stages and assess its factors using remote sensing (RS) data and cellular automata (CA)-geographical information system (GIS) with an adaptive analytic hierarchy process (AHP) to derive weights of desertification factors. The study areas (114E to 117E and 39.5to 42.2N) are one of the important agro-pastoral transitional zone, located in Beijing and its neighboring areas, marginal desertified areas in North China. Desertification information including NDVI and desertification area were derived from the satellite images of 1987TM, 1996TM (with a resolution of 28.5), and 2006 CBERS-(with a resolution of 19.5 m) in study areas. The ancillary data in terms of meteorology, geology, 30m-DEM, hydrography can be statistical analyzed with GIS technology. A CA model based on the desertification factors with AHP-derived weights was built by AML program in ArcGIS workstation to assess the evolution of desertification in different stages (from 1987 to 1996, and from 1996 to 2006). The research results show that desertified areas was increased by 3.28% per year from 1987 to 1996, so was 0.51% per year from 1996 to 2006. Although the weights of desertification factors have some changes in different stages, the main factors including climate, NDVI, and terrain did not change except the values in study areas.
NASA Astrophysics Data System (ADS)
Cea, L.; Garrido, M.; Puertas, J.
2010-03-01
SummaryThis paper presents an experimental validation of two widely used numerical models in urban flood inundation studies, the two-dimensional dynamic and diffusive wave models. Instead of using the common approach in flood inundation modelling, which consists of computing the water depth and velocity fields for a given water discharge, in this study the rainfall intensity is imposed directly in the model, the surface runoff being generated automatically. Both the dynamic and diffusive wave models are implemented in the same unstructured finite volume code, removing in such a way any differences in the numerical discretisation other than the wave approximation used to compute the water velocity. Two different methods for representing buildings are used and compared, the so-called building-block and building-hole approaches. Experimental validation of the models is presented in several simplified laboratory configurations of urban catchments, in which the surface runoff has been measured for different hyetographs. For this purpose, 72 experiments were undertaken in a rainfall simulator, including eight catchment configurations and nine hyetographs. Numerical results show that the dynamic wave model is able to predict the peak discharge and its arrival time, as well as the shape of the outlet hydrograph, while the diffusive wave model gives less accurate results. The experimental validation confirms that, when the geometry of the problem is well defined, depth-averaged dynamic wave models may be used to predict rainfall-runoff from direct precipitation data in urban environments.
ERIC Educational Resources Information Center
Young, Beverly S.
The present study was designed to determine whether conservation of number, weight, volume, area, and mass could be learned and retained by disadvantaged preschool children when taught by an inexperienced classroom teacher. An instructional sequence of 10-minute lessons was presented on alternate days over a 3 1/2 week period by preservice
Code of Federal Regulations, 2010 CFR
2010-10-01
..., by Management Area (weights in metric tons) 2a Table 2a to Part 660, Subpart G Wildlife and Fisheries... COMMERCE (CONTINUED) FISHERIES OFF WEST COAST STATES Pt. 660. Subt. G, Table 2a Table 2a to Part 660... Date Note: At 75 FR 60995, Oct. 1, 2010, subpart G was amended by removing Tables 1a through 2c...
Code of Federal Regulations, 2010 CFR
2010-10-01
... 50 Wildlife and Fisheries 9 2010-10-01 2010-10-01 false 2010, Specifications of ABCs, OYs, and HGs, by Management Area (weights in metric tons) 2a Table 2a to Part 660, Subpart C Wildlife and Fisheries... COMMERCE (CONTINUED) FISHERIES OFF WEST COAST STATES Pt. 660, Subpt. C, Table 2a Table 2a to Part...
Code of Federal Regulations, 2010 CFR
2010-10-01
... 50 Wildlife and Fisheries 9 2010-10-01 2010-10-01 false 2009, Specifications of ABCs, OYs, and HGs, by Management Area (weights in metric tons) 1a Table 1a to Part 660, Subpart C Wildlife and Fisheries... COMMERCE (CONTINUED) FISHERIES OFF WEST COAST STATES Pt. 660, Subpt. C, Table 1a Table 1a to Part...
Code of Federal Regulations, 2010 CFR
2010-10-01
..., by Management Area (weights in metric tons) 1a Table 1a to Part 660, Subpart G Wildlife and Fisheries... COMMERCE (CONTINUED) FISHERIES OFF WEST COAST STATES Pt. 660, Subpt. G, Table 1a Table 1a to Part 660... Date Note: At 75 FR 60995, Oct. 1, 2010, subpart G was amended by removing Tables 1a through 2c...
NASA Astrophysics Data System (ADS)
Ferraris, Stefano; Agnese, Carmelo; Baiamonte, Giorgio; Canone, Davide; Previati, Maurizio; Cat Berro, Daniele; Mercalli, Luca
2015-04-01
Modeling of rainfall statistical structure represents an important research area in hydrology, meteorology, atmospheric physics and climatology, because of the several theoretical and practical implications. The statistical inference of the alternation of wet periods (WP) and dry periods (DP) in daily rainfall records can be achieved through the modelling of inter-arrival time-series (IT), defined as the succession of times elapsed from a rainy day and the one immediately preceding it. It has been shown previously that the statistical structure of IT can be well described by the 3-parameter Lerch distribution (Lch). In this work, Lch was successfully applied to IT data belonging to a sub-alpine area (Piemonte and Valle d'Aosta, NW Italy); furthermore the same statistical procedure was applied to daily rainfall records to ITs associated. The analysis has been carried out for 26 daily rainfall long-series (? 90 yr of observations). The main objective of this work was to detect temporal trends of some features describing the statistical structure of both inter-arrival time-series (IT) and associated rainfall depth (H). Each time-series was divided on subsets of five years long and for each of them the estimation of the Lch parameter was performed, so to extend the trend analysis to some high quantiles.
Xaverius, Pamela; Alman, Cameron; Holtz, Lori; Yarber, Laura
2016-03-01
Objectives This study examined risk and protective factors associated with very low birth weight (VLBW) for babies born to women receiving adequate or inadequate prenatal care. Methods Birth records from St. Louis City and County from 2000 to 2009 were used (n = 152,590). Data was categorized across risk factors and stratified by adequacy of prenatal care (PNC). Multivariate logistic regression and population attributable risk (PAR) was used to explore risk factors for VLBW infants. Results Women receiving inadequate prenatal care had a higher prevalence of delivering a VLBW infant than those receiving adequate PNC (4.11 vs. 1.44 %, p < .0001). The distribution of risk factors differed between adequate and inadequate PNC regarding Black race (36.4 vs. 79.0 %, p < .0001), age under 20 (13.0 vs. 33.6 %, p < .0001), <13 years of education (35.9 vs. 77.9 %, p < .0001), Medicaid status (35.7 vs. 74.9, p < .0001), primiparity (41.6 vs. 31.4 %, p < .0001), smoking (9.7 vs. 24.5 %, p < .0001), and diabetes (4.0 vs. 2.4 %, p < .0001), respectively. Black race, advanced maternal age, primiparity and gestational hypertension were significant predictors of VLBW, regardless of adequate or inadequate PNC. Among women with inadequate PNC, Medicaid was protective against (aOR 0.671, 95 % CI 0.563-0.803; PAR -32.6 %) and smoking a risk factor for (aOR 1.23, 95 % CI 1.01, 1.49; PAR 40.1 %) VLBW. When prematurity was added to the adjusted models, the largest PAR shifts to education (44.3 %) among women with inadequate PNC. Conclusions Community actions around broader issues of racism and social determinants of health are needed to prevent VLBW in a large urban area. PMID:26537389
Haines, Aaron M.; Leu, Matthias; Svancara, Leona K.; Wilson, Gina; Scott, J. Michael
2010-01-01
Identification of biodiversity hotspots (hereafter, hotspots) has become a common strategy to delineate important areas for wildlife conservation. However, the use of hotspots has not often incorporated important habitat types, ecosystem services, anthropogenic activity, or consistency in identifying important conservation areas. The purpose of this study was to identify hotspots to improve avian conservation efforts for Species of Greatest Conservation Need (SGCN) in the state of Idaho, United States. We evaluated multiple approaches to define hotspots and used a unique approach based on weighting species by their distribution size and conservation status to identify hotspot areas. All hotspot approaches identified bodies of water (Bear Lake, Grays Lake, and American Falls Reservoir) as important hotspots for Idaho avian SGCN, but we found that the weighted approach produced more congruent hotspot areas when compared to other hotspot approaches. To incorporate anthropogenic activity into hotspot analysis, we grouped species based on their sensitivity to specific human threats (i.e., urban development, agriculture, fire suppression, grazing, roads, and logging) and identified ecological sections within Idaho that may require specific conservation actions to address these human threats using the weighted approach. The Snake River Basalts and Overthrust Mountains ecological sections were important areas for potential implementation of conservation actions to conserve biodiversity. Our approach to identifying hotspots may be useful as part of a larger conservation strategy to aid land managers or local governments in applying conservation actions on the ground.
2012-01-01
Background The study conducts statistical and spatial analyses to investigate amounts and types of permitted surface water pollution discharges in relation to population mortality rates for cancer and non-cancer causes nationwide and by urban-rural setting. Data from the Environmental Protection Agency's (EPA) Discharge Monitoring Report (DMR) were used to measure the location, type, and quantity of a selected set of 38 discharge chemicals for 10,395 facilities across the contiguous US. Exposures were refined by weighting amounts of chemical discharges by their estimated toxicity to human health, and by estimating the discharges that occur not only in a local county, but area-weighted discharges occurring upstream in the same watershed. Centers for Disease Control and Prevention (CDC) mortality files were used to measure age-adjusted population mortality rates for cancer, kidney disease, and total non-cancer causes. Analysis included multiple linear regressions to adjust for population health risk covariates. Spatial analyses were conducted by applying geographically weighted regression to examine the geographic relationships between releases and mortality. Results Greater non-carcinogenic chemical discharge quantities were associated with significantly higher non-cancer mortality rates, regardless of toxicity weighting or upstream discharge weighting. Cancer mortality was higher in association with carcinogenic discharges only after applying toxicity weights. Kidney disease mortality was related to higher non-carcinogenic discharges only when both applying toxicity weights and including upstream discharges. Effects for kidney mortality and total non-cancer mortality were stronger in rural areas than urban areas. Spatial results show correlations between non-carcinogenic discharges and cancer mortality for much of the contiguous United States, suggesting that chemicals not currently recognized as carcinogens may contribute to cancer mortality risk. The geographically weighted regression results suggest spatial variability in effects, and also indicate that some rural communities may be impacted by upstream urban discharges. Conclusions There is evidence that permitted surface water chemical discharges are related to population mortality. Toxicity weights and upstream discharges are important for understanding some mortality effects. Chemicals not currently recognized as carcinogens may nevertheless play a role in contributing to cancer mortality risk. Spatial models allow for the examination of geographic variability not captured through the regression models. PMID:22471926
ERIC Educational Resources Information Center
Gutirrez-Zornoza, Myriam; Snchez-Lpez, Mairena; Garca-Hermoso, Antonio; Gonzlez-Garca, Alberto; Chilln, Palma; Martnez-Vizcano, Vicente
2015-01-01
Purpose: The aim of this study was to examine (a) whether distance from home to school is a determinant of active commuting to school (ACS), (b) the relationship between distance from home to heavily used facilities (school, green spaces, and sports facilities) and the weight status and cardiometabolic risk categories, and (c) whether ACS has a
ERIC Educational Resources Information Center
Gutiérrez-Zornoza, Myriam; Sánchez-López, Mairena; García-Hermoso, Antonio; González-García, Alberto; Chillón, Palma; Martínez-Vizcaíno, Vicente
2015-01-01
Purpose: The aim of this study was to examine (a) whether distance from home to school is a determinant of active commuting to school (ACS), (b) the relationship between distance from home to heavily used facilities (school, green spaces, and sports facilities) and the weight status and cardiometabolic risk categories, and (c) whether ACS has a…
Cosmological Measures with Volume Averaging
NASA Astrophysics Data System (ADS)
Page, Don N.
It has been common for cosmologists to advocate volume weighting for the cosmological measure problem, weighting spatial hypersurfaces by their volume. However, this often leads to the Boltzmann brain problem, that almost all observations would be by momentary Boltzmann brains that arise very briefly as quantum fluctuations in the late universe when it has expanded to a huge size, so that our observations (too ordered for Boltzmann brains) would be highly atypical and unlikely. Here it is suggested that volume weighting may be a mistake. Volume averaging is advocated as an alternative. One consequence may be a loss of the argument that eternal inflation gives a nonzero probability that our universe now has infinite volume.
Vansteenkiste, Ewout; Govaert, Paul; Conneman, Nikk; Lequin, Maarten; Philips, Wilfried
2009-06-01
In this article, we present an interactive algorithm segmenting white brain matter, visible as hyperechoic flaring areas in ultrasound (US) images of preterm infants with periventricular leukomalacia (PVL). The algorithm combines both the textural properties of pathological brain tissue and mathematical morphology operations. An initial flaring area estimate is derived from a multifeature multiclassifier tissue texture classifier. This area is refined based on the structural properties of the choroid plexus, a brain feature known to have characteristics similar to flaring. Subsequently, a combination of a morphological closing, gradient and opening by reconstruction operation determines the final flaring area boundaries. Experimental results are compared with a gold standard constructed from manual flaring area delineations of 12 medical experts. In addition, we compared our algorithm to an existing active contour method. The results show our technique agrees to the gold standard with statistical significance and outperforms the existing method in accuracy. Finally, using the flaring area as a criterion we improve the sensitivity of PVL detection up to 98% as compared with the state of the art. PMID:19251355
ERIC Educational Resources Information Center
Mitchem, John
1989-01-01
Examples used to illustrate Simpson's paradox for secondary students include probabilities, university admissions, batting averages, student-faculty ratios, and average and expected class sizes. Each result is explained. (DC)
Faustino, Laura I; Bulfe, Nardia M L; Pinazo, Martn A; Monteoliva, Silvia E; Graciano, Corina
2013-03-01
Plants of Pinus taeda L. from each of four families were fertilized with nitrogen (N), phosphorus (P) or N?+?P at planting. The H family had the highest growth in dry mass while the L family had the lowest growth. Measurements of plant hydraulic architecture traits were performed during the first year after planting. Stomatal conductance (gs), water potential at predawn (?predawn) and at midday (?midday), branch hydraulic conductivity (ks and kl) and shoot hydraulic conductance (K) were measured. One year after planting, dry weight partitioning of all aboveground organs was performed. Phosphorus fertilization increased growth in all four families, while N fertilization had a negative effect on growth. L family plants were more negatively affected than H family plants. This negative effect was not due to limitations in N or P uptake because plants from all the families and treatments had the same N and P concentration in the needles. Phosphorus fertilization changed some hydraulic parameters, but those changes did not affect growth. However, the negative effect of N can be explained by changes in hydraulic traits. L family plants had a high leaf dry weight per branch, which was increased by N fertilization. This change occurred together with a decrease in shoot conductance. Therefore, the reduction in gs was not enough to avoid the drop in ?midday. Consequently, stomatal closure and the deficient water status of the needles resulted in a reduction in growth. In H family plants, the increase in the number of needles per branch due to N fertilization was counteracted by a reduction in gs and also by a reduction in tracheid lumen size and length. Because of these two changes, ?midday did not drop and water availability in the needles was adequate for sustained growth. In conclusion, fertilization affects the hydraulic architecture of plants, and different families develop different strategies. Some of the hydraulic changes can explain the negative effect of N fertilization on growth. PMID:23355634
2013-01-01
Background Malaria, which frequently occurs in pregnant women in the tropics, is a leading cause of maternal anaemia and low birth weight (LBW) in infants. Few data exist concerning malaria infections that are present at submicroscopic levels during pregnancy and their LBW delivery in babies. Methods A casecontrol study (87 in each group) was conducted at the Medani Hospital, Central Sudan. Cases were women who had LBW deliveries where the infants weighed?weights were recorded immediately after birth. Results Plasmodium falciparum-positive blood films were not obtained from any of the women (cases or controls). Twenty-seven (31.0%) versus 22 (25.3%) (P?=?0.500) of the cases and controls, respectively, had placental malaria infections as determined by histological examination. In comparison to the controls, the submicroscopic malaria infection prevalence rates were significantly higher in the cases; 24 (27.6%) vs six (7.0%), P?
NASA Astrophysics Data System (ADS)
Gnanvo, Kondo; Bai, Xinzhan; Gu, Chao; Liyanage, Nilanga; Nelyubin, Vladimir; Zhao, Yuxiang
2016-02-01
A large-area and light-weight gas electron multiplier (GEM) detector was built at the University of Virginia as a prototype for the detector R&D program of the future Electron Ion Collider. The prototype has a trapezoidal geometry designed as a generic sector module in a disk layer configuration of a forward tracker in collider detectors. It is based on light-weight material and narrow support frames in order to minimize multiple scattering and dead-to-sensitive area ratio. The chamber has a novel type of two dimensional (2D) stereo-angle readout board with U-V strips that provides (r,φ) position information in the cylindrical coordinate system of a collider environment. The prototype was tested at the Fermilab Test Beam Facility in October 2013 and the analysis of the test beam data demonstrates an excellent response uniformity of the large area chamber with an efficiency higher than 95%. An angular resolution of 60 μrad in the azimuthal direction and a position resolution better than 550 μm in the radial direction were achieved with the U-V strip readout board. The results are discussed in this paper.
Hmelnitsky, I; Nettheim, N
1987-06-01
Functional anatomy and physiology have naturally attended mainly to those functions which occur most commonly in everyday life. Piano playing is a more specialized area, where functions arise which have so far been neglected in medical science. These functions are here described by a pianist (IH) in the hope that medical researchers will respond to fill the gaps. The importance of this lies not only in the understanding of skilled manipulative activity but also in the avoidance of overuse syndrome (OUS) or repetitive strain injury (RSI). PMID:3614013
States' Average College Tuition.
ERIC Educational Resources Information Center
Eglin, Joseph J., Jr.; And Others
This report presents statistical data on trends in tuition costs from 1980-81 through 1995-96. The average tuition for in-state undergraduate students of 4-year public colleges and universities for academic year 1995-96 was approximately 8.9 percent of median household income. This figure was obtained by dividing the students' average annual…
NASA Astrophysics Data System (ADS)
Zavala, M.; Herndon, S. C.; Slott, R. S.; Dunlea, E. J.; Marr, L. C.; Shorter, J. H.; Zahniser, M.; Knighton, W. B.; Rogers, T. M.; Kolb, C. E.; Molina, L. T.; Molina, M. J.
2006-06-01
A mobile laboratory was used to measure on-road vehicle emission ratios during the MCMA-2003 field campaign held during the spring of 2003 in the Mexico City Metropolitan Area (MCMA). The measured emission ratios represent a sample of emissions of in-use vehicles under real world driving conditions for the MCMA. From the relative amounts of NOx and selected VOC's sampled, the results indicate that the technique is capable of differentiating among vehicle categories and fuel type in real world driving conditions. Emission ratios for NOx, NOy, NH3, H2CO, CH3CHO, and other selected volatile organic compounds (VOCs) are presented for chase sampled vehicles and fleet averaged emissions. Results indicate that colectivos, particularly CNG-powered colectivos, are potentially significant contributors of NOx and aldehydes in the MCMA. Similarly, ratios of selected VOCs and NOy showed a strong dependence on traffic mode. These results are compared with the vehicle emissions inventory for the MCMA, other vehicle emissions measurements in the MCMA, and measurements of on-road emissions in US cities. Our estimates for motor vehicle emissions of benzene, toluene, formaldehyde, and acetaldehyde in the MCMA indicate these species are present in concentrations higher than previously reported. The high motor vehicle aldehyde emissions may have an impact on the photochemistry of urban areas.
Hakkinen, P.J.; Kelling, C.K.; Callender, J.C. )
1991-02-01
A thorough understanding of the routes and magnitudes of chemical exposures that consumers experience during the use of a household product is needed as part of a well-founded risk assessment for that product and its components. This review describes some sources of generic consumer data (eg, relevant body weight or total body surface area for a given human age), and exposure-related data (eg, task frequency and duration) for specific product types needed for exposure assessments. The review also contains a discussion of the importance of statistical characterization of the consumer data (eg, does its range follow a normal, log-normal, or other type of distribution ). The importance of examining these data for correlative interactions is emphasized.25 references.
ERIC Educational Resources Information Center
Young, Vershawn Ashanti
2004-01-01
"Your Average Nigga" contends that just as exaggerating the differences between black and white language leaves some black speakers, especially those from the ghetto, at an impasse, so exaggerating and reifying the differences between the races leaves blacks in the impossible position of either having to try to be white or forever struggling to
From cellular doses to average lung dose.
Hofmann, W; Winkler-Heil, R
2015-11-01
Sensitive basal and secretory cells receive a wide range of doses in human bronchial and bronchiolar airways. Variations of cellular doses arise from the location of target cells in the bronchial epithelium of a given airway and the asymmetry and variability of airway dimensions of the lung among airways in a given airway generation and among bronchial and bronchiolar airway generations. To derive a single value for the average lung dose which can be related to epidemiologically observed lung cancer risk, appropriate weighting scenarios have to be applied. Potential biological weighting parameters are the relative frequency of target cells, the number of progenitor cells, the contribution of dose enhancement at airway bifurcations, the promotional effect of cigarette smoking and, finally, the application of appropriate regional apportionment factors. Depending on the choice of weighting parameters, detriment-weighted average lung doses can vary by a factor of up to 4 for given radon progeny exposure conditions. PMID:25920789
Covariant approximation averaging
NASA Astrophysics Data System (ADS)
Shintani, Eigo; Arthur, Rudy; Blum, Thomas; Izubuchi, Taku; Jung, Chulwoo; Lehner, Christoph
2015-06-01
We present a new class of statistical error reduction techniques for Monte Carlo simulations. Using covariant symmetries, we show that correlation functions can be constructed from inexpensive approximations without introducing any systematic bias in the final result. We introduce a new class of covariant approximation averaging techniques, known as all-mode averaging (AMA), in which the approximation takes account of contributions of all eigenmodes through the inverse of the Dirac operator computed from the conjugate gradient method with a relaxed stopping condition. In this paper we compare the performance and computational cost of our new method with traditional methods using correlation functions and masses of the pion, nucleon, and vector meson in Nf=2 +1 lattice QCD using domain-wall fermions. This comparison indicates that AMA significantly reduces statistical errors in Monte Carlo calculations over conventional methods for the same cost.
Wire weight is lowered to water surface to measure stage at a site. Levels are made to the wire weights elevation from known benchmarks to ensure correct readings. This wire weight is located along the Missouri River in Bismarck, ND....
NASA Astrophysics Data System (ADS)
Zavala, M.; Herndon, S. C.; Slott, R. S.; Dunlea, E. J.; Marr, L. C.; Shorter, J. H.; Zahniser, M.; Knighton, W. B.; Rogers, T. M.; Kolb, C. E.; Molina, L. T.; Molina, M. J.
2006-11-01
A mobile laboratory was used to measure on-road vehicle emission ratios during the MCMA-2003 field campaign held during the spring of 2003 in the Mexico City Metropolitan Area (MCMA). The measured emission ratios represent a sample of emissions of in-use vehicles under real world driving conditions for the MCMA. From the relative amounts of NOx and selected VOC's sampled, the results indicate that the technique is capable of differentiating among vehicle categories and fuel type in real world driving conditions. Emission ratios for NOx, NOy, NH3, H2CO, CH3CHO, and other selected volatile organic compounds (VOCs) are presented for chase sampled vehicles in the form of frequency distributions as well as estimates for the fleet averaged emissions. Our measurements of emission ratios for both CNG and gasoline powered "colectivos" (public transportation buses that are intensively used in the MCMA) indicate that - in a mole per mole basis - have significantly larger NOx and aldehydes emissions ratios as compared to other sampled vehicles in the MCMA. Similarly, ratios of selected VOCs and NOy showed a strong dependence on traffic mode. These results are compared with the vehicle emissions inventory for the MCMA, other vehicle emissions measurements in the MCMA, and measurements of on-road emissions in U.S. cities. We estimate NOx emissions as 100 600±29 200 metric tons per year for light duty gasoline vehicles in the MCMA for 2003. According to these results, annual NOx emissions estimated in the emissions inventory for this category are within the range of our estimated NOx annual emissions. Our estimates for motor vehicle emissions of benzene, toluene, formaldehyde, and acetaldehyde in the MCMA indicate these species are present in concentrations higher than previously reported. The high motor vehicle aldehyde emissions may have an impact on the photochemistry of urban areas.
Gong, Lunli; Zhou, Xiao; Wu, Yaohao; Zhang, Yun; Wang, Chen; Zhou, Heng; Guo, Fangfang; Cui, Lei
2014-02-01
The present study was designed to investigate the possibility of full-thickness defects repair in porcine articular cartilage (AC) weight-bearing area using chondrogenic differentiated autologous adipose-derived stem cells (ASCs) with a follow-up of 3 and 6 months, which is successive to our previous study on nonweight-bearing area. The isolated ASCs were seeded onto the phosphoglycerate/polylactic acid (PGA/PLA) with chondrogenic induction in vitro for 2 weeks as the experimental group prior to implantation in porcine AC defects (8 mm in diameter, deep to subchondral bone), with PGA/PLA only as control. With follow-up time being 3 and 6 months, both neo-cartilages of postimplantation integrated well with the neighboring normal cartilage and subchondral bone histologically in experimental group, whereas only fibrous tissue in control group. Immunohistochemical and toluidine blue staining confirmed similar distribution of COL II and glycosaminoglycan in the regenerated cartilage to the native one. A vivid remolding process with repair time was also witnessed in the neo-cartilage as the compressive modulus significantly increased from 70% of the normal cartilage at 3 months to nearly 90% at 6 months, which is similar to our former research. Nevertheless, differences of the regenerated cartilages still could be detected from the native one. Meanwhile, the exact mechanism involved in chondrogenic differentiation from ASCs seeded on PGA/PLA is still unknown. Therefore, proteome is resorted leading to 43 proteins differentially identified from 20 chosen two-dimensional spots, which do help us further our research on some committed factors. In conclusion, the comparison via proteome provided a thorough understanding of mechanisms implicating ASC differentiation toward chondrocytes, which is further substantiated by the present study as a perfect supplement to the former one in nonweight-bearing area. PMID:24044689
Gong, Lunli; Zhou, Xiao; Wu, Yaohao; Zhang, Yun; Wang, Chen; Zhou, Heng; Guo, Fangfang
2014-01-01
The present study was designed to investigate the possibility of full-thickness defects repair in porcine articular cartilage (AC) weight-bearing area using chondrogenic differentiated autologous adipose-derived stem cells (ASCs) with a follow-up of 3 and 6 months, which is successive to our previous study on nonweight-bearing area. The isolated ASCs were seeded onto the phosphoglycerate/polylactic acid (PGA/PLA) with chondrogenic induction in vitro for 2 weeks as the experimental group prior to implantation in porcine AC defects (8?mm in diameter, deep to subchondral bone), with PGA/PLA only as control. With follow-up time being 3 and 6 months, both neo-cartilages of postimplantation integrated well with the neighboring normal cartilage and subchondral bone histologically in experimental group, whereas only fibrous tissue in control group. Immunohistochemical and toluidine blue staining confirmed similar distribution of COL II and glycosaminoglycan in the regenerated cartilage to the native one. A vivid remolding process with repair time was also witnessed in the neo-cartilage as the compressive modulus significantly increased from 70% of the normal cartilage at 3 months to nearly 90% at 6 months, which is similar to our former research. Nevertheless, differences of the regenerated cartilages still could be detected from the native one. Meanwhile, the exact mechanism involved in chondrogenic differentiation from ASCs seeded on PGA/PLA is still unknown. Therefore, proteome is resorted leading to 43 proteins differentially identified from 20 chosen two-dimensional spots, which do help us further our research on some committed factors. In conclusion, the comparison via proteome provided a thorough understanding of mechanisms implicating ASC differentiation toward chondrocytes, which is further substantiated by the present study as a perfect supplement to the former one in nonweight-bearing area. PMID:24044689
Srivatsav, Siddhart; Webster, Jacquelyn; Webster, Michael
2015-01-01
The average color in a scene is a potentially important cue to the illuminant and thus for color constancy, but it remains unknown how well and in what ways observers can estimate the mean chromaticity. We examined this by measuring the variability in "achromatic" settings for stimuli composed of different distributions of colors. The displays consisted of a 15 by 15 palette of colors shown on a gray background on a monitor, with each chip subtending 0.5 deg. Individual colors were randomly sampled from varying contrast ranges along the luminance, S and LM cardinal axes. Observers were instructed to adjust the chromaticity of the palette so that the mean was gray, with variability estimated from 20 or more repeated settings. This variability increased progressively with increasing contrast in the distributions, with large increases for chromatic contrast but also weak effects for added luminance contrast. Signals along the cardinal axes are relatively independent in many detection and discrimination tasks, but showed strong interference in the white estimates. Specifically, adding S contrast increased variability in the white settings along both the S and LM axes, and vice versa. This "cross-masking" and the effects of chromatic variance in general may occur because observers cannot explicitly perceive or represent the mean of a set of qualitatively different hues (e.g. that red and green hues average to gray), and thus may infer the mean only indirectly (e.g. from the relative saturation of different hues). Meeting abstract presented at VSS 2015. PMID:26326088
Approaches for averaging surface parameters and fluxes over heterogeneous terrain
NASA Technical Reports Server (NTRS)
Chehbouni, A.; Njoku, E. G.; Lhomme, J.-P.; Kerr, Y. H.
1995-01-01
Successful prediction of possible climate change depends on realistic parameterization of land surface processes in climate models. Such parameterizations must take appropriate account of the heterogeneities that are found in most earth surfaces. In this study, different averaging strategies for aggregating patch-scale heterogeneities to scales that are appropriate for mesoscale and climate model grids have been explored. A simple model for estimating area-average 'effective' surface flux parameters is evaluated. The model satisfies the energy balance equation and leads to a set of relationships between local and effective parameters in the governing equations for the surface energy balance. One outcome is that the resulting effective surface temperature is not a simple area-weighted average of component temperatures, but is a function of a specific combination of different resistances of the individual surface elements. A set of heterogeneous surfaces has been simulated to study the effective fluxes obtained using the described model. A comparison with results obtained by other investigators using different averaging methods is also performed.
Americans' Average Radiation Exposure
NA
2000-08-11
We live with radiation every day. We receive radiation exposures from cosmic rays, from outer space, from radon gas, and from other naturally radioactive elements in the earth. This is called natural background radiation. It includes the radiation we get from plants, animals, and from our own bodies. We also are exposed to man-made sources of radiation, including medical and dental treatments, television sets and emission from coal-fired power plants. Generally, radiation exposures from man-made sources are only a fraction of those received from natural sources. One exception is high exposures used by doctors to treat cancer patients. Each year in the United States, the average dose to people from natural and man-made radiation sources is about 360 millirem. A millirem is an extremely tiny amount of energy absorbed by tissues in the body.
Temperature averaging thermal probe
NASA Technical Reports Server (NTRS)
Kalil, L. F.; Reinhardt, V. (inventors)
1985-01-01
A thermal probe to average temperature fluctuations over a prolonged period was formed with a temperature sensor embedded inside a solid object of a thermally conducting material. The solid object is held in a position equidistantly spaced apart from the interior surfaces of a closed housing by a mount made of a thermally insulating material. The housing is sealed to trap a vacuum or mass of air inside and thereby prevent transfer of heat directly between the environment outside of the housing and the solid object. Electrical leads couple the temperature sensor with a connector on the outside of the housing. Other solid objects of different sizes and materials may be substituted for the cylindrically-shaped object to vary the time constant of the probe.
Córdova-Palomera, Aldo; Fatjó-Vilas, Mar; Falcón, Carles; Bargalló, Nuria; Alemany, Silvia; Crespo-Facorro, Benedicto; Nenadic, Igor; Fañanás, Lourdes
2015-01-01
Background Previous research suggests that low birth weight (BW) induces reduced brain cortical surface area (SA) which would persist until at least early adulthood. Moreover, low BW has been linked to psychiatric disorders such as depression and psychological distress, and to altered neurocognitive profiles. Aims We present novel findings obtained by analysing high-resolution structural MRI scans of 48 twins; specifically, we aimed: i) to test the BW-SA association in a middle-aged adult sample; and ii) to assess whether either depression/anxiety disorders or intellectual quotient (IQ) influence the BW-SA link, using a monozygotic (MZ) twin design to separate environmental and genetic effects. Results Both lower BW and decreased IQ were associated with smaller total and regional cortical SA in adulthood. Within a twin pair, lower BW was related to smaller total cortical and regional SA. In contrast, MZ twin differences in SA were not related to differences in either IQ or depression/anxiety disorders. Conclusion The present study supports findings indicating that i) BW has a long-lasting effect on cortical SA, where some familial and environmental influences alter both foetal growth and brain morphology; ii) uniquely environmental factors affecting BW also alter SA; iii) higher IQ correlates with larger SA; and iv) these effects are not modified by internalizing psychopathology. PMID:26086820
Wang, Tingting; Li, Wenhua; Wu, Xiangru; Yin, Bing; Chu, Caiting; Ding, Ming; Cui, Yanfen
2016-01-01
Objective To assess the added value of diffusion-weighted magnetic resonance imaging (DWI) with apparent diffusion coefficient (ADC) values compared to MRI, for characterizing the tubo-ovarian abscesses (TOA) mimicking ovarian malignancy. Materials and Methods Patients with TOA (or ovarian abscess alone; n = 34) or ovarian malignancy (n = 35) who underwent DWI and MRI were retrospectively reviewed. The signal intensity of cystic and solid component of TOAs and ovarian malignant tumors on DWI and the corresponding ADC values were evaluated, as well as clinical characteristics, morphological features, MRI findings were comparatively analyzed. Receiver operating characteristic (ROC) curve analysis based on logistic regression was applied to identify different imaging characteristics between the two patient groups and assess the predictive value of combination diagnosis with area under the curve (AUC) analysis. Results The mean ADC value of the cystic component in TOA was significantly lower than in malignant tumors (1.04 ± 0 .41 × 10−3 mm2/s vs. 2.42 ± 0.38 × 10−3 mm2/s; p < 0.001). The mean ADC value of the enhanced solid component in 26 TOAs was 1.43 ± 0.16×10−3mm2/s, and 46.2% (12 TOAs; pseudotumor areas) showed significantly higher signal intensity on DW-MRI than in ovarian malignancy (mean ADC value 1.44 ± 0.20×10−3 mm2/s vs.1.18 ± 0.36 × 10−3 mm2/s; p = 0.043). The combination diagnosis of ADC value and dilated tubal structure achieved the best AUC of 0.996. Sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and accuracy of MRI vs. DWI with ADC values for predicting TOA were 47.1%, 91.4%, 84.2%, 64%, and 69.6% vs. 100%, 97.1%, 97.1%, 100%, and 98.6%, respectively. Conclusions DW-MRI is superior to MRI in the assessment of TOA mimicking ovarian malignancy, and the ADC values aid in discriminating the pseudotumor area of TOA from the solid portion of ovarian malignancy. PMID:26894926
42 CFR 423.279 - National average monthly bid amount.
Code of Federal Regulations, 2013 CFR
2013-10-01
... each MA-PD plan described in section 1851(a)(2)(A)(i) of the Act. The calculation does not include bids... equal weighting to PDP sponsors (other than fallback entities) and assigns MA-PD plans included in the national average bid a weight based on prior enrollment (new MA-PD plans are assigned zero weight)....
42 CFR 423.279 - National average monthly bid amount.
Code of Federal Regulations, 2012 CFR
2012-10-01
... each MA-PD plan described in section 1851(a)(2)(A)(i) of the Act. The calculation does not include bids... equal weighting to PDP sponsors (other than fallback entities) and assigns MA-PD plans included in the national average bid a weight based on prior enrollment (new MA-PD plans are assigned zero weight)....
42 CFR 423.279 - National average monthly bid amount.
Code of Federal Regulations, 2014 CFR
2014-10-01
... each MA-PD plan described in section 1851(a)(2)(A)(i) of the Act. The calculation does not include bids... equal weighting to PDP sponsors (other than fallback entities) and assigns MA-PD plans included in the national average bid a weight based on prior enrollment (new MA-PD plans are assigned zero weight)....
... weight, the calories you eat must equal the energy you burn. To lose weight, you must use more calories than you eat. A weight-control strategy might include Choosing low-fat, low-calorie foods Eating smaller portions Drinking water instead of sugary ...
Dissociating Averageness and Attractiveness: Attractive Faces Are Not Always Average
ERIC Educational Resources Information Center
DeBruine, Lisa M.; Jones, Benedict C.; Unger, Layla; Little, Anthony C.; Feinberg, David R.
2007-01-01
Although the averageness hypothesis of facial attractiveness proposes that the attractiveness of faces is mostly a consequence of their averageness, 1 study has shown that caricaturing highly attractive faces makes them mathematically less average but more attractive. Here the authors systematically test the averageness hypothesis in 5 experiments
Combining remotely sensed and other measurements for hydrologic areal averages
NASA Technical Reports Server (NTRS)
Johnson, E. R.; Peck, E. L.; Keefer, T. N.
1982-01-01
A method is described for combining measurements of hydrologic variables of various sampling geometries and measurement accuracies to produce an estimated mean areal value over a watershed and a measure of the accuracy of the mean areal value. The method provides a means to integrate measurements from conventional hydrological networks and remote sensing. The resulting areal averages can be used to enhance a wide variety of hydrological applications including basin modeling. The correlation area method assigns weights to each available measurement (point, line, or areal) based on the area of the basin most accurately represented by the measurement. The statistical characteristics of the accuracy of the various measurement technologies and of the random fields of the hydrologic variables used in the study (water equivalent of the snow cover and soil moisture) required to implement the method are discussed.
NASA Astrophysics Data System (ADS)
Du, Wen-Bo; Cao, Xian-Bin; Zhao, Lin; Zhou, Hong
2009-05-01
We investigate the evolutionary prisoner's dilemma game (PDG) on weighted Newman-Watts (NW) networks. In weighted NW networks, the link weight wij is assigned to the link between the nodes i and j as: wij = (?i ?j)?, where ?i(?j) is the degree of node i(j) and ? represents the strength of the correlations. Obviously, the link weight can be tuned by only one parameter ?. We focus on the cooperative behavior and wealth distribution in the system. Simulation results show that the cooperator frequency is promoted by a large range of ? and there is a minimal cooperation frequency around ? = - 1. Moreover, we also employ the Gini coefficient to study the wealth distribution in the population. Numerical results show that the Gini coefficient reaches its minimum when ? approx - 1. Our work may be helpful in understanding the emergence of cooperation and unequal wealth distribution in society.
... is more than 8.8 pounds. A low birth weight baby can be born too small, too early (premature), or both. This can happen for many different reasons. They include health problems in the mother, genetic factors, problems ... by the mother. Some low birth weight babies may be more at risk for ...
NASA Technical Reports Server (NTRS)
Moore, R. D.; Urasek, D. C.; Kovich, G.
1973-01-01
The overall and blade-element performances are presented over the stable flow operating range from 50 to 100 percent of design speed. Stage peak efficiency of 0.834 was obtained at a weight flow of 26.4 kg/sec (58.3 lb/sec) and a pressure ratio of 1.581. The stall margin for the stage was 7.5 percent based on weight flow and pressure ratio at stall and peak efficiency conditions. The rotor minimum losses were approximately equal to design except in the blade vibration damper region. Stator minimum losses were less than design except in the tip and damper regions.
Model averaging and muddled multimodel inferences.
Cade, Brian S
2015-09-01
Three flawed practices associated with model averaging coefficients for predictor variables in regression models commonly occur when making multimodel inferences in analyses of ecological data. Model-averaged regression coefficients based on Akaike information criterion (AIC) weights have been recommended for addressing model uncertainty but they are not valid, interpretable estimates of partial effects for individual predictors when there is multicollinearity among the predictor variables. Multicollinearity implies that the scaling of units in the denominators of the regression coefficients may change across models such that neither the parameters nor their estimates have common scales, therefore averaging them makes no sense. The associated sums of AIC model weights recommended to assess relative importance of individual predictors are really a measure of relative importance of models, with little information about contributions by individual predictors compared to other measures of relative importance based on effects size or variance reduction. Sometimes the model-averaged regression coefficients for predictor variables are incorrectly used to make model-averaged predictions of the response variable when the models are not linear in the parameters. I demonstrate the issues with the first two practices using the college grade point average example extensively analyzed by Burnham and Anderson. I show how partial standard deviations of the predictor variables can be used to detect changing scales of their estimates with multicollinearity. Standardizing estimates based on partial standard deviations for their variables can be used to make the scaling of the estimates commensurate across models, a necessary but not sufficient condition for model averaging of the estimates to be sensible. A unimodal distribution of estimates and valid interpretation of individual parameters are additional requisite conditions. The standardized estimates or equivalently the t statistics on unstandardized estimates also can be used to provide more informative measures of relative importance than sums of AIC weights. Finally, I illustrate how seriously compromised statistical interpretations and predictions can be for all three of these flawed practices by critiquing their use in a recent species distribution modeling technique developed for predicting Greater Sage-Grouse (Centrocercus urophasianus) distribution in Colorado, USA. These model averaging issues are common in other ecological literature and ought to be discontinued if we are to make effective scientific contributions to ecological knowledge and conservation of natural resources. PMID:26594695
Model Averaging Method for Supersaturated Experimental Design
NASA Astrophysics Data System (ADS)
Salaki, Deiby T.; Kurnia, Anang; Sartono, Bagus
2016-01-01
In this paper, a new modified model averaging method was proposed. The candidate model construction was performed by distinguishing the covariates into focus variables and auxiliary variables whereas the weights selection was implemented using Mallows criterion. In addition, the illustration result shows that the applied model averaging method could be considered as a new alternative method for supersaturated experimental design as a typical form of high dimensional data. A supersaturated factorial design is an experimental series in which the number of factors exceeds the number of runs, so its size is not enough to estimate all the main effect. By using the model averaging method, the estimation or prediction power is significantly enhanced. In our illustration, the main factors are regarded as focus variables in order to give more attention to them whereas the lesser factors are regarded as auxiliary variables, which is along with the hierarchical ordering principle in experimental research. The limited empirical study shows that this method produces good prediction.
NASA Technical Reports Server (NTRS)
Howard, W. H.; Young, D. R.
1972-01-01
Device applies compressive force to bone to minimize loss of bone calcium during weightlessness or bedrest. Force is applied through weights, or hydraulic, pneumatic or electrically actuated devices. Device is lightweight and easy to maintain and operate.
Weight and weddings. Engaged men's body weight ideals and wedding weight management behaviors.
Klos, Lori A; Sobal, Jeffery
2013-01-01
Most adults marry at some point in life, and many invest substantial resources in a wedding ceremony. Previous research reports that brides often strive towards culturally-bound appearance norms and engage in weight management behaviors in preparation for their wedding. However, little is known about wedding weight ideals and behaviors among engaged men. A cross-sectional survey of 163 engaged men asked them to complete a questionnaire about their current height and weight, ideal wedding body weight, wedding weight importance, weight management behaviors, formality of their upcoming wedding ceremony, and demographics. Results indicated that the discrepancy between men's current weight and reported ideal wedding weight averaged 9.61 lb. Most men considered being at a certain weight at their wedding to be somewhat important. About 39% were attempting to lose weight for their wedding, and 37% were not trying to change their weight. Attempting weight loss was more frequent among men with higher BMI's, those planning more formal weddings, and those who considered being the right weight at their wedding as important. Overall, these findings suggest that weight-related appearance norms and weight loss behaviors are evident among engaged men. PMID:23063607
NASA Technical Reports Server (NTRS)
Chen, Fei; Yates, David; LeMone, Margaret
2001-01-01
To understand the effects of land-surface heterogeneity and the interactions between the land-surface and the planetary boundary layer at different scales, we develop a multiscale data set. This data set, based on the Cooperative Atmosphere-Surface Exchange Study (CASES97) observations, includes atmospheric, surface, and sub-surface observations obtained from a dense observation network covering a large region on the order of 100 km. We use this data set to drive three land-surface models (LSMs) to generate multi-scale (with three resolutions of 1, 5, and 10 kilometers) gridded surface heat flux maps for the CASES area. Upon validating these flux maps with measurements from surface station and aircraft, we utilize them to investigate several approaches for estimating the area-integrated surface heat flux for the CASES97 domain of 71x74 square kilometers, which is crucial for land surface model development/validation and area water and energy budget studies. This research is aimed at understanding the relative contribution of random turbulence versus organized mesoscale circulations to the area-integrated surface flux at the scale of 100 kilometers, and identifying the most important effective parameters for characterizing the subgrid-scale variability for large-scale atmosphere-hydrology models.
The Molecular Weight Distribution of Polymer Samples
ERIC Educational Resources Information Center
Horta, Arturo; Pastoriza, M. Alejandra
2007-01-01
Various methods for the determination of the molecular weight distribution (MWD) of different polymer samples are presented. The study shows that the molecular weight averages and distribution of a polymerization completely depend on the characteristics of the reaction itself.
42 CFR 423.279 - National average monthly bid amount.
Code of Federal Regulations, 2011 CFR
2011-10-01
... bid amounts for each prescription drug plan (not including fallbacks) and for each MA-PD plan... to PDP sponsors (other than fallback entities) and assigns MA-PD plans included in the national average bid a weight based on prior enrollment (new MA-PD plans are assigned zero weight). (c)...
Chen, Chieh-Li; Ishikawa, Hiroshi; Wollstein, Gadi; Bilonick, Richard A.; Kagemann, Larry; Schuman, Joel S.
2016-01-01
Purpose Developing a novel image enhancement method so that nonframe-averaged optical coherence tomography (OCT) images become comparable to active eye-tracking frame-averaged OCT images. Methods Twenty-one eyes of 21 healthy volunteers were scanned with noneye-tracking nonframe-averaged OCT device and active eye-tracking frame-averaged OCT device. Virtual averaging was applied to nonframe-averaged images with voxel resampling and adding amplitude deviation with 15-time repetitions. Signal-to-noise (SNR), contrast-to-noise ratios (CNR), and the distance between the end of visible nasal retinal nerve fiber layer (RNFL) and the foveola were assessed to evaluate the image enhancement effect and retinal layer visibility. Retinal thicknesses before and after processing were also measured. Results All virtual-averaged nonframe-averaged images showed notable improvement and clear resemblance to active eye-tracking frame-averaged images. Signal-to-noise and CNR were significantly improved (SNR: 30.5 vs. 47.6 dB, CNR: 4.4 vs. 6.4 dB, original versus processed, P < 0.0001, paired t-test). The distance between the end of visible nasal RNFL and the foveola was significantly different before (681.4 vs. 446.5 μm, Cirrus versus Spectralis, P < 0.0001) but not after processing (442.9 vs. 446.5 μm, P = 0.76). Sectoral macular total retinal and circumpapillary RNFL thicknesses showed systematic differences between Cirrus and Spectralis that became not significant after processing. Conclusion The virtual averaging method successfully improved nontracking nonframe-averaged OCT image quality and made the images comparable to active eye-tracking frame-averaged OCT images. Translational Relevance Virtual averaging may enable detailed retinal structure studies on images acquired using a mixture of nonframe-averaged and frame-averaged OCT devices without concerning about systematic differences in both qualitative and quantitative aspects. PMID:26835180
Spacetime averaged null energy condition
Urban, Douglas; Olum, Ken D.
2010-06-15
The averaged null energy condition has known violations for quantum fields in curved space, even when one considers only achronal geodesics. Many such examples involve rapid variation in the stress-energy tensor in the vicinity of the geodesic under consideration, giving rise to the possibility that averaging in additional dimensions would yield a principle universally obeyed by quantum fields. However, after discussing various procedures for additional averaging, including integrating over all dimensions of the manifold, we give here a class of examples that violate any such averaged condition.
Prediction of shelled shrimp weight by machine vision
Pan, Peng-min; Li, Jian-ping; Lv, Gu-lai; Yang, Hui; Zhu, Song-ming; Lou, Jian-zhong
2009-01-01
The weight of shelled shrimp is an important parameter for grading process. The weight prediction of shelled shrimp by contour area is not accurate enough because of the ignorance of the shrimp thickness. In this paper, a multivariate prediction model containing area, perimeter, length, and width was established. A new calibration algorithm for extracting length of shelled shrimp was proposed, which contains binary image thinning, branch recognition and elimination, and length reconstruction, while its width was calculated during the process of length extracting. The model was further validated with another set of images from 30 shelled shrimps. For a comparison purpose, artificial neural network (ANN) was used for the shrimp weight predication. The ANN model resulted in a better prediction accuracy (with the average relative error at 2.67%), but took a tenfold increase in calculation time compared with the weight-area-perimeter (WAP) model (with the average relative error at 3.02%). We thus conclude that the WAP model is a better method for the prediction of the weight of shelled red shrimp. PMID:19650197
Model averaging, optimal inference, and habit formation
FitzGerald, Thomas H. B.; Dolan, Raymond J.; Friston, Karl J.
2014-01-01
Postulating that the brain performs approximate Bayesian inference generates principled and empirically testable models of neuronal functionthe subject of much current interest in neuroscience and related disciplines. Current formulations address inference and learning under some assumed and particular model. In reality, organisms are often faced with an additional challengethat of determining which model or models of their environment are the best for guiding behavior. Bayesian model averagingwhich says that an agent should weight the predictions of different models according to their evidenceprovides a principled way to solve this problem. Importantly, because model evidence is determined by both the accuracy and complexity of the model, optimal inference requires that these be traded off against one another. This means an agent's behavior should show an equivalent balance. We hypothesize that Bayesian model averaging plays an important role in cognition, given that it is both optimal and realizable within a plausible neuronal architecture. We outline model averaging and how it might be implemented, and then explore a number of implications for brain and behavior. In particular, we propose that model averaging can explain a number of apparently suboptimal phenomena within the framework of approximate (bounded) Bayesian inference, focusing particularly upon the relationship between goal-directed and habitual behavior. PMID:25018724
NASA Technical Reports Server (NTRS)
Ustino, Eugene A.
2006-01-01
This slide presentation reviews the observable radiances as functions of atmospheric parameters and of surface parameters; the mathematics of atmospheric weighting functions (WFs) and surface partial derivatives (PDs) are presented; and the equation of the forward radiative transfer (RT) problem is presented. For non-scattering atmospheres this can be done analytically, and all WFs and PDs can be computed analytically using the direct linearization approach. For scattering atmospheres, in general case, the solution of the forward RT problem can be obtained only numerically, but we need only two numerical solutions: one of the forward RT problem and one of the adjoint RT problem to compute all WFs and PDs we can think of. In this presentation we discuss applications of both the linearization and adjoint approaches
High average power pockels cell
Daly, Thomas P.
1991-01-01
A high average power pockels cell is disclosed which reduces the effect of thermally induced strains in high average power laser technology. The pockels cell includes an elongated, substantially rectangular crystalline structure formed from a KDP-type material to eliminate shear strains. The X- and Y-axes are oriented substantially perpendicular to the edges of the crystal cross-section and to the C-axis direction of propagation to eliminate shear strains.
Averaging in spherically symmetric cosmology
Coley, A. A.; Pelavas, N.
2007-02-15
The averaging problem in cosmology is of fundamental importance. When applied to study cosmological evolution, the theory of macroscopic gravity (MG) can be regarded as a long-distance modification of general relativity. In the MG approach to the averaging problem in cosmology, the Einstein field equations on cosmological scales are modified by appropriate gravitational correlation terms. We study the averaging problem within the class of spherically symmetric cosmological models. That is, we shall take the microscopic equations and effect the averaging procedure to determine the precise form of the correlation tensor in this case. In particular, by working in volume-preserving coordinates, we calculate the form of the correlation tensor under some reasonable assumptions on the form for the inhomogeneous gravitational field and matter distribution. We find that the correlation tensor in a Friedmann-Lemaitre-Robertson-Walker (FLRW) background must be of the form of a spatial curvature. Inhomogeneities and spatial averaging, through this spatial curvature correction term, can have a very significant dynamical effect on the dynamics of the Universe and cosmological observations; in particular, we discuss whether spatial averaging might lead to a more conservative explanation of the observed acceleration of the Universe (without the introduction of exotic dark matter fields). We also find that the correlation tensor for a non-FLRW background can be interpreted as the sum of a spatial curvature and an anisotropic fluid. This may lead to interesting effects of averaging on astrophysical scales. We also discuss the results of averaging an inhomogeneous Lemaitre-Tolman-Bondi solution as well as calculations of linear perturbations (that is, the backreaction) in an FLRW background, which support the main conclusions of the analysis.
Lorcaserin for weight management.
Taylor, James R; Dietrich, Eric; Powell, Jason
2013-01-01
Type 2 diabetes and obesity commonly occur together. Obesity contributes to insulin resistance, a main cause of type 2 diabetes. Modest weight loss reduces glucose, lipids, blood pressure, need for medications, and cardiovascular risk. A number of approaches can be used to achieve weight loss, including lifestyle modification, surgery, and medication. Lorcaserin, a novel antiobesity agent, affects central serotonin subtype 2A receptors, resulting in decreased food intake and increased satiety. It has been studied in obese patients with type 2 diabetes and results in an approximately 5.5 kg weight loss, on average, when used for one year. Headache, back pain, nasopharyngitis, and nausea were the most common adverse effects noted with lorcaserin. Hypoglycemia was more common in the lorcaserin groups in the clinical trials, but none of the episodes were categorized as severe. Based on the results of these studies, lorcaserin was approved at a dose of 10 mg twice daily in patients with a body mass index ≥30 kg/m(2) or ≥27 kg/m(2) with at least one weight-related comorbidity, such as hypertension, type 2 diabetes mellitus, or dyslipidemia, in addition to a reduced calorie diet and increased physical activity. Lorcaserin is effective for weight loss in obese patients with and without type 2 diabetes, although its specific role in the management of obesity is unclear at this time. This paper reviews the clinical trials of lorcaserin, its use from the patient perspective, and its potential role in the treatment of obesity. PMID:23788837
Lorcaserin for weight management
Taylor, James R; Dietrich, Eric; Powell, Jason
2013-01-01
Type 2 diabetes and obesity commonly occur together. Obesity contributes to insulin resistance, a main cause of type 2 diabetes. Modest weight loss reduces glucose, lipids, blood pressure, need for medications, and cardiovascular risk. A number of approaches can be used to achieve weight loss, including lifestyle modification, surgery, and medication. Lorcaserin, a novel antiobesity agent, affects central serotonin subtype 2A receptors, resulting in decreased food intake and increased satiety. It has been studied in obese patients with type 2 diabetes and results in an approximately 5.5 kg weight loss, on average, when used for one year. Headache, back pain, nasopharyngitis, and nausea were the most common adverse effects noted with lorcaserin. Hypoglycemia was more common in the lorcaserin groups in the clinical trials, but none of the episodes were categorized as severe. Based on the results of these studies, lorcaserin was approved at a dose of 10 mg twice daily in patients with a body mass index ≥30 kg/m2 or ≥27 kg/m2 with at least one weight-related comorbidity, such as hypertension, type 2 diabetes mellitus, or dyslipidemia, in addition to a reduced calorie diet and increased physical activity. Lorcaserin is effective for weight loss in obese patients with and without type 2 diabetes, although its specific role in the management of obesity is unclear at this time. This paper reviews the clinical trials of lorcaserin, its use from the patient perspective, and its potential role in the treatment of obesity. PMID:23788837
Application Bayesian Model Averaging method for ensemble system for Poland
NASA Astrophysics Data System (ADS)
Guzikowski, Jakub; Czerwinska, Agnieszka
2014-05-01
The aim of the project is to evaluate methods for generating numerical ensemble weather prediction using a meteorological data from The Weather Research & Forecasting Model and calibrating this data by means of Bayesian Model Averaging (WRF BMA) approach. We are constructing height resolution short range ensemble forecasts using meteorological data (temperature) generated by nine WRF's models. WRF models have 35 vertical levels and 2.5 km x 2.5 km horizontal resolution. The main emphasis is that the used ensemble members has a different parameterization of the physical phenomena occurring in the boundary layer. To calibrate an ensemble forecast we use Bayesian Model Averaging (BMA) approach. The BMA predictive Probability Density Function (PDF) is a weighted average of predictive PDFs associated with each individual ensemble member, with weights that reflect the member's relative skill. For test we chose a case with heat wave and convective weather conditions in Poland area from 23th July to 1st August 2013. From 23th July to 29th July 2013 temperature oscillated below or above 30 Celsius degree in many meteorology stations and new temperature records were added. During this time the growth of the hospitalized patients with cardiovascular system problems was registered. On 29th July 2013 an advection of moist tropical air masses was recorded in the area of Poland causes strong convection event with mesoscale convection system (MCS). MCS caused local flooding, damage to the transport infrastructure, destroyed buildings, trees and injuries and direct threat of life. Comparison of the meteorological data from ensemble system with the data recorded on 74 weather stations localized in Poland is made. We prepare a set of the model - observations pairs. Then, the obtained data from single ensemble members and median from WRF BMA system are evaluated on the basis of the deterministic statistical error Root Mean Square Error (RMSE), Mean Absolute Error (MAE). To evaluation probabilistic data The Brier Score (BS) and Continuous Ranked Probability Score (CRPS) were used. Finally comparison between BMA calibrated data and data from ensemble members will be displayed.
Effect of high-speed jet on flow behavior, retrogradation, and molecular weight of rice starch.
Fu, Zhen; Luo, Shun-Jing; BeMiller, James N; Liu, Wei; Liu, Cheng-Mei
2015-11-20
Effects of high-speed jet (HSJ) treatment on flow behavior, retrogradation, and degradation of the molecular structure of indica rice starch were investigated. Decreasing with the number of HSJ treatment passes were the turbidity of pastes (degree of retrogradation), the enthalpy of melting of retrograded rice starch, weight-average molecular weights and weight-average root-mean square radii of gyration of the starch polysaccharides, and the amylopectin peak areas of SEC profiles. The areas of lower-molecular-weight polymers increased. The chain-length distribution was not significantly changed. Pastes of all starch samples exhibited pseudoplastic, shear-thinning behavior. HSJ treatment increased the flow behavior index and decreased the consistency coefficient and viscosity. The data suggested that degradation of amylopectin was mainly involved and that breakdown preferentially occurred in chains between clusters. PMID:26344255
NASA Technical Reports Server (NTRS)
1995-01-01
The Attitude Adjuster is a system for weight repositioning corresponding to a SCUBA diver's changing positions. Compact tubes on the diver's air tank permit controlled movement of lead balls within the Adjuster, automatically repositioning when the diver changes position. Manufactured by Think Tank Technologies, the system is light and small, reducing drag and energy requirements and contributing to lower air consumption. The Mid-Continent Technology Transfer Center helped the company with both technical and business information and arranged for the testing at Marshall Space Flight Center's Weightlessness Environmental Training Facility for astronauts.
Vibrational averages along thermal lines
NASA Astrophysics Data System (ADS)
Monserrat, Bartomeu
2016-01-01
A method is proposed for the calculation of vibrational quantum and thermal expectation values of physical properties from first principles. Thermal lines are introduced: these are lines in configuration space parametrized by temperature, such that the value of any physical property along them is approximately equal to the vibrational average of that property. The number of sampling points needed to explore the vibrational phase space is reduced by up to an order of magnitude when the full vibrational density is replaced by thermal lines. Calculations of the vibrational averages of several properties and systems are reported, namely, the internal energy and the electronic band gap of diamond and silicon, and the chemical shielding tensor of L-alanine. Thermal lines pave the way for complex calculations of vibrational averages, including large systems and methods beyond semilocal density functional theory.
Evaluations of average level spacings
Liou, H.I.
1980-01-01
The average level spacing for highly excited nuclei is a key parameter in cross section formulas based on statistical nuclear models, and also plays an important role in determining many physics quantities. Various methods to evaluate average level spacings are reviewed. Because of the finite experimental resolution, to detect a complete sequence of levels without mixing other parities is extremely difficult, if not totally impossible. Most methods derive the average level spacings by applying a fit, with different degrees of generality, to the truncated Porter-Thomas distribution for reduced neutron widths. A method that tests both distributions of level widths and positions is discussed extensivey with an example of /sup 168/Er data. 19 figures, 2 tables.
Combining forecast weights: Why and how?
NASA Astrophysics Data System (ADS)
Yin, Yip Chee; Kok-Haur, Ng; Hock-Eam, Lim
2012-09-01
This paper proposes a procedure called forecast weight averaging which is a specific combination of forecast weights obtained from different methods of constructing forecast weights for the purpose of improving the accuracy of pseudo out of sample forecasting. It is found that under certain specified conditions, forecast weight averaging can lower the mean squared forecast error obtained from model averaging. In addition, we show that in a linear and homoskedastic environment, this superior predictive ability of forecast weight averaging holds true irrespective whether the coefficients are tested by t statistic or z statistic provided the significant level is within the 10% range. By theoretical proofs and simulation study, we have shown that model averaging like, variance model averaging, simple model averaging and standard error model averaging, each produces mean squared forecast error larger than that of forecast weight averaging. Finally, this result also holds true marginally when applied to business and economic empirical data sets, Gross Domestic Product (GDP growth rate), Consumer Price Index (CPI) and Average Lending Rate (ALR) of Malaysia.
Modeling operating weight and axle weight distributions for highway vehicles
Greene, D.L.; Liang, J.C.
1988-07-01
The estimation of highway cost responsibility requires detailed information on vehicle operating weights and axle weights by type of vehicle. Typically, 10--20 vehicle types must be cross-classified by 10--20 registered weight classes and again by 20 or more operating weight categories, resulting in 100--400 relative frequencies to be determined for each vehicle type. For each of these, gross operating weight must be distributed to each axle or axle unit. Given the rarity of many of the heaviest vehicle types, direct estimation of these frequencies and axle weights from traffic classification count statistics and truck weight data may exceed the reliability of even the largest (e.g., 250,000 record) data sources. An alternative is to estimate statistical models of operating weight distributions as functions of registered weight, and models of axle weight shares as functions of operating weight. This paper describes the estimation of such functions using the multinomial logit model (a log-linear model) and the implementation of the modeling framework as a PC-based FORTRAN program. Areas for further research include the addition of highway class and region as explanatory variables in operating weight distribution models, and the development of theory for including registration costs and costs of operating overweight in the modeling framework. 14 refs., 45 figs., 5 tabs.
Averaging facial expression over time
Haberman, Jason; Harp, Tom; Whitney, David
2010-01-01
The visual system groups similar features, objects, and motion (e.g., Gestalt grouping). Recent work suggests that the computation underlying perceptual grouping may be one of summary statistical representation. Summary representation occurs for low-level features, such as size, motion, and position, and even for high level stimuli, including faces; for example, observers accurately perceive the average expression in a group of faces (J. Haberman & D. Whitney, 2007, 2009). The purpose of the present experiments was to characterize the time-course of this facial integration mechanism. In a series of three experiments, we measured observers abilities to recognize the average expression of a temporal sequence of distinct faces. Faces were presented in sets of 4, 12, or 20, at temporal frequencies ranging from 1.6 to 21.3 Hz. The results revealed that observers perceived the average expression in a temporal sequence of different faces as precisely as they perceived a single face presented repeatedly. The facial averaging was independent of temporal frequency or set size, but depended on the total duration of exposed faces, with a time constant of ~800 ms. These experiments provide evidence that the visual system is sensitive to the ensemble characteristics of complex objects presented over time. PMID:20053064
Temperature-Averaging Thermal Probe
NASA Technical Reports Server (NTRS)
Kalil, L. F.; Reinhardt, V.
1984-01-01
Temperature-averaging thermal probe measures long-term temperature fluctuations in fluid environment. Consists of temperature probe embedded inside thermally massive material. Probe measurements used to estimate powerplant heating and cooling loads, map temperature profiles, and calibrate more-sensitive temperature probes.
Macleod, Maureen; Craigie, Angela M; Barton, Karen L; Treweek, Shaun; Anderson, Annie S
2013-07-01
Little is known about the response of post-partum women from deprived backgrounds to weight management interventions, however behavioural intervention trials in disadvantaged communities are often characterised by recruitment difficulties. Recruitment and retention is key to the robust conduct of an effective trial, and exploratory work is essential prior to a definitive randomised controlled trial. This paper describes strategies used to recruit to the WeighWell feasibility study, which aimed to recruit 60 overweight or obese post-partum women living in areas of deprivation to a trial of a weight-loss intervention. Recruitment strategies included the following: (1) distribution of posters and 'business cards'; (2) newspaper advertisements; (3) visits to community groups; and (4) personalised letters of invitation sent via the National Health Service (NHS). Potential participants were screened for eligibility following response to a Freephone number. Body mass index was calculated using self-reported body weight and height. Over 6 months, 142 women responded of whom 65 (46%) met the eligibility criteria. The most effective methods for recruiting eligible women and those who went on to complete the study (n = 36) were visits to community groups (37% and 42%, respectively), personalised letters (26% and 17%, respectively) and posters and 'business cards' (22% and 31%, respectively). These results emphasise the need to utilise a range of strategies beyond traditional NHS settings. Current approaches might be enhanced by sending personal contact letters via their General Practitioner to women identified as eligible at post-natal discharge. Under-reporting of body weight by self-report suggests that a threshold lower than 25 kg/m(2) should be utilised for screening purposes. PMID:22284216
Constrained randomization of weighted networks.
Ansmann, Gerrit; Lehnertz, Klaus
2011-08-01
We propose a Markov chain method to efficiently generate surrogate networks that are random under the constraint of given vertex strengths. With these strength-preserving surrogates and with edge-weight-preserving surrogates we investigate the clustering coefficient and the average shortest path length of functional networks of the human brain as well as of the International Trade Networks. We demonstrate that surrogate networks can provide additional information about network-specific characteristics and thus help interpreting empirical weighted networks. PMID:21929060
Annual Average Changes in Adult Obesity as a Risk Factor for Papillary Thyroid Cancer
Hwang, Yunji; Lee, Kyu Eun; Park, Young Joo; Kim, Su-Jin; Kwon, Hyungju; Park, Do Joon; Cho, Belong; Choi, Ho-Chun; Kang, Daehee; Park, Sue K.
2016-01-01
Abstract We evaluated the association between weight change in middle-aged adults and papillary thyroid cancer (PTC) based on a large-scale case-control study. Our study included data from 1551 PTC patients (19.3% men and 80.7% women) who underwent thyroidectomy at the 3 general hospitals in Korea and 15,510 individually matched control subjects. The subjects’ weight history, epidemiologic information, and tumor characteristics confirmed after thyroidectomy were analyzed. Odds ratios (ORs) and 95% confidence intervals (95% CIs) were determined for the annual average changes in weight and obesity indicators (body mass index (BMI), body surface area, and body fat percentage (BF%) in subjects since the age of 35 years. Subjects with a total weight gain ≥10 kg after age 35 years were more likely to have PTC (men, OR, 5.39, 95% CI, 3.88–7.49; women, OR, 3.36, 95% CI, 2.87–3.93) compared with subjects with a stable weight (loss or gain <5 kg). A marked increase in BMI since age 35 years (annual average change of BMI ≥0.3 kg/m2/yr) was related to an elevated PTC risk, and the association was more pronounced for large-sized PTC risks (<1 cm, OR, 2.34, 95% CI, 1.92–2.85; ≥1 cm, OR, 4.00, 95% CI, 2.91–5.49, P heterogeneity = 0.005) compared with low PTC risks. Weight gain and annual increases in obesity indicators in middle-aged adults may increase the risk of developing PTC. PMID:26945379
Ultrahigh molecular weight aromatic siloxane polymers
NASA Technical Reports Server (NTRS)
Ludwick, L. M.
1982-01-01
The condensation of a diol with a silane in toluene yields a silphenylene-siloxane polymer. The reaction of stiochiometric amounts of the diol and silane produced products with molecular weights in the range 2.0 - 6.0 x 10 to the 5th power. The molecular weight of the product was greatly increased by a multistep technique. The methodology for synthesis of high molecular weight polymers using a two step procedure was refined. Polymers with weight average molecular weights in excess of 1.0 x 10 to the 6th power produced by this method. Two more reactive silanes, bis(pyrrolidinyl)dimethylsilane and bis(gamma butyrolactam)dimethylsilane, are compared with the dimethyleminodimethylsilane in ability to advance the molecular weight of the prepolymer. The polymers produced are characterized by intrinsic viscosity in tetrahydrofuran. Weight and number average molecular weights and polydispersity are determined by gel permeation chromatography.
Unexplained weight loss is a decrease in body weight, when you did not try to lose the weight on your own. Many people gain and lose weight. Unintentional weight loss is loss of 10 pounds OR 5% of ...
... loss-rapid weight loss; Overweight-rapid weight loss; Obesity-rapid weight loss; Diet-rapid weight loss ... for people who have health problems because of obesity. For these people, losing a lot of weight ...
Cosmological measures without volume weighting
Page, Don N
2008-10-15
Many cosmologists (myself included) have advocated volume weighting for the cosmological measure problem, weighting spatial hypersurfaces by their volume. However, this often leads to the Boltzmann brain problem, that almost all observations would be by momentary Boltzmann brains that arise very briefly as quantum fluctuations in the late universe when it has expanded to a huge size, so that our observations (too ordered for Boltzmann brains) would be highly atypical and unlikely. Here it is suggested that volume weighting may be a mistake. Volume averaging is advocated as an alternative. One consequence may be a loss of the argument that eternal inflation gives a nonzero probability that our universe now has infinite volume.
Averaging Robertson-Walker cosmologies
Brown, Iain A.; Robbers, Georg; Behrend, Juliane E-mail: G.Robbers@thphys.uni-heidelberg.de
2009-04-15
The cosmological backreaction arises when one directly averages the Einstein equations to recover an effective Robertson-Walker cosmology, rather than assuming a background a priori. While usually discussed in the context of dark energy, strictly speaking any cosmological model should be recovered from such a procedure. We apply the scalar spatial averaging formalism for the first time to linear Robertson-Walker universes containing matter, radiation and dark energy. The formalism employed is general and incorporates systems of multiple fluids with ease, allowing us to consider quantitatively the universe from deep radiation domination up to the present day in a natural, unified manner. Employing modified Boltzmann codes we evaluate numerically the discrepancies between the assumed and the averaged behaviour arising from the quadratic terms, finding the largest deviations for an Einstein-de Sitter universe, increasing rapidly with Hubble rate to a 0.01% effect for h = 0.701. For the {Lambda}CDM concordance model, the backreaction is of the order of {Omega}{sub eff}{sup 0} Almost-Equal-To 4 Multiplication-Sign 10{sup -6}, with those for dark energy models being within a factor of two or three. The impacts at recombination are of the order of 10{sup -8} and those in deep radiation domination asymptote to a constant value. While the effective equations of state of the backreactions in Einstein-de Sitter, concordance and quintessence models are generally dust-like, a backreaction with an equation of state w{sub eff} < -1/3 can be found for strongly phantom models.
Ensemble averaging of acoustic data
NASA Technical Reports Server (NTRS)
Stefanski, P. K.
1982-01-01
A computer program called Ensemble Averaging of Acoustic Data is documented. The program samples analog data, analyzes the data, and displays them in the time and frequency domains. Hard copies of the displays are the program's output. The documentation includes a description of the program and detailed user instructions for the program. This software was developed for use on the Ames 40- by 80-Foot Wind Tunnel's Dynamic Analysis System consisting of a PDP-11/45 computer, two RK05 disk drives, a tektronix 611 keyboard/display terminal, and FPE-4 Fourier Processing Element, and an analog-to-digital converter.
Weight loss attempts in adults: goals, duration, and rate of weight loss.
Williamson, D F; Serdula, M K; Anda, R F; Levy, A; Byers, T
1992-01-01
OBJECTIVES: Although attempted weight loss is common, little is known about the goals and durations of weight loss attempts and the rates of achieved weight loss in the general population. METHODS. Data were collected by telephone in 1989 from adults aged 18 years and older in 39 states and the District of Columbia. Analyses were carried out separately for the 6758 men and 14,915 women who reported currently trying to lose weight. RESULTS. Approximately 25% of the men respondents and 40% of the women respondents reported that they were currently trying to lose weight. Among men, a higher percentage of Hispanics (31%) than of Whites (25%) or Blacks (23%) reported trying to lose weight. Among women, however, there were no ethnic differences in prevalence. The average man wanted to lose 30 pounds and to weigh 178 pounds; the average woman wanted to lose 31 pounds and to weigh 133 pounds. Black women wanted to lose an average of 8 pounds more than did White women, but Black women's goal weight was 10 pounds heavier. The average rate of achieved weight loss was 1.4 pounds per week for men and 1.1 pounds per week for women; these averages, however, may reflect only the experience of those most successful at losing weight. CONCLUSIONS. Attempted weight loss is a common behavior, regardless of age, gender, or ethnicity, and weight loss goals are substantial; however, obesity remains a major public health problem in the United States. PMID:1503167
Achronal averaged null energy condition
Graham, Noah; Olum, Ken D.
2007-09-15
The averaged null energy condition (ANEC) requires that the integral over a complete null geodesic of the stress-energy tensor projected onto the geodesic tangent vector is never negative. This condition is sufficient to prove many important theorems in general relativity, but it is violated by quantum fields in curved spacetime. However there is a weaker condition, which is free of known violations, requiring only that there is no self-consistent spacetime in semiclassical gravity in which ANEC is violated on a complete, achronal null geodesic. We indicate why such a condition might be expected to hold and show that it is sufficient to rule out closed timelike curves and wormholes connecting different asymptotically flat regions.
How global is the global effect? The spatial characteristics of saccade averaging.
Van der Stigchel, S; Nijboer, T C W
2013-05-24
When a target and a distractor are presented in close proximity, an eye movement will generally land in between these two elements. This is known as the 'global effect' and has been claimed to be a reflection of the averaged saccade programs towards both locations. The aim of the present study was to systematically investigate whether there is only a limited area in the saccade map in which saccade averaging occurs. To this end, we examined various distances between target and distractor in two experiments and investigated whether the majority of eye movements landed in between the target and the distractor. Results indicated that the endpoint distribution was unimodal for distances up to 35 (in polar coordinates), with saccades generally landing in between the target and the distractor. When the distance was higher than 45, the saccade endpoint distribution was predominantly bimodal, with saccades landing either on the target or on the distractor. The decrease in saccade averaging was linear until almost no averaging saccades were observed for the longest distances. As saccades landing in between target and distractor reflect a weak, or absent, top-down signal, the present study indicated that top-down information is unable to strongly influence the oculomotor system when target and distractor are presented in close proximity. In this situation, the resulting eye movement is determined by the weighted average of saccade vectors present in a restricted region in the motor map. PMID:23523571
Flexible time domain averaging technique
NASA Astrophysics Data System (ADS)
Zhao, Ming; Lin, Jing; Lei, Yaguo; Wang, Xiufeng
2013-09-01
Time domain averaging(TDA) is essentially a comb filter, it cannot extract the specified harmonics which may be caused by some faults, such as gear eccentric. Meanwhile, TDA always suffers from period cutting error(PCE) to different extent. Several improved TDA methods have been proposed, however they cannot completely eliminate the waveform reconstruction error caused by PCE. In order to overcome the shortcomings of conventional methods, a flexible time domain averaging(FTDA) technique is established, which adapts to the analyzed signal through adjusting each harmonic of the comb filter. In this technique, the explicit form of FTDA is first constructed by frequency domain sampling. Subsequently, chirp Z-transform(CZT) is employed in the algorithm of FTDA, which can improve the calculating efficiency significantly. Since the signal is reconstructed in the continuous time domain, there is no PCE in the FTDA. To validate the effectiveness of FTDA in the signal de-noising, interpolation and harmonic reconstruction, a simulated multi-components periodic signal that corrupted by noise is processed by FTDA. The simulation results show that the FTDA is capable of recovering the periodic components from the background noise effectively. Moreover, it can improve the signal-to-noise ratio by 7.9 dB compared with conventional ones. Experiments are also carried out on gearbox test rigs with chipped tooth and eccentricity gear, respectively. It is shown that the FTDA can identify the direction and severity of the eccentricity gear, and further enhances the amplitudes of impulses by 35%. The proposed technique not only solves the problem of PCE, but also provides a useful tool for the fault symptom extraction of rotating machinery.
A pure bending exact nodal-averaged shear strain method for finite element plate analysis
NASA Astrophysics Data System (ADS)
Wu, C. T.; Guo, Y.; Wang, D.
2014-05-01
An averaged shear strain method, based on a nodal integration approach, is presented for the finite element analysis of Reissner-Mindlin plates. In this work, we combine the shear interpolation method from the MITC4 plate element with an area-weighted averaging technique for the nodal integration of shear energy to relieve shear locking in the thin plate analysis as well as to pass the pure bending patch test. In order to resolve the numerical instability caused by the direct nodal integration, the bending strain field is computed by a sub-domain nodal integration approach based on the Sub-domain Stabilized Conforming Integration and a modified curvature smoothing scheme. The resulting nodally integrated smoothed strain formulation is shown to contain only the primitive variables and thus can be easily implemented in the existing displacement-based finite element plate formulation. Several numerical examples are presented to demonstrate the accuracy of the present method.
Weight-ing: the experience of waiting on weight loss.
Glenn, Nicole M
2013-03-01
Perhaps we want to be perfect, strive for health, beauty, and the admiring gaze of others. Maybe we desire the body of our youth, the "healthy" body, the body that has just the right fit. Regardless of the motivation, we might find ourselves striving, wanting, and waiting on weight loss. What is it to wait on weight loss? I explore the meaning of this experience-as-lived using van Manen's guide to phenomenological reflection and writing. Weight has become an increasing focus of contemporary culture, demonstrated, for example, by a growing weight-loss industry and global obesity "epidemic." Weight has become synonymous with health status, and weight loss with "healthier." I examine the weight wait through experiences of the common and uncommon, considering relations to time, body, space, and the other with the aim of evoking a felt, embodied, emotive understanding of the meaning of waiting on weight loss. I also discuss the implications of the findings. PMID:23202478
Non-Homogeneous Fractal Hierarchical Weighted Networks
Dong, Yujuan; Dai, Meifeng; Ye, Dandan
2015-01-01
A model of fractal hierarchical structures that share the property of non-homogeneous weighted networks is introduced. These networks can be completely and analytically characterized in terms of the involved parameters, i.e., the size of the original graph Nk and the non-homogeneous weight scaling factors r1, r2, rM. We also study the average weighted shortest path (AWSP), the average degree and the average node strength, taking place on the non-homogeneous hierarchical weighted networks. Moreover the AWSP is scrupulously calculated. We show that the AWSP depends on the number of copies and the sum of all non-homogeneous weight scaling factors in the infinite network order limit. PMID:25849619
College Freshman Stress and Weight Change: Differences by Gender
ERIC Educational Resources Information Center
Economos, Christina D.; Hildebrandt, M. Lise; Hyatt, Raymond R.
2008-01-01
Objectives: To examine how stress and health-related behaviors affect freshman weight change by gender. Methods: Three hundred ninety-six freshmen completed a 40-item health behavior survey and height and weight were collected at baseline and follow-up. Results: Average weight change was 5.04 lbs for males, 5.49 lbs for females. Weight gain was…
College Freshman Stress and Weight Change: Differences by Gender
ERIC Educational Resources Information Center
Economos, Christina D.; Hildebrandt, M. Lise; Hyatt, Raymond R.
2008-01-01
Objectives: To examine how stress and health-related behaviors affect freshman weight change by gender. Methods: Three hundred ninety-six freshmen completed a 40-item health behavior survey and height and weight were collected at baseline and follow-up. Results: Average weight change was 5.04 lbs for males, 5.49 lbs for females. Weight gain was
The Case for a National Standard of Grade Weighting.
ERIC Educational Resources Information Center
Talley, Norma R.; Mohr, Joan Isaac
1993-01-01
Surveyed high schools as to their models and practices of grade weighting, sought opinions from college respondents who participated in past national surveys on weighting averages, and determined feasibility of recommending national grade weighting system. Found that weighted grading system is preferred by college admission professionals. (NB)
40 CFR 63.5710 - How do I demonstrate compliance using emissions averaging?
Code of Federal Regulations, 2013 CFR
2013-07-01
... CATEGORIES (CONTINUED) National Emission Standards for Hazardous Air Pollutants for Boat Manufacturing... the average.) ER22AU01.012 Where: HAP emissions= Organic HAP emissions calculated using MACT model point values for each operation included in the average, kilograms. PVR= Weighted-average MACT...
40 CFR 63.5710 - How do I demonstrate compliance using emissions averaging?
Code of Federal Regulations, 2012 CFR
2012-07-01
... CATEGORIES (CONTINUED) National Emission Standards for Hazardous Air Pollutants for Boat Manufacturing... the average.) ER22AU01.012 Where: HAP emissions= Organic HAP emissions calculated using MACT model point values for each operation included in the average, kilograms. PVR= Weighted-average MACT...
40 CFR 63.5710 - How do I demonstrate compliance using emissions averaging?
Code of Federal Regulations, 2014 CFR
2014-07-01
... CATEGORIES (CONTINUED) National Emission Standards for Hazardous Air Pollutants for Boat Manufacturing... the average.) ER22AU01.012 Where: HAP emissions= Organic HAP emissions calculated using MACT model point values for each operation included in the average, kilograms. PVR = Weighted-average MACT...
Comparative effectiveness of accommodating and weight resistance training modes.
O'Hagan, F T; Sale, D G; MacDougall, J D; Garner, S H
1995-08-01
Six men and six women trained the elbow flexors of both arms 3 d.wk-1 for 20 wk. In each training session, one arm did 3-5 sets of 10 maximal concentric actions on an accommodating resistance device (ARD), the other arm 3-5 sets of 8-12 coupled eccentric/concentric actions (repetitions) to volitional failure (8-12 RM) on a weight resistance device (WRD). The average "intensity" (force of concentric actions) was approximately 1.25 times greater in ARD training, the average "volume" (number of actions x force of actions) 1.6 times greater in WRD training, and the time required to complete a training session the same for each. Both types of training produced significant increases in a single maximum weight lift (1 RM on the WRD), in the peak force of a single maximal concentric action measured on the ARD and an isovelocity dynamometer, and in biceps, brachialis, and total elbow flexor cross-sectional area (CSA). Biceps Type I and II fiber area did not change significantly. WRD training produced greater increases than ARD training in the 1 RM test on the WRD and in brachialis CSA. The data indicate that both of these common training regimens effectively increase strength and muscle mass, but the weight training regimen may be more effective for increasing muscle mass. PMID:7476067
Effect of clothing weight on body weight
Technology Transfer Automated Retrieval System (TEKTRAN)
Background: In clinical settings, it is common to measure weight of clothed patients and estimate a correction for the weight of clothing, but we can find no papers in the medical literature regarding the variability in clothing weight with weather, season, and gender. Methods: Fifty adults (35 wom...
Model Averaging for Improving Inference from Causal Diagrams.
Hamra, Ghassan B; Kaufman, Jay S; Vahratian, Anjel
2015-08-01
Model selection is an integral, yet contentious, component of epidemiologic research. Unfortunately, there remains no consensus on how to identify a single, best model among multiple candidate models. Researchers may be prone to selecting the model that best supports their a priori, preferred result; a phenomenon referred to as "wish bias". Directed acyclic graphs (DAGs), based on background causal and substantive knowledge, are a useful tool for specifying a subset of adjustment variables to obtain a causal effect estimate. In many cases, however, a DAG will support multiple, sufficient or minimally-sufficient adjustment sets. Even though all of these may theoretically produce unbiased effect estimates they may, in practice, yield somewhat distinct values, and the need to select between these models once again makes the research enterprise vulnerable to wish bias. In this work, we suggest combining adjustment sets with model averaging techniques to obtain causal estimates based on multiple, theoretically-unbiased models. We use three techniques for averaging the results among multiple candidate models: information criteria weighting, inverse variance weighting, and bootstrapping. We illustrate these approaches with an example from the Pregnancy, Infection, and Nutrition (PIN) study. We show that each averaging technique returns similar, model averaged causal estimates. An a priori strategy of model averaging provides a means of integrating uncertainty in selection among candidate, causal models, while also avoiding the temptation to report the most attractive estimate from a suite of equally valid alternatives. PMID:26270672
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Averaging. 91.1304 Section 91.1304... Averaging. (a) A manufacturer may use averaging across engine families to demonstrate a zero or positive credit balance for a model year. Positive credits to be used in averaging may be obtained from...
Code of Federal Regulations, 2013 CFR
2013-07-01
... 40 Protection of Environment 21 2013-07-01 2013-07-01 false Averaging. 91.1304 Section 91.1304... Averaging. (a) A manufacturer may use averaging across engine families to demonstrate a zero or positive credit balance for a model year. Positive credits to be used in averaging may be obtained from...
Code of Federal Regulations, 2012 CFR
2012-07-01
... 40 Protection of Environment 21 2012-07-01 2012-07-01 false Averaging. 91.1304 Section 91.1304... Averaging. (a) A manufacturer may use averaging across engine families to demonstrate a zero or positive credit balance for a model year. Positive credits to be used in averaging may be obtained from...
Informed Test Component Weighting.
ERIC Educational Resources Information Center
Rudner, Lawrence M.
2001-01-01
Identifies and evaluates alternative methods for weighting tests. Presents formulas for composite reliability and validity as a function of component weights and suggests a rational process that identifies and considers trade-offs in determining weights. Discusses drawbacks to implicit weighting and explicit weighting and the difficulty of…
Programmable noise bandwidth reduction by means of digital averaging
NASA Technical Reports Server (NTRS)
Poklemba, John J. (Inventor)
1993-01-01
Predetection noise bandwidth reduction is effected by a pre-averager capable of digitally averaging the samples of an input data signal over two or more symbols, the averaging interval being defined by the input sampling rate divided by the output sampling rate. As the averaged sample is clocked to a suitable detector at a much slower rate than the input signal sampling rate the noise bandwidth at the input to the detector is reduced, the input to the detector having an improved signal to noise ratio as a result of the averaging process, and the rate at which such subsequent processing must operate is correspondingly reduced. The pre-averager forms a data filter having an output sampling rate of one sample per symbol of received data. More specifically, selected ones of a plurality of samples accumulated over two or more symbol intervals are output in response to clock signals at a rate of one sample per symbol interval. The pre-averager includes circuitry for weighting digitized signal samples using stored finite impulse response (FIR) filter coefficients. A method according to the present invention is also disclosed.
5 CFR 591.210 - What are weights?
Code of Federal Regulations, 2011 CFR
2011-01-01
... What are weights? (a) A weight is the relative importance or share of a subpart of a group compared... compared with the whole pie. (b) OPM uses two kinds of weights: Consumer expenditure weights and employment.... The employment weight is the relative employment population of the survey area compared with...
Weight Management and Calories
... Food Safety Newsroom Dietary Guidelines Communicator’s Guide Weight Management You are here Home / ADULTS Weight Management Print Share Why is weight management important? In addition to helping you feel and ...
... I Help a Friend Who Cuts? Weight and Diabetes KidsHealth > For Teens > Weight and Diabetes Print A ... person's diabetes management plan. Weight and Type 1 Diabetes If a person has type 1 diabetes but ...
Weight loss surgery helps people with extreme obesity to lose weight. It may be an option if you ... caused by obesity. There are different types of weight loss surgery. They often limit the amount of food ...
... Can I Help a Friend Who Cuts? Weight Loss Surgery KidsHealth > For Teens > Weight Loss Surgery Print ... Any Risks and Side Effects? What Is Weight Loss Surgery? For some people, being overweight is about ...
Jupiter's Radio Rotation Period: A 50-year Average
NASA Astrophysics Data System (ADS)
Higgins, C. A.; Reyes, F.; Solus, D.
2011-12-01
Using 50 years of continuous seasonal observations of Jupiter's decametric radio emissions from 18-22 MHz collected at the University of Florida Radio Observatory (UFRO), we calculate a new radio rotation period of Jupiter. The new period is the weighted mean of more than 20 independent measurements. Each measurement is found by determining the drift of the histograms of probability of occurrence versus the System III (1965) central meridian longitude (CML) over intervals of approximately 12, 24, 36, and 48 years. This multiple 12-year average technique is employed to reduce the uncertainty in the longitudes of the radio sources caused by Jupiter's 11.86 year orbit. Our weighted mean is 9 h 55 m 29.689 s, with a standard deviation of the weighted mean of 0.004 s. Our calculations show remarkably stable radio sources. An upper limit of any radio rotation period drift is discussed.
Effect of molecular weight on polyphenylquinoxaline properties
NASA Technical Reports Server (NTRS)
Jensen, Brian J.
1991-01-01
A series of polyphenyl quinoxalines with different molecular weight and end-groups were prepared by varying monomer stoichiometry. Thus, 4,4'-oxydibenzil and 3,3'-diaminobenzidine were reacted in a 50/50 mixture of m-cresol and xylenes. Reaction concentration, temperature, and stir rate were studied and found to have an effect on polymer properties. Number and weight average molecular weights were determined and correlated well with viscosity data. Glass transition temperatures were determined and found to vary with molecular weight and end-groups. Mechanical properties of films from polymers with different molecular weights were essentially identical at room temperature but showed significant differences at 232 C. Diamine terminated polymers were found to be much less thermooxidatively stable than benzil terminated polymers when aged at 316 C even though dynamic thermogravimetric analysis revealed only slight differences. Lower molecular weight polymers exhibited better processability than higher molecular weight polymers.
The uncertainty of simple spatial averages using rain gauge networks
Morrissey, M.L.; Maliekal, J.A.; Greene, J.S.
1995-08-01
A technique is presented whereby the standard error of a simple spatial average can be computed given any spatial configuration of measuring sites. The method should improve useful to those wishing to assess the time-averaged performance of a network of rain gauges where the site locations are fixed within the averaging area. It is demonstrated that in addition to the site density and the spatial structure of the measured variable, the spatial geometry of the network with respect to the averaging area must be accounted for in the standard error expression. By incorporating estimates of the long-term population mean in the method, bias and random error in the sample statistics are minimized. 18 refs., 9 figs.
ERIC Educational Resources Information Center
Molfese, Dennis L.; Key, Alexandra Fonaryova; Kelly, Spencer; Cunningham, Natalie; Terrell, Shona; Ferguson, Melissa; Molfese, Victoria J.; Bonebright, Terri
2006-01-01
Event-related potentials (ERPs) were recorded from 27 children (14 girls, 13 boys) who varied in their reading skill levels. Both behavior performance measures recorded during the ERP word classification task and the ERP responses themselves discriminated between children with above-average, average, and below-average reading skills. ERP
Do Diurnal Aerosol Changes Affect Daily Average Radiative Forcing?
Kassianov, Evgueni I.; Barnard, James C.; Pekour, Mikhail S.; Berg, Larry K.; Michalsky, Joseph J.; Lantz, K.; Hodges, G. B.
2013-06-17
Strong diurnal variability of aerosol has been observed frequently for many urban/industrial regions. How this variability may alter the direct aerosol radiative forcing (DARF), however, is largely unknown. To quantify changes in the time-averaged DARF, we perform an assessment of 29 days of high temporal resolution ground-based data collected during the Two-Column Aerosol Project (TCAP) on Cape Cod, which is downwind of metropolitan areas. We demonstrate that strong diurnal changes of aerosol loading (about 20% on average) have a negligible impact on the 24-h average DARF, when daily averaged optical properties are used to find this quantity. However, when there is a sparse temporal sampling of aerosol properties, which may preclude the calculation of daily averaged optical properties, large errors (up to 100%) in the computed DARF may occur. We describe a simple way of reducing these errors, which suggests the minimal temporal sampling needed to accurately find the forcing.
Do diurnal aerosol changes affect daily average radiative forcing?
NASA Astrophysics Data System (ADS)
Kassianov, Evgueni; Barnard, James; Pekour, Mikhail; Berg, Larry K.; Michalsky, Joseph; Lantz, Kathy; Hodges, Gary
2013-06-01
diurnal variability of aerosol has been observed frequently for many urban/industrial regions. How this variability may alter the direct aerosol radiative forcing (DARF), however, is largely unknown. To quantify changes in the time-averaged DARF, we perform an assessment of 29 days of high temporal resolution ground-based data collected during the Two-Column Aerosol Project on Cape Cod, which is downwind of metropolitan areas. We demonstrate that strong diurnal changes of aerosol loading (about 20% on average) have a negligible impact on the 24-h average DARF when daily averaged optical properties are used to find this quantity. However, when there is a sparse temporal sampling of aerosol properties, which may preclude the calculation of daily averaged optical properties, large errors (up to 100%) in the computed DARF may occur. We describe a simple way of reducing these errors, which suggests the minimal temporal sampling needed to accurately find the forcing.
Body Weight Relationships in Early Marriage: Weight Relevance, Weight Comparisons, and Weight Talk
Bove, Caron F.; Sobal, Jeffery
2011-01-01
This investigation uncovered processes underlying the dynamics of body weight and body image among individuals involved in nascent heterosexual marital relationships in Upstate New York. In-depth, semi-structured qualitative interviews conducted with 34 informants, 20 women and 14 men, just prior to marriage and again one year later were used to explore continuity and change in cognitive, affective, and behavioral factors relating to body weight and body image at the time of marriage, an important transition in the life course. Three major conceptual themes operated in the process of developing and enacting informants’ body weight relationships with their partner: weight relevance, weight comparisons, and weight talk. Weight relevance encompassed the changing significance of weight during early marriage and included attracting and capturing a mate, relaxing about weight, living healthily, and concentrating on weight. Weight comparisons between partners involved weight relativism, weight competition, weight envy, and weight role models. Weight talk employed pragmatic talk, active and passive reassurance, and complaining and critiquing criticism. Concepts emerging from this investigation may be useful in designing future studies of and approaches to managing body weight in adulthood. PMID:21864601
Mechanisms of Weight Regain following Weight Loss
Blomain, Erik Scott; Dirhan, Dara Anne; Valentino, Michael Anthony; Kim, Gilbert Won; Waldman, Scott Arthur
2013-01-01
Obesity is a world-wide pandemic and its incidence is on the rise along with associated comorbidities. Currently, there are few effective therapies to combat obesity. The use of lifestyle modification therapy, namely, improvements in diet and exercise, is preferable over bariatric surgery or pharmacotherapy due to surgical risks and issues with drug efficacy and safety. Although they are initially successful in producing weight loss, such lifestyle intervention strategies are generally unsuccessful in achieving long-term weight maintenance, with the vast majority of obese patients regaining their lost weight during followup. Recently, various compensatory mechanisms have been elucidated by which the body may oppose new weight loss, and this compensation may result in weight regain back to the obese baseline. The present review summarizes the available evidence on these compensatory mechanisms, with a focus on weight loss-induced changes in energy expenditure, neuroendocrine pathways, nutrient metabolism, and gut physiology. These findings have added a major focus to the field of antiobesity research. In addition to investigating pathways that induce weight loss, the present work also focuses on pathways that may instead prevent weight regain. Such strategies will be necessary for improving long-term weight loss maintenance and outcomes for patients who struggle with obesity. PMID:24533218
Demonstration of a Model Averaging Capability in FRAMES
NASA Astrophysics Data System (ADS)
Meyer, P. D.; Castleton, K. J.
2009-12-01
Uncertainty in model structure can be incorporated in risk assessment using multiple alternative models and model averaging. To facilitate application of this approach to regulatory applications based on risk or dose assessment, a model averaging capability was integrated with the Framework for Risk Analysis in Multimedia Environmental Systems (FRAMES) version 2 software. FRAMES is a software platform that allows the non-parochial communication between disparate models, databases, and other frameworks. Users have the ability to implement and select environmental models for specific risk assessment and management problems. Standards are implemented so that models produce information that is readable by other downstream models and accept information from upstream models. Models can be linked across multiple media and from source terms to quantitative risk/dose estimates. Parameter sensitivity and uncertainty analysis tools are integrated. A model averaging module was implemented to accept output from multiple models and produce average results. These results can be deterministic quantities or probability distributions obtained from an analysis of parameter uncertainty. Output from alternative models is averaged using weights determined from user input and/or model calibration results. A model calibration module based on the PEST code was implemented to provide FRAMES with a general calibration capability. An application illustrates the implementation, user interfaces, execution, and results of the FRAMES model averaging capabilities.
40 CFR 76.11 - Emissions averaging.
Code of Federal Regulations, 2014 CFR
2014-07-01
...) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General... averaging plan is in compliance with the Acid Rain emission limitation for NOX under the plan only if...
40 CFR 76.11 - Emissions averaging.
Code of Federal Regulations, 2013 CFR
2013-07-01
...) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General... averaging plan is in compliance with the Acid Rain emission limitation for NOX under the plan only if...
RHIC BPM system average orbit calculations
Michnoff,R.; Cerniglia, P.; Degen, C.; Hulsart, R.; et al.
2009-05-04
RHIC beam position monitor (BPM) system average orbit was originally calculated by averaging positions of 10000 consecutive turns for a single selected bunch. Known perturbations in RHIC particle trajectories, with multiple frequencies around 10 Hz, contribute to observed average orbit fluctuations. In 2006, the number of turns for average orbit calculations was made programmable; this was used to explore averaging over single periods near 10 Hz. Although this has provided an average orbit signal quality improvement, an average over many periods would further improve the accuracy of the measured closed orbit. A new continuous average orbit calculation was developed just prior to the 2009 RHIC run and was made operational in March 2009. This paper discusses the new algorithm and performance with beam.
Dynamic averaging and foraging decisions in horses (Equus callabus).
Devenport, Jill A; Patterson, Megan R; Devenport, Lynn D
2005-08-01
The variability of most environments taxes foraging decisions by increasing the uncertainty of the information available. One solution to the problem is to use dynamic averaging, as do some granivores and carnivores. Arguably, the same strategy could be useful for grazing herbivores, even though their food renews and is more homogeneously distributed. Horses (Equus callabus) were given choices between variable patches after short or long delays. When patch information was current, horses returned to the patch that was recently best, whereas those without current information matched choices to the long-term average values of the patches. These results demonstrate that a grazing species uses dynamic averaging and indicate that, like granivores and carnivores, they can use temporal weighting to optimize foraging decisions. PMID:16131264
Averaging in LRS class II spacetimes
NASA Astrophysics Data System (ADS)
Kapar, Petr; Svtek, Otakar
2015-02-01
We generalize Buchert's averaged equations (Gen Relativ Gravit 32; 105, 2000; Gen Relativ Gravit 33; 1381, 2001) to LRS class II dust model in the sense that all Einstein equations are averaged, not only the trace part. We derive the relevant averaged equations and we investigate backreaction on expansion and shear scalars in an approximate LTB model. Finally we propose a way to close the system of averaged equations.
Averaging and Adding in Children's Worth Judgements
ERIC Educational Resources Information Center
Schlottmann, Anne; Harman, Rachel M.; Paine, Julie
2012-01-01
Under the normative Expected Value (EV) model, multiple outcomes are additive, but in everyday worth judgement intuitive averaging prevails. Young children also use averaging in EV judgements, leading to a disordinal, crossover violation of utility when children average the part worths of simple gambles involving independent events (Schlottmann,…
Designing Digital Control Systems With Averaged Measurements
NASA Technical Reports Server (NTRS)
Polites, Michael E.; Beale, Guy O.
1990-01-01
Rational criteria represent improvement over "cut-and-try" approach. Recent development in theory of control systems yields improvements in mathematical modeling and design of digital feedback controllers using time-averaged measurements. By using one of new formulations for systems with time-averaged measurements, designer takes averaging effect into account when modeling plant, eliminating need to iterate design and simulation phases.
NASA Astrophysics Data System (ADS)
Ding, F.; Theobald, M.; Vollmer, B.; Savtchenko, A. K.; Hearty, T. J.; Esfandiari, A. E.
2012-12-01
Producing timely and accurate water forecast and information is the mission of National Weather Service River Forecast Centers (NWS RFCs) of National Oceanic and Atmospheric Administration (NOAA). The river forecast system in RFCs requires average surface temperature in the fixed 6-hour period 000-0600, 0600-1200, 1200-1800, and 1200-0000 UTC. The current logic of RFC temperature forecast relies on ingest of point values of daytime maximum and nighttime minimum temperature. Meanwhile, the mean temperature for the 6-hour period is estimated from a weighted average of daytime maximum and nighttime minimum temperature. The Atmospheric Infrared Sounder (AIRS) in the first high spectral resolution infrared sounder on board the Aqua satellite which was launched in May 2002 and follows a Sun-synchronous polar orbit. It is aimed to produce high resolution atmospheric profile and surface atmospheric parameters. As Aqua crosses the equator at about 1330 and 0130 local time, the AIRS retrieved surface temperature may represent daytime maximum and nighttime minimum value. Comparing to point observation from surface weather stations which are often sparse over the less-populated area and are unevenly distributed, satellite may obtain better area averaged observation. This test study assesses the potential of using AIRS retrieved surface temperature to forecast 6-hour average temperature for NWS RFCs. The California Nevada RFC is selected due to the poor coverage of surface observation in the mountainous region and spring snow melting. The study focuses on the March to May spring season when water from snowpack melting often plays important role in flood. AIRS retrieved temperature and surface weather station data set will be used to derive statistical weighting coefficient for 6-hour average temperature forecast. The resulting forecast biases and errors will be the main indicators of the potential usage. All study results will be presented in the meeting.
ERIC Educational Resources Information Center
Ryan, Kevin Michael
2011-01-01
Research on syllable weight in generative phonology has focused almost exclusively on systems in which weight is treated as an ordinal hierarchy of clearly delineated categories (e.g. light and heavy). As I discuss, canonical weight-sensitive phenomena in phonology, including quantitative meter and quantity-sensitive stress, can also treat weight
... Weight Body Mass Index (BMI) About Adult BMI Adult BMI Calculator Metric Version Â About Child & Teen BMI Measuring Children's Height and Weight Accurately At Home Child & Teen BMI Calculator Children's BMI Tool for Schools Finding a Balance Other Factors in Weight Gain Preventing Weight Gain ...
... Weight Body Mass Index (BMI) About Adult BMI Adult BMI Calculator Metric Version Â About Child & Teen BMI Measuring Children's Height and Weight Accurately At Home Child & Teen BMI Calculator Children's BMI Tool for Schools Finding a Balance Other Factors in Weight Gain Preventing Weight Gain ...
Fact Sheet Proven Weight Loss Methods What can weight loss do for you? Losing weight can improve your health in a number of ways. ... limiting calories) usually isn’t enough to cause weight loss. But exercise plays an important part in helping ...
PREVENTING WEIGHT REGAIN AFTER WEIGHT LOSS
Technology Transfer Automated Retrieval System (TEKTRAN)
For most dieters, a regaining of lost weight is an all too common experience. Indeed, virtually all interventions for weight loss show limited or even poor long-term effectiveness. This sobering reality was reflected in a comprehensive review of nonsurgical treatments of obesity conducted by the Ins...
40 CFR 86.129-80 - Road load power, test weight, and inertia weight class determination.
Code of Federal Regulations, 2014 CFR
2014-07-01
... load setting is determined from the equivalent test weight, the reference frontal area, the body shape... = vehicle equivalent test weight (lbs) from the table in paragraph (a). a = 0.43 for fastback-shaped... 40 Protection of Environment 19 2014-07-01 2014-07-01 false Road load power, test weight,...
The experimental evidence for weight-loss treatment of essential hypertension: a critical review.
Hovell, M F
1982-01-01
The empirical evidence concerning the therapeutic effects of weight loss for hypertension treatment was reviewed. Interventions were critically reviewed for strength of measures and experimental design. Six of 21 intervention studies proved to be methodologically strong. However, only one study was considered a randomized clinical trial, testing the combined effects of weight reduction and pharmacological treatment of hypertension. Average blood pressure decrease obtained from the methodologically strongest studies was -21 mmHg and -13 mmHg, for systolic and diastolic measures, respectively. This magnitude change suggests that weight loss may be a clinically and statistically significant treatment. Confounding and bias variables, such as adherence to diet, medication, salt consumption, etc., were discussed and future areas of research were outlined. It was concluded that weight loss appears to be an effective and safe treatment of hypertension. PMID:7039371
Data Point Averaging for Computational Fluid Dynamics Data
NASA Technical Reports Server (NTRS)
Norman, David, Jr. (Inventor)
2014-01-01
A system and method for generating fluid flow parameter data for use in aerodynamic heating analysis. Computational fluid dynamics data is generated for a number of points in an area on a surface to be analyzed. Sub-areas corresponding to areas of the surface for which an aerodynamic heating analysis is to be performed are identified. A computer system automatically determines a sub-set of the number of points corresponding to each of the number of sub-areas and determines a value for each of the number of sub-areas using the data for the sub-set of points corresponding to each of the number of sub-areas. The value is determined as an average of the data for the sub-set of points corresponding to each of the number of sub-areas. The resulting parameter values then may be used to perform an aerodynamic heating analysis.
A model-averaging method for assessing groundwater conceptual model uncertainty.
Ye, Ming; Pohlmann, Karl F; Chapman, Jenny B; Pohll, Greg M; Reeves, Donald M
2010-01-01
This study evaluates alternative groundwater models with different recharge and geologic components at the northern Yucca Flat area of the Death Valley Regional Flow System (DVRFS), USA. Recharge over the DVRFS has been estimated using five methods, and five geological interpretations are available at the northern Yucca Flat area. Combining the recharge and geological components together with additional modeling components that represent other hydrogeological conditions yields a total of 25 groundwater flow models. As all the models are plausible given available data and information, evaluating model uncertainty becomes inevitable. On the other hand, hydraulic parameters (e.g., hydraulic conductivity) are uncertain in each model, giving rise to parametric uncertainty. Propagation of the uncertainty in the models and model parameters through groundwater modeling causes predictive uncertainty in model predictions (e.g., hydraulic head and flow). Parametric uncertainty within each model is assessed using Monte Carlo simulation, and model uncertainty is evaluated using the model averaging method. Two model-averaging techniques (on the basis of information criteria and GLUE) are discussed. This study shows that contribution of model uncertainty to predictive uncertainty is significantly larger than that of parametric uncertainty. For the recharge and geological components, uncertainty in the geological interpretations has more significant effect on model predictions than uncertainty in the recharge estimates. In addition, weighted residuals vary more for the different geological models than for different recharge models. Most of the calibrated observations are not important for discriminating between the alternative models, because their weighted residuals vary only slightly from one model to another. PMID:19788638
Federal Register 2010, 2011, 2012, 2013, 2014
2010-12-28
... Department did not grant offsets for non-dumped comparisons. Several World Trade Organization (``WTO... methodologies have been challenged as being inconsistent with the World Trade Organization (``WTO'') General... International Trade Administration 19 CFR Part 351 RIN 0625-AA87 Antidumping Proceedings: Calculation of...
Federal Register 2010, 2011, 2012, 2013, 2014
2012-02-14
...-Year (``Sunset'') Reviews of Antidumping Duty Orders, 76 FR 66893 (Oct. 28, 2011), and accompanying... currently applies in investigations.\\21\\ \\21\\ Proposed Modification for Reviews, 75 FR at 81534. In this.... Several World Trade Organization (``WTO'') dispute settlement reports have found that the United...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-02-01
... certain antidumping duty proceedings (75 FR 81533). That proposed rule and proposed modification indicated... Ronald K. Lorentzen, Deputy Assistant Secretary for Import Administration, Room 1870, Department...
ERIC Educational Resources Information Center
Jenny, Mirjam A.; Rieskamp, Jrg; Nilsson, Hkan
2014-01-01
Judging whether multiple events will co-occur is an important aspect of everyday decision making. The underlying probabilities of occurrence are usually unknown and have to be inferred from experience. Using a rigorous, quantitative model comparison, we investigate how people judge the conjunctive probabilities of multiple events to co-occur. In 2
Implications of the method of capital cost payment on the weighted average cost of capital.
Boles, K E
1986-01-01
The author develops a theoretical and mathematical model, based on published financial management literature, to describe the cost of capital structure for health care delivery entities. This model is then used to generate the implications of changing the capital cost reimbursement mechanism from a cost basis to a prospective basis. The implications are that the cost of capital is increased substantially, the use of debt must be restricted, interest rates for borrowed funds will increase, and, initially, firms utilizing debt efficiently under cost-basis reimbursement will be restricted to the generation of funds from equity only under a prospective system. PMID:3525468
Cernicchiaro, N; Renter, D G; Xiang, S; White, B J; Bello, N M
2013-06-01
Variability in ADG of feedlot cattle can affect profits, thus making overall returns more unstable. Hence, knowledge of the factors that contribute to heterogeneity of variances in animal performance can help feedlot managers evaluate risks and minimize profit volatility when making managerial and economic decisions in commercial feedlots. The objectives of the present study were to evaluate heteroskedasticity, defined as heterogeneity of variances, in ADG of cohorts of commercial feedlot cattle, and to identify cattle demographic factors at feedlot arrival as potential sources of variance heterogeneity, accounting for cohort- and feedlot-level information in the data structure. An operational dataset compiled from 24,050 cohorts from 25 U. S. commercial feedlots in 2005 and 2006 was used for this study. Inference was based on a hierarchical Bayesian model implemented with Markov chain Monte Carlo, whereby cohorts were modeled at the residual level and feedlot-year clusters were modeled as random effects. Forward model selection based on deviance information criteria was used to screen potentially important explanatory variables for heteroskedasticity at cohort- and feedlot-year levels. The Bayesian modeling framework was preferred as it naturally accommodates the inherently hierarchical structure of feedlot data whereby cohorts are nested within feedlot-year clusters. Evidence for heterogeneity of variance components of ADG was substantial and primarily concentrated at the cohort level. Feedlot-year specific effects were, by far, the greatest contributors to ADG heteroskedasticity among cohorts, with an estimated ∼12-fold change in dispersion between most and least extreme feedlot-year clusters. In addition, identifiable demographic factors associated with greater heterogeneity of cohort-level variance included smaller cohort sizes, fewer days on feed, and greater arrival BW, as well as feedlot arrival during summer months. These results support that heterogeneity of variances in ADG is prevalent in feedlot performance and indicate potential sources of heteroskedasticity. Further investigation of factors associated with heteroskedasticity in feedlot performance is warranted to increase consistency and uniformity in commercial beef cattle production and subsequent profitability. PMID:23482583
Random time averaged diffusivities for Lvy walks
NASA Astrophysics Data System (ADS)
Froemberg, D.; Barkai, E.
2013-07-01
We investigate a Lvy walk alternating between velocities v0 with opposite sign. The sojourn time probability distribution at large times is a power law lacking its mean or second moment. The first case corresponds to a ballistic regime where the ensemble averaged mean squared displacement (MSD) at large times is ?x2? ? t2, the latter to enhanced diffusion with ?x2? ? t?, 1 < ? < 2. The correlation function and the time averaged MSD are calculated. In the ballistic case, the deviations of the time averaged MSD from a purely ballistic behavior are shown to be distributed according to a Mittag-Leffler density function. In the enhanced diffusion regime, the fluctuations of the time averages MSD vanish at large times, yet very slowly. In both cases we quantify the discrepancy between the time averaged and ensemble averaged MSDs.
Average-cost based robust structural control
NASA Technical Reports Server (NTRS)
Hagood, Nesbitt W.
1993-01-01
A method is presented for the synthesis of robust controllers for linear time invariant structural systems with parameterized uncertainty. The method involves minimizing quantities related to the quadratic cost (H2-norm) averaged over a set of systems described by real parameters such as natural frequencies and modal residues. Bounded average cost is shown to imply stability over the set of systems. Approximations for the exact average are derived and proposed as cost functionals. The properties of these approximate average cost functionals are established. The exact average and approximate average cost functionals are used to derive dynamic controllers which can provide stability robustness. The robustness properties of these controllers are demonstrated in illustrative numerical examples and tested in a simple SISO experiment on the MIT multi-point alignment testbed.
Calculating ensemble averaged descriptions of protein rigidity without sampling.
González, Luis C; Wang, Hui; Livesay, Dennis R; Jacobs, Donald J
2012-01-01
Previous works have demonstrated that protein rigidity is related to thermodynamic stability, especially under conditions that favor formation of native structure. Mechanical network rigidity properties of a single conformation are efficiently calculated using the integer body-bar Pebble Game (PG) algorithm. However, thermodynamic properties require averaging over many samples from the ensemble of accessible conformations to accurately account for fluctuations in network topology. We have developed a mean field Virtual Pebble Game (VPG) that represents the ensemble of networks by a single effective network. That is, all possible number of distance constraints (or bars) that can form between a pair of rigid bodies is replaced by the average number. The resulting effective network is viewed as having weighted edges, where the weight of an edge quantifies its capacity to absorb degrees of freedom. The VPG is interpreted as a flow problem on this effective network, which eliminates the need to sample. Across a nonredundant dataset of 272 protein structures, we apply the VPG to proteins for the first time. Our results show numerically and visually that the rigidity characterizations of the VPG accurately reflect the ensemble averaged [Formula: see text] properties. This result positions the VPG as an efficient alternative to understand the mechanical role that chemical interactions play in maintaining protein stability. PMID:22383947
Recovery of petroleum with chemically treated high molecular weight polymers
Gibb, C.L.; Rhudy, J.S.
1980-11-18
Plugging of reservoirs with high molecular weight polymers, e.g. Partially hydrolyzed polyacrylamide, is overcome by chemically treating a polymer having an excessively high average molecular weight prior to injection into a reservoir with an oxidizing chemical, e.g. sodium hypochlorite, and thereafter incorporating a reducing chemical, e.g., sodium sulfite, to stop degradation of the polymer when a desired lower average molecular weight and flooding characteristics are attained.
Statistics of time averaged atmospheric scintillation
Stroud, P.
1994-02-01
A formulation has been constructed to recover the statistics of the moving average of the scintillation Strehl from a discrete set of measurements. A program of airborne atmospheric propagation measurements was analyzed to find the correlation function of the relative intensity over displaced propagation paths. The variance in continuous moving averages of the relative intensity was then found in terms of the correlation functions. An empirical formulation of the variance of the continuous moving average of the scintillation Strehl has been constructed. The resulting characterization of the variance of the finite time averaged Strehl ratios is being used to assess the performance of an airborne laser system.
General periodic average structures of decagonal quasicrystals.
Cervellino, Antonio; Steurer, Walter
2002-03-01
The concept of periodic average structure is mutated from the theory of incommensurately modulated structures. For quasicrystals, this concept (up to now explored only in few cases) is becoming increasingly useful to understand their properties and to interpret some important structural features. The peculiar property of quasicrystals is that they admit not one but many (infinite) possible different average structures. Few of them, however, will be meaningful. Here are given a simple method (based on reciprocal space) for generating all the possible periodic average structures of decagonal quasicrystals and some new ideas about their meaning. By this method, the most significant average structures can be recognized from the diffraction pattern. PMID:11832588
Cell averaging Chebyshev methods for hyperbolic problems
NASA Technical Reports Server (NTRS)
Wei, Cai; Gottlieb, David; Harten, Ami
1990-01-01
A cell averaging method for the Chebyshev approximations of first order hyperbolic equations in conservation form is described. Formulas are presented for transforming between pointwise data at the collocation points and cell averaged quantities, and vice-versa. This step, trivial for the finite difference and Fourier methods, is nontrivial for the global polynomials used in spectral methods. The cell averaging methods presented are proven stable for linear scalar hyperbolic equations and present numerical simulations of shock-density wave interaction using the new cell averaging Chebyshev methods.
Spectral and parametric averaging for integrable systems
NASA Astrophysics Data System (ADS)
Ma, Tao; Serota, R. A.
2015-05-01
We analyze two theoretical approaches to ensemble averaging for integrable systems in quantum chaos, spectral averaging (SA) and parametric averaging (PA). For SA, we introduce a new procedure, namely, rescaled spectral averaging (RSA). Unlike traditional SA, it can describe the correlation function of spectral staircase (CFSS) and produce persistent oscillations of the interval level number variance (IV). PA while not as accurate as RSA for the CFSS and IV, can also produce persistent oscillations of the global level number variance (GV) and better describes saturation level rigidity as a function of the running energy. Overall, it is the most reliable method for a wide range of statistics.
NASA Astrophysics Data System (ADS)
Arsenault, Richard; Gatien, Philippe; Renaud, Benoit; Brissette, François; Martel, Jean-Luc
2015-10-01
This study aims to test whether a weighted combination of several hydrological models can simulate flows more accurately than the models taken individually. In addition, the project attempts to identify the most efficient model averaging method and the optimal number of models to include in the weighting scheme. In order to address the first objective, streamflow was simulated using four lumped hydrological models (HSAMI, HMETS, MOHYSE and GR4J-6), each of which were calibrated with three different objective functions on 429 watersheds. The resulting 12 hydrographs (4 models × 3 metrics) were weighted and combined with the help of 9 averaging methods which are the simple arithmetic mean (SAM), Akaike information criterion (AICA), Bates-Granger (BGA), Bayes information criterion (BICA), Bayesian model averaging (BMA), Granger-Ramanathan average variant A, B and C (GRA, GRB and GRC) and the average by SCE-UA optimization (SCA). The same weights were then applied to the hydrographs in validation mode, and the Nash-Sutcliffe Efficiency metric was measured between the averaged and observed hydrographs. Statistical analyses were performed to compare the accuracy of weighted methods to that of individual models. A Kruskal-Wallis test and a multi-objective optimization algorithm were then used to identify the most efficient weighted method and the optimal number of models to integrate. Results suggest that the GRA, GRB, GRC and SCA weighted methods perform better than the individual members. Model averaging from these four methods were superior to the best of the individual members in 76% of the cases. Optimal combinations on all watersheds included at least one of each of the four hydrological models. None of the optimal combinations included all members of the ensemble of 12 hydrographs. The Granger-Ramanathan average variant C (GRC) is recommended as the best compromise between accuracy, speed of execution, and simplicity.
Several weight-loss medicines are available. Ask your health care provider if any are right for you. About 5 to ... whether the cost, side effects, and the small weight loss you can expect are worth it to you. ...
... a healthy eating plan and set realistic weight goals. Do not stop any medicines that may be ... al. Position of the American Dietetic Association: weight management. J Am Diet Assoc . 2009;109:330-46. ...
... and during pregnancy can reduce the likelihood of negative effects for mothers and babies We’ve heard ... that prepregnancy weight and pregnancy weight gain can impact not only pregnancy outcomes for mother and baby, ...
... get us off the hook. How Our Minds Influence Weight Loss Why is it so hard to ... is where counselors (like psychologists or weight-management social workers) can help. They are trained to understand ...
Dynamic Multiscale Averaging (DMA) of Turbulent Flow
Richard W. Johnson
2012-09-01
A new approach called dynamic multiscale averaging (DMA) for computing the effects of turbulent flow is described. The new method encompasses multiple applications of temporal and spatial averaging, that is, multiscale operations. Initially, a direct numerical simulation (DNS) is performed for a relatively short time; it is envisioned that this short time should be long enough to capture several fluctuating time periods of the smallest scales. The flow field variables are subject to running time averaging during the DNS. After the relatively short time, the time-averaged variables are volume averaged onto a coarser grid. Both time and volume averaging of the describing equations generate correlations in the averaged equations. These correlations are computed from the flow field and added as source terms to the computation on the next coarser mesh. They represent coupling between the two adjacent scales. Since they are computed directly from first principles, there is no modeling involved. However, there is approximation involved in the coupling correlations as the flow field has been computed for only a relatively short time. After the time and spatial averaging operations are applied at a given stage, new computations are performed on the next coarser mesh using a larger time step. The process continues until the coarsest scale needed is reached. New correlations are created for each averaging procedure. The number of averaging operations needed is expected to be problem dependent. The new DMA approach is applied to a relatively low Reynolds number flow in a square duct segment. Time-averaged stream-wise velocity and vorticity contours from the DMA approach appear to be very similar to a full DNS for a similar flow reported in the literature. Expected symmetry for the final results is produced for the DMA method. The results obtained indicate that DMA holds significant potential in being able to accurately compute turbulent flow without modeling for practical engineering applications.
Skeleton of weighted social network
NASA Astrophysics Data System (ADS)
Zhang, X.; Zhu, J.
2013-03-01
In the literature of social networks, understanding topological structure is an important scientific issue. In this paper, we construct a network from mobile phone call records and use the cumulative number of calls as a measure of the weight of a social tie. We extract skeletons from the weighted social network on the basis of the weights of ties, and we study their properties. We find that strong ties can support the skeleton in the network by studying the percolation characters. We explore the centrality of w-skeletons based on the correlation between some centrality measures and the skeleton index w of a vertex, and we find that the average centrality of a w-skeleton increases as w increases. We also study the cumulative degree distribution of the successive w-skeletons and find that as w increases, the w-skeleton tends to become more self-similar. Furthermore, fractal characteristics appear in higher w-skeletons. We also explore the global information diffusion efficiency of w-skeletons using simulations, from which we can see that the ties in the high w-skeletons play important roles in information diffusion. Identifying such a simple structure of a w-skeleton is a step forward toward understanding and representing the topological structure of weighted social networks.
ERIC Educational Resources Information Center
Ryan, Kevin Michael
2011-01-01
Research on syllable weight in generative phonology has focused almost exclusively on systems in which weight is treated as an ordinal hierarchy of clearly delineated categories (e.g. light and heavy). As I discuss, canonical weight-sensitive phenomena in phonology, including quantitative meter and quantity-sensitive stress, can also treat weight…
... inches (25 centimeters) during puberty before reaching full adult height. Most kids gain weight more rapidly during this ... her weight and height. The BMI formula uses height and weight measurements to ... is the same for adults and children, figuring out what the BMI number ...
NASA Technical Reports Server (NTRS)
Dyer, A., Jr.; Ferrara, P. W.; Luke, H. P.
1969-01-01
Weight Control System, a set of linked computer programs which provides weight and balance reports from magnetic tape files, provides weight control and reporting on launch vehicle programs. With minor format modifications the program is applicable to aerospace, marine, automotive and other land transportation industries.
40 CFR 63.846 - Emission averaging.
Code of Federal Regulations, 2014 CFR
2014-07-01
... averaging. (a) General. The owner or operator of an existing potline or anode bake furnace in a State that... by total aluminum production. (c) Anode bake furnaces. The owner or operator may average TF emissions from anode bake furnaces and demonstrate compliance with the limits in Table 3 of this subpart...
40 CFR 63.846 - Emission averaging.
Code of Federal Regulations, 2012 CFR
2012-07-01
... averaging. (a) General. The owner or operator of an existing potline or anode bake furnace in a State that... by total aluminum production. (c) Anode bake furnaces. The owner or operator may average TF emissions from anode bake furnaces and demonstrate compliance with the limits in Table 3 of this subpart...
40 CFR 63.846 - Emission averaging.
Code of Federal Regulations, 2013 CFR
2013-07-01
... averaging. (a) General. The owner or operator of an existing potline or anode bake furnace in a State that... by total aluminum production. (c) Anode bake furnaces. The owner or operator may average TF emissions from anode bake furnaces and demonstrate compliance with the limits in Table 3 of this subpart...
40 CFR 76.11 - Emissions averaging.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 16 2010-07-01 2010-07-01 false Emissions averaging. 76.11 Section 76.11 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General provisions. In lieu of complying with the...
Average Transmission Probability of a Random Stack
ERIC Educational Resources Information Center
Lu, Yin; Miniatura, Christian; Englert, Berthold-Georg
2010-01-01
The transmission through a stack of identical slabs that are separated by gaps with random widths is usually treated by calculating the average of the logarithm of the transmission probability. We show how to calculate the average of the transmission probability itself with the aid of a recurrence relation and derive analytical upper and lower
Code of Federal Regulations, 2012 CFR
2012-07-01
... 40 Protection of Environment 21 2012-07-01 2012-07-01 false Averaging. 90.204 Section 90.204 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NONROAD SPARK-IGNITION ENGINES AT OR BELOW 19 KILOWATTS Certification Averaging, Banking,...
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Averaging. 90.204 Section 90.204 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NONROAD SPARK-IGNITION ENGINES AT OR BELOW 19 KILOWATTS Certification Averaging, Banking,...
Code of Federal Regulations, 2013 CFR
2013-07-01
... 40 Protection of Environment 21 2013-07-01 2013-07-01 false Averaging. 90.204 Section 90.204 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NONROAD SPARK-IGNITION ENGINES AT OR BELOW 19 KILOWATTS Certification Averaging, Banking,...
Code of Federal Regulations, 2011 CFR
2011-07-01
... 40 Protection of Environment 20 2011-07-01 2011-07-01 false Averaging. 90.204 Section 90.204 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NONROAD SPARK-IGNITION ENGINES AT OR BELOW 19 KILOWATTS Certification Averaging, Banking,...
Code of Federal Regulations, 2014 CFR
2014-07-01
... 40 Protection of Environment 20 2014-07-01 2013-07-01 true Averaging. 90.204 Section 90.204 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NONROAD SPARK-IGNITION ENGINES AT OR BELOW 19 KILOWATTS Certification Averaging, Banking,...
Determinants of College Grade Point Averages
ERIC Educational Resources Information Center
Bailey, Paul Dean
2012-01-01
Chapter 2: The Role of Class Difficulty in College Grade Point Averages. Grade Point Averages (GPAs) are widely used as a measure of college students' ability. Low GPAs can remove a students from eligibility for scholarships, and even continued enrollment at a university. However, GPAs are determined not only by student ability but also by
40 CFR 63.846 - Emission averaging.
Code of Federal Regulations, 2011 CFR
2011-07-01
... operator may average TF emissions from potlines and demonstrate compliance with the limits in Table 1 of... operator also may average POM emissions from potlines and demonstrate compliance with the limits in Table 2... limit in Table 1 of this subpart (for TF emissions) and/or Table 2 of this subpart (for POM...
Determinants of College Grade Point Averages
ERIC Educational Resources Information Center
Bailey, Paul Dean
2012-01-01
Chapter 2: The Role of Class Difficulty in College Grade Point Averages. Grade Point Averages (GPAs) are widely used as a measure of college students' ability. Low GPAs can remove a students from eligibility for scholarships, and even continued enrollment at a university. However, GPAs are determined not only by student ability but also by…
40 CFR 63.846 - Emission averaging.
Code of Federal Regulations, 2010 CFR
2010-07-01
... averaging. (a) General. The owner or operator of an existing potline or anode bake furnace in a State that... by total aluminum production. (c) Anode bake furnaces. The owner or operator may average TF emissions from anode bake furnaces and demonstrate compliance with the limits in Table 3 of this subpart...
The Hubble rate in averaged cosmology
Umeh, Obinna; Larena, Julien; Clarkson, Chris E-mail: julien.larena@gmail.com
2011-03-01
The calculation of the averaged Hubble expansion rate in an averaged perturbed Friedmann-Lematre-Robertson-Walker cosmology leads to small corrections to the background value of the expansion rate, which could be important for measuring the Hubble constant from local observations. It also predicts an intrinsic variance associated with the finite scale of any measurement of H{sub 0}, the Hubble rate today. Both the mean Hubble rate and its variance depend on both the definition of the Hubble rate and the spatial surface on which the average is performed. We quantitatively study different definitions of the averaged Hubble rate encountered in the literature by consistently calculating the backreaction effect at second order in perturbation theory, and compare the results. We employ for the first time a recently developed gauge-invariant definition of an averaged scalar. We also discuss the variance of the Hubble rate for the different definitions.
Evaluating template bias when synthesizing population averages
NASA Astrophysics Data System (ADS)
Carlson, Blake L.; Christensen, Gary E.; Johnson, Hans J.; Vannier, Michael W.
2001-07-01
Establishing the average shape and spatial variability for a set of similar anatomical objects is important for detecting and discriminating morphological differences between populations. This may be done using deformable templates to synthesize a 3D CT/MRI image of the average anatomy from a set of CT/MRI images collected from a population of similar anatomical objects. This paper investigates the error associated with the choice of template selected from the population used to synthesize the average population shape. Population averages were synthesized for a population of five infant skulls with sagittal synostosis and a population of six normal adult brains using a consistent linear-elastic image registration algorithm. Each data set from the populations was used as the template to synthesize a population average. This resulted in five different population averages for the skull population and six different population averages for the brain population. The displacement variance distance from a skull within the population to the other skulls in the population ranged from 5.5 to 9.9 mm2 while the displacement variance distance from the synthesized average skulls to the population ranged from 2.2 to 2.7 mm2. The displacement variance distance from a brain within the population to the other brains in the population ranged from 9.3 to 14.2 mm2 while the displacement variance distance from the synthesized average brains to the population ranged from 3.2 to 3.6 mm2. These results suggest that there was no significant difference between the choice of template with respect to the shape of the synthesized average data set for these two populations.
Minimizing Average Shortest Path Distances via Shortcut Edge Addition
NASA Astrophysics Data System (ADS)
Meyerson, Adam; Tagiku, Brian
We consider adding k shortcut edges (i.e. edges of small fixed length ? ? 0) to a graph so as to minimize the weighted average shortest path distance over all pairs of vertices. We explore several variations of the problem and give O(1)-approximations for each. We also improve the best known approximation ratio for metric k-median with penalties, as many of our approximations depend upon this bound. We give a (1+2(p+1)/?(p+1)-1,?)-approximation with runtime exponential in p. If we set ?= 1 (to be exact on the number of medians), this matches the best current k-median (without penalties) result.
Overweight, obesity, and weight change among incarcerated women.
Clarke, Jennifer G; Waring, Molly E
2012-10-01
Excessive weight gain among inmates is frequently observed by correctional health care providers; however, there is little published research on weight change during incarceration. This study describes the weight and weight changes among women incarcerated in a unified correctional system (prison and jail). The women were interviewed and had their height and weight measured. At baseline, 33.0% were of normal weight, 34.9% were overweight, and 32.1% were obese. Participants were reweighed after a median of 14 days; the women had gained an average of 1.1 lbs/week (SD: 2.1 lbs, range: -3.3 to +9.2 lbs) with 71% of women gaining weight. Women incarcerated for 2 weeks or less at time of study enrollment experienced higher average weight weekly gains than those incarcerated longer than 2 weeks (1.7 lbs vs. 0.8 lbs). PMID:22899812
Weighted Watson-Crick automata
Tamrin, Mohd Izzuddin Mohd; Turaev, Sherzod; Sembok, Tengku Mohd Tengku
2014-07-10
There are tremendous works in biotechnology especially in area of DNA molecules. The computer society is attempting to develop smaller computing devices through computational models which are based on the operations performed on the DNA molecules. A Watson-Crick automaton, a theoretical model for DNA based computation, has two reading heads, and works on double-stranded sequences of the input related by a complementarity relation similar with the Watson-Crick complementarity of DNA nucleotides. Over the time, several variants of Watson-Crick automata have been introduced and investigated. However, they cannot be used as suitable DNA based computational models for molecular stochastic processes and fuzzy processes that are related to important practical problems such as molecular parsing, gene disease detection, and food authentication. In this paper we define new variants of Watson-Crick automata, called weighted Watson-Crick automata, developing theoretical models for molecular stochastic and fuzzy processes. We define weighted Watson-Crick automata adapting weight restriction mechanisms associated with formal grammars and automata. We also study the generative capacities of weighted Watson-Crick automata, including probabilistic and fuzzy variants. We show that weighted variants of Watson-Crick automata increase their generative power.
Weighted Watson-Crick automata
NASA Astrophysics Data System (ADS)
Tamrin, Mohd Izzuddin Mohd; Turaev, Sherzod; Sembok, Tengku Mohd Tengku
2014-07-01
There are tremendous works in biotechnology especially in area of DNA molecules. The computer society is attempting to develop smaller computing devices through computational models which are based on the operations performed on the DNA molecules. A Watson-Crick automaton, a theoretical model for DNA based computation, has two reading heads, and works on double-stranded sequences of the input related by a complementarity relation similar with the Watson-Crick complementarity of DNA nucleotides. Over the time, several variants of Watson-Crick automata have been introduced and investigated. However, they cannot be used as suitable DNA based computational models for molecular stochastic processes and fuzzy processes that are related to important practical problems such as molecular parsing, gene disease detection, and food authentication. In this paper we define new variants of Watson-Crick automata, called weighted Watson-Crick automata, developing theoretical models for molecular stochastic and fuzzy processes. We define weighted Watson-Crick automata adapting weight restriction mechanisms associated with formal grammars and automata. We also study the generative capacities of weighted Watson-Crick automata, including probabilistic and fuzzy variants. We show that weighted variants of Watson-Crick automata increase their generative power.
Light propagation in the averaged universe
Bagheri, Samae; Schwarz, Dominik J. E-mail: dschwarz@physik.uni-bielefeld.de
2014-10-01
Cosmic structures determine how light propagates through the Universe and consequently must be taken into account in the interpretation of observations. In the standard cosmological model at the largest scales, such structures are either ignored or treated as small perturbations to an isotropic and homogeneous Universe. This isotropic and homogeneous model is commonly assumed to emerge from some averaging process at the largest scales. We assume that there exists an averaging procedure that preserves the causal structure of space-time. Based on that assumption, we study the effects of averaging the geometry of space-time and derive an averaged version of the null geodesic equation of motion. For the averaged geometry we then assume a flat Friedmann-Lematre (FL) model and find that light propagation in this averaged FL model is not given by null geodesics of that model, but rather by a modified light propagation equation that contains an effective Hubble expansion rate, which differs from the Hubble rate of the averaged space-time.
Solvated Ensemble Averaging in the Calculation of Partial Atomic Charges
BASMA, M.; SUNDARA, S.; ALGAN, D.; VERNALI, T.; WOODS, R. J.
2006-01-01
In the calculation of partial atomic charges, for use in molecular mechanics or dynamics simulations, it is common practice to select only a single conformation for the molecule of interest. For molecules that contain rotatable bonds, it is preferable to compute the charges from several relevant conformations. We present here results from a charge derivation protocol that determines the partial charges by averaging charges computed for conformations selected from explicitly solvated MD simulations, performed under periodic boundary conditions. This approach leads to partial charges that are weighted by a realistic population of conformations and that are suitable for condensed phase simulations. This protocol can, in principle, be applied to any class of molecule and to nonaqueous solvation. Carbohydrates contain numerous hydroxyl groups that exist in an ensemble of orientations in solution, and in this report we apply ensemble averaging to a series of methyl glycosides. We report the extent to which ensemble averaging leads to charge convergence among the various monosaccharides and among the constituent atoms within a given monosaccharide. Due to the large number of conformations (200) in our ensembles, we are able to compute statistically relevant standard deviations for the partial charges. An analysis of the standard deviations allows us to assess the extent to which equivalent atom types may, nevertheless, require unique partial charges. The configurations of the hydroxyl groups exert considerable influence on internal energies, and the limits of ensemble averaged charges are discussed in terms of these properties. PMID:17882310
Jacques, Paul F; Wang, Huifen
2014-05-01
A large body of observational studies and randomized controlled trials (RCTs) has examined the role of dairy products in weight loss and maintenance of healthy weight. Yogurt is a dairy product that is generally very similar to milk, but it also has some unique properties that may enhance its possible role in weight maintenance. This review summarizes the human RCT and prospective observational evidence on the relation of yogurt consumption to the management and maintenance of body weight and composition. The RCT evidence is limited to 2 small, short-term, energy-restricted trials. They both showed greater weight losses with yogurt interventions, but the difference between the yogurt intervention and the control diet was only significant in one of these trials. There are 5 prospective observational studies that have examined the association between yogurt and weight gain. The results of these studies are equivocal. Two of these studies reported that individuals with higher yogurt consumption gained less weight over time. One of these same studies also considered changes in waist circumference (WC) and showed that higher yogurt consumption was associated with smaller increases in WC. A third study was inconclusive because of low statistical power. A fourth study observed no association between changes in yogurt intake and weight gain, but the results suggested that those with the largest increases in yogurt intake during the study also had the highest increase in WC. The final study examined weight and WC change separately by sex and baseline weight status and showed benefits for both weight and WC changes for higher yogurt consumption in overweight men, but it also found that higher yogurt consumption in normal-weight women was associated with a greater increase in weight over follow-up. Potential underlying mechanisms for the action of yogurt on weight are briefly discussed. PMID:24695887
Effects of wildfire disaster exposure on male birth weight in an Australian population
O’Donnell, M. H.; Behie, A. M.
2015-01-01
Background and objectives: Maternal stress can depress birth weight and gestational age, with potential health effects. A growing number of studies examine the effect of maternal stress caused by environmental disasters on birth outcomes. These changes may indicate an adaptive response. In this study, we examine the effects of maternal exposure to wildfire on birth weight and gestational age, hypothesising that maternal stress will negatively influence these measures. Methodology: Using data from the Australian Capital Territory, we employed Analysis of Variance to examine the influence of the 2003 Canberra wildfires on the weight of babies born to mothers resident in fire-affected regions, while considering the role of other factors. Results: We found that male infants born in the most severely fire-affected area had significantly higher average birth weights than their less exposed peers and were also heavier than males born in the same areas in non-fire years. Higher average weights were attributable to an increase in the number of macrosomic infants. There was no significant effect on the weight of female infants or on gestational age for either sex. Conclusions and implications: Our findings indicate heightened environmental responsivity in the male cohort. We find that elevated maternal stress acted to accelerate the growth of male fetuses, potentially through an elevation of maternal blood glucose levels. Like previous studies, our work finds effects of disaster exposure and suggests that fetal growth patterns respond to maternal signals. However, the direction of the change in birth weight is opposite to that of many earlier studies. PMID:26574560
Cosmic inhomogeneities and averaged cosmological dynamics.
Paranjape, Aseem; Singh, T P
2008-10-31
If general relativity (GR) describes the expansion of the Universe, the observed cosmic acceleration implies the existence of a "dark energy." However, while the Universe is on average homogeneous on large scales, it is inhomogeneous on smaller scales. While GR governs the dynamics of the inhomogeneous Universe, the averaged homogeneous Universe obeys modified Einstein equations. Can such modifications alone explain the acceleration? For a simple generic model with realistic initial conditions, we show the answer to be "no." Averaging effects negligibly influence the cosmological dynamics. PMID:18999811
Spacetime Average Density (SAD) cosmological measures
Page, Don N.
2014-11-01
The measure problem of cosmology is how to obtain normalized probabilities of observations from the quantum state of the universe. This is particularly a problem when eternal inflation leads to a universe of unbounded size so that there are apparently infinitely many realizations or occurrences of observations of each of many different kinds or types, making the ratios ambiguous. There is also the danger of domination by Boltzmann Brains. Here two new Spacetime Average Density (SAD) measures are proposed, Maximal Average Density (MAD) and Biased Average Density (BAD), for getting a finite number of observation occurrences by using properties of the Spacetime Average Density (SAD) of observation occurrences to restrict to finite regions of spacetimes that have a preferred beginning or bounce hypersurface. These measures avoid Boltzmann brain domination and appear to give results consistent with other observations that are problematic for other widely used measures, such as the observation of a positive cosmological constant.
40 CFR 76.11 - Emissions averaging.
Code of Federal Regulations, 2012 CFR
2012-07-01
...) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General... compliance with the Acid Rain emission limitation for NOX under the plan only if the following...
40 CFR 76.11 - Emissions averaging.
Code of Federal Regulations, 2011 CFR
2011-07-01
...) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General... compliance with the Acid Rain emission limitation for NOX under the plan only if the following...
Rotational averaging of multiphoton absorption cross sections
Friese, Daniel H. Beerepoot, Maarten T. P.; Ruud, Kenneth
2014-11-28
Rotational averaging of tensors is a crucial step in the calculation of molecular properties in isotropic media. We present a scheme for the rotational averaging of multiphoton absorption cross sections. We extend existing literature on rotational averaging to even-rank tensors of arbitrary order and derive equations that require only the number of photons as input. In particular, we derive the first explicit expressions for the rotational average of five-, six-, and seven-photon absorption cross sections. This work is one of the required steps in making the calculation of these higher-order absorption properties possible. The results can be applied to any even-rank tensor provided linearly polarized light is used.
Modeling Plants With Moving-Average Outputs
NASA Technical Reports Server (NTRS)
Polites, Michael E.
1989-01-01
Three discrete-state-variable representations derived. Paper discusses mathematical modeling of digital control systems for plants in which outputs include combinations of instantaneous and moving-average-prefiltered measurements.
Tailoring dietary approaches for weight loss.
Gardner, C D
2012-07-01
Although the 'Low-Fat' diet was the predominant public health recommendation for weight loss and weight control for the past several decades, the obesity epidemic continued to grow during this time period. An alternative 'low-carbohydrate' (Low-Carb) approach, although originally dismissed and even vilified, was comparatively tested in a series of studies over the past decade, and has been found in general to be as effective, if not more, as the Low-Fat approach for weight loss and for several related metabolic health measures. From a glass half full perspective, this suggests that there is more than one choice for a dietary approach to lose weight, and that Low-Fat and Low-Carb diets may be equally effective. From a glass half empty perspective, the average amount of weight lost on either of these two dietary approaches under the conditions studied, particularly when followed beyond 1 year, has been modest at best and negligible at worst, suggesting that the two approaches may be equally ineffective. One could resign themselves at this point to focusing on calories and energy intake restriction, regardless of macronutrient distributions. However, before throwing out the half-glass of water, it is worthwhile to consider that focusing on average results may mask important subgroup successes and failures. In all weight-loss studies, without exception, the range of individual differences in weight change within any particular diet groups is orders of magnitude greater than the average group differences between diet groups. Several studies have now reported that adults with greater insulin resistance are more successful with weight loss on a lower-carbohydrate diet compared with a lower-fat diet, whereas adults with greater insulin sensitivity are equally or more successful with weight loss on a lower-fat diet compared with a lower-carbohydrate diet. Other preliminary findings suggest that there may be some promise with matching individuals with certain genotypes to one type of diet over another for increasing weight-loss success. Future research to address the macronutrient intake component of the obesity epidemic should build on these recent insights and be directed toward effectively classifying individuals who can be differentially matched to alternate types of weight-loss diets that maximize weight-loss and weight-control success. PMID:25089189
Tailoring dietary approaches for weight loss
Gardner, C D
2012-01-01
Although the ‘Low-Fat' diet was the predominant public health recommendation for weight loss and weight control for the past several decades, the obesity epidemic continued to grow during this time period. An alternative ‘low-carbohydrate' (Low-Carb) approach, although originally dismissed and even vilified, was comparatively tested in a series of studies over the past decade, and has been found in general to be as effective, if not more, as the Low-Fat approach for weight loss and for several related metabolic health measures. From a glass half full perspective, this suggests that there is more than one choice for a dietary approach to lose weight, and that Low-Fat and Low-Carb diets may be equally effective. From a glass half empty perspective, the average amount of weight lost on either of these two dietary approaches under the conditions studied, particularly when followed beyond 1 year, has been modest at best and negligible at worst, suggesting that the two approaches may be equally ineffective. One could resign themselves at this point to focusing on calories and energy intake restriction, regardless of macronutrient distributions. However, before throwing out the half-glass of water, it is worthwhile to consider that focusing on average results may mask important subgroup successes and failures. In all weight-loss studies, without exception, the range of individual differences in weight change within any particular diet groups is orders of magnitude greater than the average group differences between diet groups. Several studies have now reported that adults with greater insulin resistance are more successful with weight loss on a lower-carbohydrate diet compared with a lower-fat diet, whereas adults with greater insulin sensitivity are equally or more successful with weight loss on a lower-fat diet compared with a lower-carbohydrate diet. Other preliminary findings suggest that there may be some promise with matching individuals with certain genotypes to one type of diet over another for increasing weight-loss success. Future research to address the macronutrient intake component of the obesity epidemic should build on these recent insights and be directed toward effectively classifying individuals who can be differentially matched to alternate types of weight-loss diets that maximize weight-loss and weight-control success. PMID:25089189
Applications of high average power nonlinear optics
Velsko, S.P.; Krupke, W.F.
1996-02-05
Nonlinear optical frequency convertors (harmonic generators and optical parametric oscillators are reviewed with an emphasis on high average power performance and limitations. NLO materials issues and NLO device designs are discussed in reference to several emerging scientific, military and industrial commercial applications requiring {approx} 100 watt average power level in the visible and infrared spectral regions. Research efforts required to enable practical {approx} 100 watt class NLO based laser systems are identified.
Monthly average polar sea-ice concentration
Schweitzer, Peter N.
1995-01-01
The data contained in this CD-ROM depict monthly averages of sea-ice concentration in the modern polar oceans. These averages were derived from the Scanning Multichannel Microwave Radiometer (SMMR) and Special Sensor Microwave/Imager (SSM/I) instruments aboard satellites of the U.S. Air Force Defense Meteorological Satellite Program from 1978 through 1992. The data are provided as 8-bit images using the Hierarchical Data Format (HDF) developed by the National Center for Supercomputing Applications.
Instrument to average 100 data sets
NASA Technical Reports Server (NTRS)
Tuma, G. B.; Birchenough, A. G.; Rice, W. J.
1977-01-01
An instrumentation system is currently under development which will measure many of the important parameters associated with the operation of an internal combustion engine. Some of these parameters include mass-fraction burn rate, ignition energy, and the indicated mean effective pressure. One of the characteristics of an internal combustion engine is the cycle-to-cycle variation of these parameters. A curve-averaging instrument has been produced which will generate the average curve, over 100 cycles, of any engine parameter. the average curve is described by 2048 discrete points which are displayed on an oscilloscope screen to facilitate recording and is available in real time. Input can be any parameter which is expressed as a + or - 10-volt signal. Operation of the curve-averaging instrument is defined between 100 and 6000 rpm. Provisions have also been made for averaging as many as four parameters simultaneously, with a subsequent decrease in resolution. This provides the means to correlate and perhaps interrelate the phenomena occurring in an internal combustion engine. This instrument has been used successfully on a 1975 Chevrolet V8 engine, and on a Continental 6-cylinder aircraft engine. While this instrument was designed for use on an internal combustion engine, with some modification it can be used to average any cyclically varying waveform.
Computation of vertically averaged velocities in irregular sections of straight channels
NASA Astrophysics Data System (ADS)
Spada, E.; Tucciarelli, T.; Sinagra, M.; Sammartano, V.; Corato, G.
2015-09-01
Two new methods for vertically averaged velocity computation are presented, validated and compared with other available formulas. The first method derives from the well-known Huthoff algorithm, which is first shown to be dependent on the way the river cross section is discretized into several subsections. The second method assumes the vertically averaged longitudinal velocity to be a function only of the friction factor and of the so-called "local hydraulic radius", computed as the ratio between the integral of the elementary areas around a given vertical and the integral of the elementary solid boundaries around the same vertical. Both integrals are weighted with a linear shape function equal to zero at a distance from the integration variable which is proportional to the water depth according to an empirical coefficient β. Both formulas are validated against (1) laboratory experimental data, (2) discharge hydrographs measured in a real site, where the friction factor is estimated from an unsteady-state analysis of water levels recorded in two different river cross sections, and (3) the 3-D solution obtained using the commercial ANSYS CFX code, computing the steady-state uniform flow in a cross section of the Alzette River.
Zhao, Kaiguang; Valle, Denis; Popescu, Sorin; Zhang, Xuesong; Malick, Bani
2013-05-15
Model specification remains challenging in spectroscopy of plant biochemistry, as exemplified by the availability of various spectral indices or band combinations for estimating the same biochemical. This lack of consensus in model choice across applications argues for a paradigm shift in hyperspectral methods to address model uncertainty and misspecification. We demonstrated one such method using Bayesian model averaging (BMA), which performs variable/band selection and quantifies the relative merits of many candidate models to synthesize a weighted average model with improved predictive performances. The utility of BMA was examined using a portfolio of 27 foliage spectral–chemical datasets representing over 80 species across the globe to estimate multiple biochemical properties, including nitrogen, hydrogen, carbon, cellulose, lignin, chlorophyll (a or b), carotenoid, polar and nonpolar extractives, leaf mass per area, and equivalent water thickness. We also compared BMA with partial least squares (PLS) and stepwise multiple regression (SMR). Results showed that all the biochemicals except carotenoid were accurately estimated from hyerspectral data with R2 values > 0.80.
Theory of optimal weighting of data to detect climatic change
NASA Technical Reports Server (NTRS)
Bell, T. L.
1986-01-01
A search for climatic change predicted by climate models can easily yield unconvincing results because of 'climatic noise,' the inherent, unpredictable variability of time-average atmospheric data. A weighted average of data that maximizes the probability of detecting predicted climatic change is presented. To obtain the optimal weights, an estimate of the covariance matrix of the data from a prior data set is needed. This introduces additional sampling error into the method. This is presently taken into account. A form of the weighted average is found whose probability distribution is independent of the true (but unknown) covariance statistics of the data and of the climate model prediction.
Ensemble bayesian model averaging using markov chain Monte Carlo sampling
Vrugt, Jasper A; Diks, Cees G H; Clark, Martyn P
2008-01-01
Bayesian model averaging (BMA) has recently been proposed as a statistical method to calibrate forecast ensembles from numerical weather models. Successful implementation of BMA however, requires accurate estimates of the weights and variances of the individual competing models in the ensemble. In their seminal paper (Raftery etal. Mon Weather Rev 133: 1155-1174, 2(05)) has recommended the Expectation-Maximization (EM) algorithm for BMA model training, even though global convergence of this algorithm cannot be guaranteed. In this paper, we compare the performance of the EM algorithm and the recently developed Differential Evolution Adaptive Metropolis (DREAM) Markov Chain Monte Carlo (MCMC) algorithm for estimating the BMA weights and variances. Simulation experiments using 48-hour ensemble data of surface temperature and multi-model stream-flow forecasts show that both methods produce similar results, and that their performance is unaffected by the length of the training data set. However, MCMC simulation with DREAM is capable of efficiently handling a wide variety of BMA predictive distributions, and provides useful information about the uncertainty associated with the estimated BMA weights and variances.
Unbiased Average Age-Appropriate Atlases for Pediatric Studies
Fonov, Vladimir; Evans, Alan C.; Botteron, Kelly; Almli, C. Robert; McKinstry, Robert C.; Collins, D. Louis
2010-01-01
Spatial normalization, registration, and segmentation techniques for Magnetic Resonance Imaging (MRI) often use a target or template volume to facilitate processing, take advantage of prior information, and define a common coordinate system for analysis. In the neuroimaging literature, the MNI305 Talairach-like coordinate system is often used as a standard template. However, when studying pediatric populations, variation from the adult brain makes the MNI305 suboptimal for processing brain images of children. Morphological changes occurring during development render the use of age-appropriate templates desirable to reduce potential errors and minimize bias during processing of pediatric data. This paper presents the methods used to create unbiased, age-appropriate MRI atlas templates for pediatric studies that represent the average anatomy for the age range of 4.5–18.5 years, while maintaining a high level of anatomical detail and contrast. The creation of anatomical T1-weighted, T2-weighted, and proton density-weighted templates for specific developmentally important age-ranges, used data derived from the largest epidemiological, representative (healthy and normal) sample of the U.S. population, where each subject was carefully screened for medical and psychiatric factors and characterized using established neuropsychological and behavioral assessments. . Use of these age-specific templates was evaluated by computing average tissue maps for gray matter, white matter, and cerebrospinal fluid for each specific age range, and by conducting an exemplar voxel-wise deformation-based morphometry study using 66 young (4.5–6.9 years) participants to demonstrate the benefits of using the age-appropriate templates. The public availability of these atlases/templates will facilitate analysis of pediatric MRI data and enable comparison of results between studies in a common standardized space specific to pediatric research. PMID:20656036
Average luminosity distance in inhomogeneous universes
Kostov, Valentin
2010-04-01
Using numerical ray tracing, the paper studies how the average distance modulus in an inhomogeneous universe differs from its homogeneous counterpart. The averaging is over all directions from a fixed observer not over all possible observers (cosmic), thus is more directly applicable to our observations. In contrast to previous studies, the averaging is exact, non-perturbative, and includes all non-linear effects. The inhomogeneous universes are represented by Swiss-cheese models containing random and simple cubic lattices of mass-compensated voids. The Earth observer is in the homogeneous cheese which has an Einstein-de Sitter metric. For the first time, the averaging is widened to include the supernovas inside the voids by assuming the probability for supernova emission from any comoving volume is proportional to the rest mass in it. Voids aligned along a certain direction give rise to a distance modulus correction which increases with redshift and is caused by cumulative gravitational lensing. That correction is present even for small voids and depends on their density contrast, not on their radius. Averaging over all directions destroys the cumulative lensing correction even in a non-randomized simple cubic lattice of voids. At low redshifts, the average distance modulus correction does not vanish due to the peculiar velocities, despite the photon flux conservation argument. A formula for the maximal possible average correction as a function of redshift is derived and shown to be in excellent agreement with the numerical results. The formula applies to voids of any size that: (a)have approximately constant densities in their interior and walls; and (b)are not in a deep nonlinear regime. The average correction calculated in random and simple cubic void lattices is severely damped below the predicted maximal one after a single void diameter. That is traced to cancellations between the corrections from the fronts and backs of different voids. The results obtained allow one to readily predict the redshift above which the direction-averaged fluctuation in the Hubble diagram falls below a required precision and suggest a method to extract the background Hubble constant from low redshift data without the need to correct for peculiar velocities.
Conditional simulation of geologically averaged block permeabilities
NASA Astrophysics Data System (ADS)
Journel, A. G.
1996-08-01
Currently available hardware and software for flow simulation can handle up to hundreds of thousands of blocks, or more comfortably tens of thousands of blocks. This limits the discretization of the reservoir model to an extremely coarse grid, say 200 × 200 × 25 for 10 6 blocks. Such a coarse grid cannot represent the structural and petrophysical variability at the resolution provided to geologists by well logs and outcrops. Thus there is no alternative to averaging the impact of all small-scale, within-block, heterogeneities into block 'pseudos' or average values. The flow simulator will account for geological description only through those pseudos, hence detailed modelling of geological heterogeneity should not go beyond the information that block pseudos can carry, at least for flow simulation purposes. It is suggested that the present drive in outcrop sampling be clearly redirected at evaluating 'geopseudos', i.e. at evaluating how small-scale variability (both structural and petrophysical) of typical depositional units averages out into large blocks' effective transmissivities and relative permeabilities. Outcrop data would allow the building of generic, high-resolution, numerical models of the geo-variability within a typical depositional unit: this is where geology intervenes. Then, this numerical model would be input into a generic flow simulator, single or multiphase, yielding genetic block averages, for blocks of various sizes and geometries: this is where the reservoir engineer intervenes. Next, the spatial statistics of these block averages (histograms, variograms, …) would be inferred: this is where the geostatistician intervenes. Last comes the problem of filling-in the actual reservoir volume with simulated block averages specific to each depositional unit. Because each reservoir is unique, random drawing of block average values from the previously inferred generic distributions would not be enough. The placement of block average values in the specific reservoir volume must be made conditional on local data whether well log, seismic or production-derived. This non-trivial task of 'conditional simulation' of block average is the challenge of both the reservoir geologist and geostatistician. This paper proposes an avenue of approach that draws from the pioneering works of Steve Begg at BP-Alaska (1992, 1994) and Jaime Gomez-Hernandez at Universidad of Valencia (1990, 1991).
A Stochastic Model of Space-Time Variability of Mesoscale Rainfall: Statistics of Spatial Averages
NASA Technical Reports Server (NTRS)
Kundu, Prasun K.; Bell, Thomas L.
2003-01-01
A characteristic feature of rainfall statistics is that they depend on the space and time scales over which rain data are averaged. A previously developed spectral model of rain statistics that is designed to capture this property, predicts power law scaling behavior for the second moment statistics of area-averaged rain rate on the averaging length scale L as L right arrow 0. In the present work a more efficient method of estimating the model parameters is presented, and used to fit the model to the statistics of area-averaged rain rate derived from gridded radar precipitation data from TOGA COARE. Statistical properties of the data and the model predictions are compared over a wide range of averaging scales. An extension of the spectral model scaling relations to describe the dependence of the average fraction of grid boxes within an area containing nonzero rain (the "rainy area fraction") on the grid scale L is also explored.
Self-averaging in complex brain neuron signals
NASA Astrophysics Data System (ADS)
Bershadskii, A.; Dremencov, E.; Fukayama, D.; Yadid, G.
2002-12-01
Nonlinear statistical properties of Ventral Tegmental Area (VTA) of limbic brain are studied in vivo. VTA plays key role in generation of pleasure and in development of psychological drug addiction. It is shown that spiking time-series of the VTA dopaminergic neurons exhibit long-range correlations with self-averaging behavior. This specific VTA phenomenon has no relation to VTA rewarding function. Last result reveals complex role of VTA in limbic brain.
A re-averaged WENO reconstruction and a third order CWENO scheme for hyperbolic conservation laws
NASA Astrophysics Data System (ADS)
Huang, Chieh-Sen; Arbogast, Todd; Hung, Chen-Hui
2014-04-01
A WENO re-averaging (or re-mapping) technique is developed that converts function averages on one grid to another grid to high order. Nonlinear weighting gives the essentially non-oscillatory property to the re-averaged function values. The new reconstruction grid is used to obtain a standard high order WENO reconstruction of the function averages at a select point. By choosing the reconstruction grid to include the point of interest, a high order function value can be reconstructed using only positive linear weights. The re-averaging technique is applied to define two variants of a classic CWENO3 scheme that combines two linear polynomials to obtain formal third order accuracy. Such a scheme cannot otherwise be defined, due to the nonexistence of linear weights for third order reconstruction at the center of a grid element. The new scheme uses a compact stencil of three solution averages, and only positive linear weights are used. The scheme extends easily to problems in higher space dimensions, essentially as a tensor product of the one-dimensional scheme. The scheme maintains formal third order accuracy in higher dimensions. Numerical results show that this CWENO3 scheme is third order accurate for smooth problems and gives good results for non-smooth problems, including those with shocks.
Average Magnification Effect of Clumping of Matter
NASA Astrophysics Data System (ADS)
Kibble, T. W. B.; Lieu, Richard
2005-10-01
The aim of this paper is to reexamine the question of the average magnification in a universe with some inhomogeneously distributed matter. We present an analytic proof, valid under rather general conditions, including clumps of any shape and size and strong lensing, that as long as the clumps are uncorrelated, the average ``reciprocal'' magnification (in one of several possible senses) is precisely the same as in a homogeneous universe with an equal mean density. From this result, we also show that a similar statement can be made about one definition of the average ``direct'' magnification. We discuss, in the context of observations of discrete and extended sources, the physical significance of the various different measures of magnification and the circumstances in which they are appropriate.
Exploiting scale dependence in cosmological averaging
Mattsson, Teppo; Ronkainen, Maria E-mail: maria.ronkainen@helsinki.fi
2008-02-15
We study the role of scale dependence in the Buchert averaging method, using the flat Lemaitre-Tolman-Bondi model as a testing ground. Within this model, a single averaging scale gives predictions that are too coarse, but by replacing it with the distance of the objects R(z) for each redshift z, we find an O(1%) precision at z<2 in the averaged luminosity and angular diameter distances compared to their exact expressions. At low redshifts, we show the improvement for generic inhomogeneity profiles, and our numerical computations further verify it up to redshifts z{approx}2. At higher redshifts, the method breaks down due to its inability to capture the time evolution of the inhomogeneities. We also demonstrate that the running smoothing scale R(z) can mimic acceleration, suggesting that it could be at least as important as the backreaction in explaining dark energy as an inhomogeneity induced illusion.
Books average previous decade of economic misery.
Bentley, R Alexander; Acerbi, Alberto; Ormerod, Paul; Lampos, Vasileios
2014-01-01
For the 20(th) century since the Depression, we find a strong correlation between a 'literary misery index' derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade. PMID:24416159
Averaged energy conditions and evaporating black holes
Ford, L.H.; Roman, T.A.
1996-02-01
In this paper the averaged weak and averaged null energy conditions, together with uncertainty-principle-type restrictions on negative energy ({open_quote}{open_quote}quantum inequalities{close_quote}{close_quote}), are examined in the context of evaporating black hole backgrounds in both two and four dimensions. In particular, integrals over only half-geodesics are studied. We determine the regions of the spacetime in which the averaged energy conditions are violated. In all cases where these conditions fail, there appear to be quantum inequalities which bound the magnitude and extent of the negative energy, and hence the degree of the violation. The possible relevance of these results for the validity of singularity theorems in evaporating black hole spacetimes is discussed. {copyright} {ital 1996 The American Physical Society.}
Books Average Previous Decade of Economic Misery
Bentley, R. Alexander; Acerbi, Alberto; Ormerod, Paul; Lampos, Vasileios
2014-01-01
For the 20th century since the Depression, we find a strong correlation between a ‘literary misery index’ derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade. PMID:24416159
ERIC Educational Resources Information Center
Iona, Mario
1975-01-01
Presents a summary and comparison of various views on the concepts of mass and weight. Includes a consideration of gravitational force in an inertial system and apparent gravitational force on a rotating earth. Discusses the units and methods for measuring mass and weight. (GS)
ERIC Educational Resources Information Center
Lakdawalla, Darius; Philipson, Tomas
2007-01-01
We use panel data from the National Longitudinal Survey of Youth to investigate on-the-job exercise and weight. For male workers, job-related exercise has causal effects on weight, but for female workers, the effects seem primarily selective. A man who spends 18 years in the most physical fitness-demanding occupation is about 25 pounds (14…
... Weight Control A calorie is a unit of energy. Most foods and beverages contain calories. To lose weight you ... Combine the two for the best results The foods you eat and the beverages you drink provide energy and nutrients. The basic required nutrients are: water, ...
ERIC Educational Resources Information Center
Katch, Victor L.
This paper describes a number of factors which go into determining weight. The paper describes what calories are, how caloric expenditure is measured, and why caloric expenditure is different for different people. The paper then outlines the way the body tends to adjust food intake and exercise to maintain a constant body weight. It is speculated…
... with age. The secret to maintaining a healthy weight is to balance “energy in” and “energy out.” Energy in means the calories you get from the food and beverages you consume. Energy ... you be to keep a healthy weight? The answer is diﬀerent for each person, but ...
Technology Transfer Automated Retrieval System (TEKTRAN)
This review evaluated the available scientific literature relative to anthocyanins and weight loss and/or obesity with mention of other effects of anthocyanins on pathologies that are closely related to obesity. Although there is considerable popular press concerning anthocyanins and weight loss, th...
ERIC Educational Resources Information Center
Lakdawalla, Darius; Philipson, Tomas
2007-01-01
We use panel data from the National Longitudinal Survey of Youth to investigate on-the-job exercise and weight. For male workers, job-related exercise has causal effects on weight, but for female workers, the effects seem primarily selective. A man who spends 18 years in the most physical fitness-demanding occupation is about 25 pounds (14
Sansone, Randy A; Sansone, Lori A
2014-07-01
Acute marijuana use is classically associated with snacking behavior (colloquially referred to as "the munchies"). In support of these acute appetite-enhancing effects, several authorities report that marijuana may increase body mass index in patients suffering from human immunodeficiency virus and cancer. However, for these medical conditions, while appetite may be stimulated, some studies indicate that weight gain is not always clinically meaningful. In addition, in a study of cancer patients in which weight gain did occur, it was less than the comparator drug (megestrol). However, data generally suggest that acute marijuana use stimulates appetite, and that marijuana use may stimulate appetite in low-weight individuals. As for large epidemiological studies in the general population, findings consistently indicate that users of marijuana tend to have lower body mass indices than nonusers. While paradoxical and somewhat perplexing, these findings may be explained by various study confounds, such as potential differences between acute versus chronic marijuana use; the tendency for marijuana use to be associated with other types of drug use; and/or the possible competition between food and drugs for the same reward sites in the brain. Likewise, perhaps the effects of marijuana are a function of initial weight status-i.e., maybe marijuana is a metabolic regulatory substance that increases body weight in low-weight individuals but not in normal-weight or overweight individuals. Only further research will clarify the complex relationships between marijuana and body weight. PMID:25337447
Wire weight is lowered to water surface to measure stage at a site. Levels are made to the wire weights elevation from known benchmarks to ensure correct readings. In the background there is housing protected with dikes along the Missouri River in Mandan, ND....
... in a person's diabetes management plan. Weight and Type 1 Diabetes If a person has type 1 diabetes but hasn't been treated yet, he or she often loses weight. In type 1 diabetes, the body can't use glucose (pronounced: GLOO- ...
SOURCE TERMS FOR AVERAGE DOE SNF CANISTERS
K. L. Goluoglu
2000-06-09
The objective of this calculation is to generate source terms for each type of Department of Energy (DOE) spent nuclear fuel (SNF) canister that may be disposed of at the potential repository at Yucca Mountain. The scope of this calculation is limited to generating source terms for average DOE SNF canisters, and is not intended to be used for subsequent calculations requiring bounding source terms. This calculation is to be used in future Performance Assessment calculations, or other shielding or thermal calculations requiring average source terms.
An improved moving average technical trading rule
NASA Astrophysics Data System (ADS)
Papailias, Fotis; Thomakos, Dimitrios D.
2015-06-01
This paper proposes a modified version of the widely used price and moving average cross-over trading strategies. The suggested approach (presented in its 'long only' version) is a combination of cross-over 'buy' signals and a dynamic threshold value which acts as a dynamic trailing stop. The trading behaviour and performance from this modified strategy are different from the standard approach with results showing that, on average, the proposed modification increases the cumulative return and the Sharpe ratio of the investor while exhibiting smaller maximum drawdown and smaller drawdown duration than the standard strategy.
Average: the juxtaposition of procedure and context
NASA Astrophysics Data System (ADS)
Watson, Jane; Chick, Helen; Callingham, Rosemary
2014-09-01
This paper presents recent data on the performance of 247 middle school students on questions concerning average in three contexts. Analysis includes considering levels of understanding linking definition and context, performance across contexts, the relative difficulty of tasks, and difference in performance for male and female students. The outcomes lead to a discussion of the expectations of the curriculum and its implementation, as well as assessment, in relation to students' skills in carrying out procedures and their understanding about the meaning of average in context.
The modulated average structure of mullite.
Birkenstock, Johannes; Pet??ek, Vclav; Pedersen, Bjoern; Schneider, Hartmut; Fischer, Reinhard X
2015-06-01
Homogeneous and inclusion-free single crystals of 2:1 mullite (Al(4.8)Si(1.2)O(9.6)) grown by the Czochralski technique were examined by X-ray and neutron diffraction methods. The observed diffuse scattering together with the pattern of satellite reflections confirm previously published data and are thus inherent features of the mullite structure. The ideal composition was closely met as confirmed by microprobe analysis (Al(4.82?(3))Si(1.18?(1))O(9.59?(5))) and by average structure refinements. 8?(5) to 20?(13)% of the available Si was found in the T* position of the tetrahedra triclusters. The strong tendencey for disorder in mullite may be understood from considerations of hypothetical superstructures which would have to be n-fivefold with respect to the three-dimensional average unit cell of 2:1 mullite and n-fourfold in case of 3:2 mullite. In any of these the possible arrangements of the vacancies and of the tetrahedral units would inevitably be unfavorable. Three directions of incommensurate modulations were determined: q1 = [0.3137?(2) 0 ], q2 = [0 0.4021?(5) 0.1834?(2)] and q3 = [0 0.4009?(5) -0.1834?(2)]. The one-dimensional incommensurately modulated crystal structure associated with q1 was refined for the first time using the superspace approach. The modulation is dominated by harmonic occupational modulations of the atoms in the di- and the triclusters of the tetrahedral units in mullite. The modulation amplitudes are small and the harmonic character implies that the modulated structure still represents an average structure in the overall disordered arrangement of the vacancies and of the tetrahedral structural units. In other words, when projecting the local assemblies at the scale of a few tens of average mullite cells into cells determined by either one of the modulation vectors q1, q2 or q3 a weak average modulation results with slightly varying average occupation factors for the tetrahedral units. As a result, the real structure of mullite is locally ordered (as previously known), but on the long-range its average is not completely disordered, the modulated structure of mullite may be denoted the true 'average structure of mullite'. PMID:26027012
Polarized electron beams at milliampere average current
Poelker, Matthew
2013-11-01
This contribution describes some of the challenges associated with developing a polarized electron source capable of uninterrupted days-long operation at milliAmpere average beam current with polarization greater than 80%. Challenges will be presented in the context of assessing the required level of extrapolation beyond the performance of today's CEBAF polarized source operating at ~ 200 uA average current. Estimates of performance at higher current will be based on hours-long demonstrations at 1 and 4 mA. Particular attention will be paid to beam-related lifetime-limiting mechanisms, and strategies to construct a photogun that operate reliably at bias voltage > 350kV.
Weight discrimination and bullying.
Puhl, Rebecca M; King, Kelly M
2013-04-01
Despite significant attention to the medical impacts of obesity, often ignored are the negative outcomes that obese children and adults experience as a result of stigma, bias, and discrimination. Obese individuals are frequently stigmatized because of their weight in many domains of daily life. Research spanning several decades has documented consistent weight bias and stigmatization in employment, health care, schools, the media, and interpersonal relationships. For overweight and obese youth, weight stigmatization translates into pervasive victimization, teasing, and bullying. Multiple adverse outcomes are associated with exposure to weight stigmatization, including depression, anxiety, low self-esteem, body dissatisfaction, suicidal ideation, poor academic performance, lower physical activity, maladaptive eating behaviors, and avoidance of health care. This review summarizes the nature and extent of weight stigmatization against overweight and obese individuals, as well as the resulting consequences that these experiences create for social, psychological, and physical health for children and adults who are targeted. PMID:23731874
NASA Astrophysics Data System (ADS)
Chang, Shu-Chiuan; Shrock, Robert
2010-02-01
We study two weighted graph coloring problems, in which one assigns q colors to the vertices of a graph such that adjacent vertices have different colors, with a vertex weighting w that either disfavors or favors a given color. We exhibit a weighted chromatic polynomial Ph( G, q, w) associated with this problem that generalizes the chromatic polynomial P( G, q). General properties of this polynomial are proved, and illustrative calculations for various families of graphs are presented. We show that the weighted chromatic polynomial is able to distinguish between certain graphs that yield the same chromatic polynomial. We give a general structural formula for Ph( G, q, w) for lattice strip graphs G with periodic longitudinal boundary conditions. The zeros of Ph( G, q, w) in the q and w planes and their accumulation sets in the limit of infinitely many vertices of G are analyzed. Finally, some related weighted graph coloring problems are mentioned.
Initial Conditions in the Averaging Cognitive Model
ERIC Educational Resources Information Center
Noventa, S.; Massidda, D.; Vidotto, G.
2010-01-01
The initial state parameters s[subscript 0] and w[subscript 0] are intricate issues of the averaging cognitive models in Information Integration Theory. Usually they are defined as a measure of prior information (Anderson, 1981; 1982) but there are no general rules to deal with them. In fact, there is no agreement as to their treatment except in
Bayesian Model Averaging for Propensity Score Analysis
ERIC Educational Resources Information Center
Kaplan, David; Chen, Jianshen
2013-01-01
The purpose of this study is to explore Bayesian model averaging in the propensity score context. Previous research on Bayesian propensity score analysis does not take into account model uncertainty. In this regard, an internally consistent Bayesian framework for model building and estimation must also account for model uncertainty. The…
Why Johnny Can Be Average Today.
ERIC Educational Resources Information Center
Sturrock, Alan
1997-01-01
During a (hypothetical) phone interview with a university researcher, an elementary principal reminisced about a lifetime of reading groups with unmemorable names, medium-paced math problems, patchworked social studies/science lessons, and totally "average" IQ and batting scores. The researcher hung up at the mention of bell-curved assembly lines
Average Annual Rainfall Over the Globe
NASA Astrophysics Data System (ADS)
Agrawal, D. C.
2013-12-01
The atmospheric recycling of water is a very important phenomenon on the globe because it not only refreshes the water but it also redistributes it over land and oceans/rivers/lakes throughout the globe. This is made possible by the solar energy intercepted by the Earth. The half of the globe facing the Sun, on the average, intercepts 1.741017 J of solar radiation per second and it is divided over various channels as given in Table 1. It keeps our planet warm and maintains its average temperature2 of 288 K with the help of the atmosphere in such a way that life can survive. It also recycles the water in the oceans/rivers/ lakes by initial evaporation and subsequent precipitation; the average annual rainfall over the globe is around one meter. According to M. King Hubbert the amount of solar power going into the evaporation and precipitation channel is 4.01016 W. Students can verify the value of average annual rainfall over the globe by utilizing this part of solar energy. This activity is described in the next section.
The periodic average structure of particular quasicrystals.
Steurer; Haibach
1999-01-01
The non-crystallographic symmetry of d-dimensional (dD) quasiperiodic structures is incompatible with lattice periodicity in dD physical space. However, dD quasiperiodic structures can be described as irrational sections of nD (n > d) periodic hypercrystal structures. By appropriate oblique projection of particular hypercrystal structures onto physical space, discrete periodic average structures can be obtained. The boundaries of the projected atomic surfaces give the maximum distance of each atom in a quasiperiodic structure from the vertices of the reference lattice of its average structure. These maximum distances turn out to be smaller than even the shortest atomic bond lengths. The metrics of the average structure of a 3D Ammann tiling, for instance, with edge lengths of the unit tiles equal to the bond lengths in elemental aluminium, correspond almost exactly to the metrics of face-centred-cubic aluminium. This is remarkable since most stable quasicrystals contain aluminium as the main constitutent. The study of the average structure of quasicrystals can be a valuable aid to the elucidation of the geometry of quasicrystal-to-crystal transformations. It can also contribute to the derivation of the physically most relevant Brillouin (Jones) zone. PMID:10927229
A Functional Measurement Study on Averaging Numerosity
ERIC Educational Resources Information Center
Tira, Michael D.; Tagliabue, Mariaelena; Vidotto, Giulio
2014-01-01
In two experiments, participants judged the average numerosity between two sequentially presented dot patterns to perform an approximate arithmetic task. In Experiment 1, the response was given on a 0-20 numerical scale (categorical scaling), and in Experiment 2, the response was given by the production of a dot pattern of the desired numerosity
Measuring Time-Averaged Blood Pressure
NASA Technical Reports Server (NTRS)
Rothman, Neil S.
1988-01-01
Device measures time-averaged component of absolute blood pressure in artery. Includes compliant cuff around artery and external monitoring unit. Ceramic construction in monitoring unit suppresses ebb and flow of pressure-transmitting fluid in sensor chamber. Transducer measures only static component of blood pressure.
Average Annual Rainfall over the Globe
ERIC Educational Resources Information Center
Agrawal, D. C.
2013-01-01
The atmospheric recycling of water is a very important phenomenon on the globe because it not only refreshes the water but it also redistributes it over land and oceans/rivers/lakes throughout the globe. This is made possible by the solar energy intercepted by the Earth. The half of the globe facing the Sun, on the average, intercepts 1.74 ׅ
Glenzinski, D.; /Fermilab
2008-01-01
This paper summarizes a talk given at the Top2008 Workshop at La Biodola, Isola d Elba, Italy. The status of the world average top-quark mass is discussed. Some comments about the challanges facing the experiments in order to further improve the precision are offered.
Why Johnny Can Be Average Today.
ERIC Educational Resources Information Center
Sturrock, Alan
1997-01-01
During a (hypothetical) phone interview with a university researcher, an elementary principal reminisced about a lifetime of reading groups with unmemorable names, medium-paced math problems, patchworked social studies/science lessons, and totally "average" IQ and batting scores. The researcher hung up at the mention of bell-curved assembly lines…
HIGH AVERAGE POWER OPTICAL FEL AMPLIFIERS.
BEN-ZVI, ILAN, DAYRAN, D.; LITVINENKO, V.
2005-08-21
Historically, the first demonstration of the optical FEL was in an amplifier configuration at Stanford University [l]. There were other notable instances of amplifying a seed laser, such as the LLNL PALADIN amplifier [2] and the BNL ATF High-Gain Harmonic Generation FEL [3]. However, for the most part FELs are operated as oscillators or self amplified spontaneous emission devices. Yet, in wavelength regimes where a conventional laser seed can be used, the FEL can be used as an amplifier. One promising application is for very high average power generation, for instance FEL's with average power of 100 kW or more. The high electron beam power, high brightness and high efficiency that can be achieved with photoinjectors and superconducting Energy Recovery Linacs (ERL) combine well with the high-gain FEL amplifier to produce unprecedented average power FELs. This combination has a number of advantages. In particular, we show that for a given FEL power, an FEL amplifier can introduce lower energy spread in the beam as compared to a traditional oscillator. This properly gives the ERL based FEL amplifier a great wall-plug to optical power efficiency advantage. The optics for an amplifier is simple and compact. In addition to the general features of the high average power FEL amplifier, we will look at a 100 kW class FEL amplifier is being designed to operate on the 0.5 ampere Energy Recovery Linac which is under construction at Brookhaven National Laboratory's Collider-Accelerator Department.
Averaging on Earth-Crossing Orbits
NASA Astrophysics Data System (ADS)
Gronchi, G. F.; Milani, A.
The orbits of planet-crossing asteroids (and comets) can undergo close approaches and collisions with some major planet. This introduces a singularity in the N-body Hamiltonian, and the averaging of the equations of motion, traditionally used to compute secular perturbations, is undefined. We show that it is possible to define in a rigorous way some generalised averaged equations of motion, in such a way that the generalised solutions are unique and piecewise smooth. This is obtained, both in the planar and in the three-dimensional case, by means of the method of extraction of the singularities by Kantorovich. The modified distance used to approximate the singularity is the one used by Wetherill in his method to compute probability of collision. Some examples of averaged dynamics have been computed; a systematic exploration of the averaged phase space to locate the secular resonances should be the next step. `Alice sighed wearily. ``I think you might do something better with the time'' she said, ``than waste it asking riddles with no answers'' (Alice in Wonderland, L. Carroll)
Fonov, V S; Le Troter, A; Taso, M; De Leener, B; Lvque, G; Benhamou, M; Sdika, M; Benali, H; Pradat, P-F; Collins, D L; Callot, V; Cohen-Adad, J
2014-11-15
The field of spinal cord MRI is lacking a common template, as existing for the brain, which would allow extraction of multi-parametric data (diffusion-weighted, magnetization transfer, etc.) without user bias, thereby facilitating group analysis and multi-center studies. This paper describes a framework to produce an unbiased average anatomical template of the human spinal cord. The template was created by co-registering T2-weighted images (N = 16 healthy volunteers) using a series of pre-processing steps followed by non-linear registration. A white and gray matter probabilistic template was then merged to the average anatomical template, yielding the MNI-Poly-AMU template, which currently covers vertebral levels C1 to T6. New subjects can be registered to the template using a dedicated image processing pipeline. Validation was conducted on 16 additional subjects by comparing an automatic template-based segmentation and manual segmentation, yielding a median Dice coefficient of 0.89. The registration pipeline is rapid (~15 min), automatic after one C2/C3 landmark manual identification, and robust, thereby reducing subjective variability and bias associated with manual segmentation. The template can notably be used for measurements of spinal cord cross-sectional area, voxel-based morphometry, identification of anatomical features (e.g., vertebral levels, white and gray matter location) and unbiased extraction of multi-parametric data. PMID:25204864
Orbit Averaging in Perturbed Planetary Rings
NASA Astrophysics Data System (ADS)
Stewart, Glen R.
2015-11-01
The orbital period is typically much shorter than the time scale for dynamical evolution of large-scale structures in planetary rings. This large separation in time scales motivates the derivation of reduced models by averaging the equations of motion over the local orbit period (Borderies et al. 1985, Shu et al. 1985). A more systematic procedure for carrying out the orbit averaging is to use Lie transform perturbation theory to remove the dependence on the fast angle variable from the problem order-by-order in epsilon, where the small parameter epsilon is proportional to the fractional radial distance from exact resonance. This powerful technique has been developed and refined over the past thirty years in the context of gyrokinetic theory in plasma physics (Brizard and Hahm, Rev. Mod. Phys. 79, 2007). When the Lie transform method is applied to resonantly forced rings near a mean motion resonance with a satellite, the resulting orbit-averaged equations contain the nonlinear terms found previously, but also contain additional nonlinear self-gravity terms of the same order that were missed by Borderies et al. and by Shu et al. The additional terms result from the fact that the self-consistent gravitational potential of the perturbed rings modifies the orbit-averaging transformation at nonlinear order. These additional terms are the gravitational analog of electrostatic ponderomotive forces caused by large amplitude waves in plasma physics. The revised orbit-averaged equations are shown to modify the behavior of nonlinear density waves in planetary rings compared to the previously published theory. This reserach was supported by NASA's Outer Planets Reserach program.
A weighted and directed interareal connectivity matrix for macaque cerebral cortex.
Markov, N T; Ercsey-Ravasz, M M; Ribeiro Gomes, A R; Lamy, C; Magrou, L; Vezoli, J; Misery, P; Falchier, A; Quilodran, R; Gariel, M A; Sallet, J; Gamanut, R; Huissoud, C; Clavagnier, S; Giroud, P; Sappey-Marinier, D; Barone, P; Dehay, C; Toroczkai, Z; Knoblauch, K; Van Essen, D C; Kennedy, H
2014-01-01
Retrograde tracer injections in 29 of the 91 areas of the macaque cerebral cortex revealed 1,615 interareal pathways, a third of which have not previously been reported. A weight index (extrinsic fraction of labeled neurons [FLNe]) was determined for each area-to-area pathway. Newly found projections were weaker on average compared with the known projections; nevertheless, the 2 sets of pathways had extensively overlapping weight distributions. Repeat injections across individuals revealed modest FLNe variability given the range of FLNe values (standard deviation <1 log unit, range 5 log units). The connectivity profile for each area conformed to a lognormal distribution, where a majority of projections are moderate or weak in strength. In the G29 × 29 interareal subgraph, two-thirds of the connections that can exist do exist. Analysis of the smallest set of areas that collects links from all 91 nodes of the G29 × 91 subgraph (dominating set analysis) confirms the dense (66%) structure of the cortical matrix. The G29 × 29 subgraph suggests an unexpectedly high incidence of unidirectional links. The directed and weighted G29 × 91 connectivity matrix for the macaque will be valuable for comparison with connectivity analyses in other species, including humans. It will also inform future modeling studies that explore the regularities of cortical networks. PMID:23010748
A Weighted and Directed Interareal Connectivity Matrix for Macaque Cerebral Cortex
Markov, N. T.; Ercsey-Ravasz, M. M.; Ribeiro Gomes, A. R.; Lamy, C.; Magrou, L.; Vezoli, J.; Misery, P.; Falchier, A.; Quilodran, R.; Gariel, M. A.; Sallet, J.; Gamanut, R.; Huissoud, C.; Clavagnier, S.; Giroud, P.; Sappey-Marinier, D.; Barone, P.; Dehay, C.; Toroczkai, Z.; Knoblauch, K.; Van Essen, D. C.; Kennedy, H.
2014-01-01
Retrograde tracer injections in 29 of the 91 areas of the macaque cerebral cortex revealed 1,615 interareal pathways, a third of which have not previously been reported. A weight index (extrinsic fraction of labeled neurons [FLNe]) was determined for each area-to-area pathway. Newly found projections were weaker on average compared with the known projections; nevertheless, the 2 sets of pathways had extensively overlapping weight distributions. Repeat injections across individuals revealed modest FLNe variability given the range of FLNe values (standard deviation <1 log unit, range 5 log units). The connectivity profile for each area conformed to a lognormal distribution, where a majority of projections are moderate or weak in strength. In the G29 × 29 interareal subgraph, two-thirds of the connections that can exist do exist. Analysis of the smallest set of areas that collects links from all 91 nodes of the G29 × 91 subgraph (dominating set analysis) confirms the dense (66%) structure of the cortical matrix. The G29 × 29 subgraph suggests an unexpectedly high incidence of unidirectional links. The directed and weighted G29 × 91 connectivity matrix for the macaque will be valuable for comparison with connectivity analyses in other species, including humans. It will also inform future modeling studies that explore the regularities of cortical networks. PMID:23010748
Clinical and genetic predictors of weight gain in patients diagnosed with breast cancer
Reddy, S M; Sadim, M; Li, J; Yi, N; Agarwal, S; Mantzoros, C S; Kaklamani, V G
2013-01-01
Background: Post-diagnosis weight gain in breast cancer patients has been associated with increased cancer recurrence and mortality. Our study was designed to identify risk factors for this weight gain and create a predictive model to identify a high-risk population for targeted interventions. Methods: Chart review was conducted on 459 breast cancer patients from Northwestern Robert H. Lurie Cancer Centre to obtain weights and body mass indices (BMIs) over an 18-month period from diagnosis. We also recorded tumour characteristics, demographics, clinical factors, and treatment regimens. Blood samples were genotyped for 14 single-nucleotide polymorphisms (SNPs) in fat mass and obesity-associated protein (FTO) and adiponectin pathway genes (ADIPOQ and ADIPOR1). Results: In all, 56% of patients had >0.5 kg m–2 increase in BMI from diagnosis to 18 months, with average BMI and weight gain of 1.9 kg m–2 and 5.1 kg, respectively. Our best predictive model was a primarily SNP-based model incorporating all 14 FTO and adiponectin pathway SNPs studied, their epistatic interactions, and age and BMI at diagnosis, with area under receiver operating characteristic curve of 0.85 for 18-month weight gain. Conclusion: We created a powerful risk prediction model that can identify breast cancer patients at high risk for weight gain. PMID:23922112
Blanc, Ann K.; Wardlaw, Tessa
2005-01-01
OBJECTIVE: To critically examine the data used to produce estimates of the proportion of infants with low birth weight in developing countries and to describe biases in these data. To assess the effect of adjustment procedures on the estimates and propose a modified estimation procedure for international reporting purposes. METHODS: Mothers' reports about their recent births in 62 nationally representative Demographic and Health Surveys (DHS) conducted between 1990 and 2000 were analysed. The proportion of infants weighed at birth, characteristics of those weighed, extent of misreporting, and mothers' subjective assessments of their children's size at birth were examined. FINDINGS: In many developing countries the majority of infants were not weighed at birth. Those who were weighed were more likely to have mothers who live in urban areas and are educated, and to be born in a medical facility with assistance from medically trained personnel. Birth weights reported by mothers are "heaped" on multiples of 500 grams. CONCLUSION: Current survey-based estimates of the prevalence of low birth weight are biased substantially downwards. Two adjustments to reported data are recommended: a weighting procedure that combines reported birth weights with mothers' assessment of the child's size at birth, and categorization of one-quarter of the infants reported to have a birth weight of exactly 2500 grams as having low birth weight. Averaged over all surveys, these procedures increased the proportion classified as having low birth weight by 25%. We also recommend that the proportion of infants not weighed at birth be routinely reported. Efforts are needed to increase the weighing of newborns and the recording of their weights. PMID:15798841
Hwang, Yunji; Lee, Kyu Eun; Park, Young Joo; Kim, Su-Jin; Kwon, Hyungju; Park, Do Joon; Cho, Belong; Choi, Ho-Chun; Kang, Daehee; Park, Sue K
2016-03-01
We evaluated the association between weight change in middle-aged adults and papillary thyroid cancer (PTC) based on a large-scale case-control study.Our study included data from 1551 PTC patients (19.3% men and 80.7% women) who underwent thyroidectomy at the 3 general hospitals in Korea and 15,510 individually matched control subjects. The subjects' weight history, epidemiologic information, and tumor characteristics confirmed after thyroidectomy were analyzed. Odds ratios (ORs) and 95% confidence intervals (95% CIs) were determined for the annual average changes in weight and obesity indicators (body mass index (BMI), body surface area, and body fat percentage (BF%) in subjects since the age of 35 years.Subjects with a total weight gain ≥10 kg after age 35 years were more likely to have PTC (men, OR, 5.39, 95% CI, 3.88-7.49; women, OR, 3.36, 95% CI, 2.87-3.93) compared with subjects with a stable weight (loss or gain <5 kg). A marked increase in BMI since age 35 years (annual average change of BMI ≥0.3 kg/m/yr) was related to an elevated PTC risk, and the association was more pronounced for large-sized PTC risks (<1 cm, OR, 2.34, 95% CI, 1.92-2.85; ≥1 cm, OR, 4.00, 95% CI, 2.91-5.49, P heterogeneity = 0.005) compared with low PTC risks.Weight gain and annual increases in obesity indicators in middle-aged adults may increase the risk of developing PTC. PMID:26945379
Englberger, L.
1999-01-01
A programme of weight loss competitions and associated activities in Tonga, intended to combat obesity and the noncommunicable diseases linked to it, has popular support and the potential to effect significant improvements in health. PMID:10063662
... that spark physical changes like breast development in girls, testicular enlargement in boys, and spurts in height and weight gain in both boys and girls. Once these changes start, they continue for several ...
... Differences in BMRs are associated with changes in energy balance. Energy balance reflects the difference between the amount of ... such as amphetamines, animals often have a negative energy balance which leads to weight loss. Based on ...
Average entanglement for Markovian quantum trajectories
Vogelsberger, S.; Spehner, D.
2010-11-15
We study the evolution of the entanglement of noninteracting qubits coupled to reservoirs under monitoring of the reservoirs by means of continuous measurements. We calculate the average of the concurrence of the qubits wave function over all quantum trajectories. For two qubits coupled to independent baths subjected to local measurements, this average decays exponentially with a rate depending on the measurement scheme only. This contrasts with the known disappearance of entanglement after a finite time for the density matrix in the absence of measurements. For two qubits coupled to a common bath, the mean concurrence can vanish at discrete times. Our analysis applies to arbitrary quantum jump or quantum state diffusion dynamics in the Markov limit. We discuss the best measurement schemes to protect entanglement in specific examples.
Rigid shape matching by segmentation averaging.
Wang, Hongzhi; Oliensis, John
2010-04-01
We use segmentations to match images by shape. The new matching technique does not require point-to-point edge correspondence and is robust to small shape variations and spatial shifts. To address the unreliability of segmentations computed bottom-up, we give a closed form approximation to an average over all segmentations. Our method has many extensions, yielding new algorithms for tracking, object detection, segmentation, and edge-preserving smoothing. For segmentation, instead of a maximum a posteriori approach, we compute the "central" segmentation minimizing the average distance to all segmentations of an image. For smoothing, instead of smoothing images based on local structures, we smooth based on the global optimal image structures. Our methods for segmentation, smoothing, and object detection perform competitively, and we also show promising results in shape-based tracking. PMID:20224119
Apparent and average accelerations of the Universe
Bolejko, Krzysztof; Andersson, Lars E-mail: larsa@math.miami.edu
2008-10-15
In this paper we consider the relation between the volume deceleration parameter obtained within the Buchert averaging scheme and the deceleration parameter derived from supernova observation. This work was motivated by recent findings that showed that there are models which despite having {Lambda} = 0 have volume deceleration parameter q{sup vol}<0. This opens the possibility that back-reaction and averaging effects may be used as an interesting alternative explanation to the dark energy phenomenon. We have calculated q{sup vol} in some Lemaitre-Tolman models. For those models which are chosen to be realistic and which fit the supernova data, we find that q{sup vol}>0, while those models which we have been able to find which exhibit q{sup vol}<0 turn out to be unrealistic. This indicates that care must be exercised in relating the deceleration parameter to observations.
Average gluon and quark jet multiplicities
NASA Astrophysics Data System (ADS)
Kotikov, A. V.
2016-01-01
We show the results in [1, 2] for computing the QCD contributions to the scale evolution of average gluon and quark jet multiplicities. The new results came due a recent progress in timelike small-x resummation obtained in the M S factorization scheme. They depend on two nonperturbative parameters with clear and simple physical interpretations. A global fit of these two quantities to all available experimental data sets demonstrates by its goodness how our results solve a longstandig problem of QCD. Including all the available theoretical input within our approach, ?s(5 ) (Mz)= 0.11990.0026 has been obtained in the M S scheme in an approximation equivalent to next-to-next-to-leading order enhanced by the resummations of ln x terms through the NNLL level and of ln Q2 terms by the renormalization group. This result is in excellent agreement with the present world average.
Models of space averaged energetics of plates
NASA Technical Reports Server (NTRS)
Bouthier, O. M.; Bernhard, R. J.
1990-01-01
The analysis of high frequency vibrations in plates is of particular interest in the study of structure borne noise in aircrafts. The current methods of analysis are either too expensive (finite element method) or may have a confidence band wider than desirable (Statistical Energy Analysis). An alternative technique to model the space and time averaged response of structural acoustics problems with enough detail to include all significant mechanisms of energy generation, transmission, and absorption is highly desirable. The focus of this paper is the development of a set of equations which govern the space and time averaged energy density in plates. To solve this equation, a new type of boundary value problem must be treated in terms of energy density variables using energy and intensity boundary conditions. A computer simulation verification study of the energy governing equation is performed. A finite element formulation of the new equations is also implemented and several test cases are analyzed and compared to analytical solutions.
Weight Distribution for Non-binary Cluster LDPC Code Ensemble
NASA Astrophysics Data System (ADS)
Nozaki, Takayuki; Maehara, Masaki; Kasai, Kenta; Sakaniwa, Kohichi
In this paper, we derive the average weight distributions for the irregular non-binary cluster low-density parity-check (LDPC) code ensembles. Moreover, we give the exponential growth rate of the average weight distribution in the limit of large code length. We show that there exist $(2,d_c)$-regular non-binary cluster LDPC code ensembles whose normalized typical minimum distances are strictly positive.
The Average Velocity in a Queue
ERIC Educational Resources Information Center
Frette, Vidar
2009-01-01
A number of cars drive along a narrow road that does not allow overtaking. Each driver has a certain maximum speed at which he or she will drive if alone on the road. As a result of slower cars ahead, many cars are forced to drive at speeds lower than their maximum ones. The average velocity in the queue offers a non-trivial example of a mean
The Average Velocity in a Queue
ERIC Educational Resources Information Center
Frette, Vidar
2009-01-01
A number of cars drive along a narrow road that does not allow overtaking. Each driver has a certain maximum speed at which he or she will drive if alone on the road. As a result of slower cars ahead, many cars are forced to drive at speeds lower than their maximum ones. The average velocity in the queue offers a non-trivial example of a mean…
Stochastic Games with Average Payoff Criterion
Ghosh, M. K.; Bagchi, A.
1998-11-15
We study two-person stochastic games on a Polish state and compact action spaces and with average payoff criterion under a certain ergodicity condition. For the zero-sum game we establish the existence of a value and stationary optimal strategies for both players. For the nonzero-sum case the existence of Nash equilibrium in stationary strategies is established under certain separability conditions.
Average chemical composition of the lunar surface
NASA Technical Reports Server (NTRS)
Turkevich, A. L.
1973-01-01
The available data on the chemical composition of the lunar surface at eleven sites (3 Surveyor, 5 Apollo and 3 Luna) are used to estimate the amounts of principal chemical elements (those present in more than about 0.5% by atom) in average lunar surface material. The terrae of the moon differ from the maria in having much less iron and titanium and appreciably more aluminum and calcium.
Average Annual Rainfall over the Globe
ERIC Educational Resources Information Center
Agrawal, D. C.
2013-01-01
The atmospheric recycling of water is a very important phenomenon on the globe because it not only refreshes the water but it also redistributes it over land and oceans/rivers/lakes throughout the globe. This is made possible by the solar energy intercepted by the Earth. The half of the globe facing the Sun, on the average, intercepts 1.74 ×…
Digital Averaging Phasemeter for Heterodyne Interferometry
NASA Technical Reports Server (NTRS)
Johnson, Donald; Spero, Robert; Shaklan, Stuart; Halverson, Peter; Kuhnert, Andreas
2004-01-01
A digital averaging phasemeter has been built for measuring the difference between the phases of the unknown and reference heterodyne signals in a heterodyne laser interferometer. This phasemeter performs well enough to enable interferometric measurements of distance with accuracy of the order of 100 pm and with the ability to track distance as it changes at a speed of as much as 50 cm/s. This phasemeter is unique in that it is a single, integral system capable of performing three major functions that, heretofore, have been performed by separate systems: (1) measurement of the fractional-cycle phase difference, (2) counting of multiple cycles of phase change, and (3) averaging of phase measurements over multiple cycles for improved resolution. This phasemeter also offers the advantage of making repeated measurements at a high rate: the phase is measured on every heterodyne cycle. Thus, for example, in measuring the relative phase of two signals having a heterodyne frequency of 10 kHz, the phasemeter would accumulate 10,000 measurements per second. At this high measurement rate, an accurate average phase determination can be made more quickly than is possible at a lower rate.
Disk-averaged synthetic spectra of Mars
NASA Technical Reports Server (NTRS)
Tinetti, Giovanna; Meadows, Victoria S.; Crisp, David; Fong, William; Velusamy, Thangasamy; Snively, Heather
2005-01-01
The principal goal of the NASA Terrestrial Planet Finder (TPF) and European Space Agency's Darwin mission concepts is to directly detect and characterize extrasolar terrestrial (Earthsized) planets. This first generation of instruments is expected to provide disk-averaged spectra with modest spectral resolution and signal-to-noise. Here we use a spatially and spectrally resolved model of a Mars-like planet to study the detectability of a planet's surface and atmospheric properties from disk-averaged spectra. We explore the detectability as a function of spectral resolution and wavelength range, for both the proposed visible coronograph (TPFC) and mid-infrared interferometer (TPF-I/Darwin) architectures. At the core of our model is a spectrum-resolving (line-by-line) atmospheric/surface radiative transfer model. This model uses observational data as input to generate a database of spatially resolved synthetic spectra for a range of illumination conditions and viewing geometries. The model was validated against spectra recorded by the Mars Global Surveyor-Thermal Emission Spectrometer and the Mariner 9-Infrared Interferometer Spectrometer. Results presented here include disk-averaged synthetic spectra, light curves, and the spectral variability at visible and mid-infrared wavelengths for Mars as a function of viewing angle, illumination, and season. We also considered the differences in the spectral appearance of an increasingly ice-covered Mars, as a function of spectral resolution, signal-to-noise and integration time for both TPF-C and TPFI/ Darwin.
Average System Cost Methodology : Administrator's Record of Decision.
United States. Bonneville Power Administration.
1984-06-01
Significant features of average system cost (ASC) methodology adopted are: retention of the jurisdictional approach where retail rate orders of regulartory agencies provide primary data for computing the ASC for utilities participating in the residential exchange; inclusion of transmission costs; exclusion of construction work in progress; use of a utility's weighted cost of debt securities; exclusion of income taxes; simplification of separation procedures for subsidized generation and transmission accounts from other accounts; clarification of ASC methodology rules; more generous review timetable for individual filings; phase-in of reformed methodology; and each exchanging utility must file under the new methodology within 20 days of implementation by the Federal Energy Regulatory Commission of the ten major participating utilities, the revised ASC will substantially only affect three. (PSB)
The Economic Impact of Weight Regain
Sheppard, Caroline E.; Lester, Erica L. W.; Chuck, Anderson W.; Birch, Daniel W.; Karmali, Shahzeer; de Gara, Christopher J.
2013-01-01
Background. Obesity is well known for being associated with significant economic repercussions. Bariatric surgery is the only evidence-based solution to this problem as well as a cost-effective method of addressing the concern. Numerous authors have calculated the cost effectiveness and cost savings of bariatric surgery; however, to date the economic impact of weight regain as a component of overall cost has not been addressed. Methods. The literature search was conducted to elucidate the direct costs of obesity and primary bariatric surgery, the rate of weight recidivism and surgical revision, and any costs therein. Results. The quoted cost of obesity in Canada was $2.0 billion–$6.7 billion in 2013 CAD. The median percentage of bariatric procedures that fail due to weight gain or insufficient weight loss is 20% (average: 21.1% ± 10.1%, range: 5.2–39, n = 10). Revision of primary surgeries on average ranges from 2.5% to 18.4%, and depending on the procedure accounts for an additional cost between $14,000 and $50,000 USD per patient. Discussion. There was a significant deficit of the literature pertaining to the cost of revision surgery as compared with primary bariatric surgery. As such, the cycle of weight recidivism and bariatric revisions has not as of yet been introduced into any previous cost analysis of bariatric surgery. PMID:24454339
The economic impact of weight regain.
Sheppard, Caroline E; Lester, Erica L W; Chuck, Anderson W; Birch, Daniel W; Karmali, Shahzeer; de Gara, Christopher J
2013-01-01
Background. Obesity is well known for being associated with significant economic repercussions. Bariatric surgery is the only evidence-based solution to this problem as well as a cost-effective method of addressing the concern. Numerous authors have calculated the cost effectiveness and cost savings of bariatric surgery; however, to date the economic impact of weight regain as a component of overall cost has not been addressed. Methods. The literature search was conducted to elucidate the direct costs of obesity and primary bariatric surgery, the rate of weight recidivism and surgical revision, and any costs therein. Results. The quoted cost of obesity in Canada was $2.0 billion-$6.7 billion in 2013 CAD. The median percentage of bariatric procedures that fail due to weight gain or insufficient weight loss is 20% (average: 21.1% ± 10.1%, range: 5.2-39, n = 10). Revision of primary surgeries on average ranges from 2.5% to 18.4%, and depending on the procedure accounts for an additional cost between $14,000 and $50,000 USD per patient. Discussion. There was a significant deficit of the literature pertaining to the cost of revision surgery as compared with primary bariatric surgery. As such, the cycle of weight recidivism and bariatric revisions has not as of yet been introduced into any previous cost analysis of bariatric surgery. PMID:24454339
Ensemble Averaged Probability Density Function (APDF) for Compressible Turbulent Reacting Flows
NASA Technical Reports Server (NTRS)
Shih, Tsan-Hsing; Liu, Nan-Suey
2012-01-01
In this paper, we present a concept of the averaged probability density function (APDF) for studying compressible turbulent reacting flows. The APDF is defined as an ensemble average of the fine grained probability density function (FG-PDF) with a mass density weighting. It can be used to exactly deduce the mass density weighted, ensemble averaged turbulent mean variables. The transport equation for APDF can be derived in two ways. One is the traditional way that starts from the transport equation of FG-PDF, in which the compressible Navier- Stokes equations are embedded. The resulting transport equation of APDF is then in a traditional form that contains conditional means of all terms from the right hand side of the Navier-Stokes equations except for the chemical reaction term. These conditional means are new unknown quantities that need to be modeled. Another way of deriving the transport equation of APDF is to start directly from the ensemble averaged Navier-Stokes equations. The resulting transport equation of APDF derived from this approach appears in a closed form without any need for additional modeling. The methodology of ensemble averaging presented in this paper can be extended to other averaging procedures: for example, the Reynolds time averaging for statistically steady flow and the Reynolds spatial averaging for statistically homogeneous flow. It can also be extended to a time or spatial filtering procedure to construct the filtered density function (FDF) for the large eddy simulation (LES) of compressible turbulent reacting flows.
Importance Ratings and Weighting: Old Concerns and New Perspectives
ERIC Educational Resources Information Center
Russell, Lara B.; Hubley, Anita M.
2005-01-01
This article describes key concepts, reviews empirical findings, and discusses important issues related to the use of subjective importance ratings and importance weighting. The review of empirical findings focuses on weighting achieved via the multiplicative model and on 3 areas in which weighting is commonly used: quality of life, self-esteem,…
NASA Astrophysics Data System (ADS)
Noda, H.; Lapusta, N.; Kanamori, H.
2010-12-01
Static stress drop is often estimated using the seismic moment and rupture area based on a model for uniform stress drop distribution; we denote this estimate by ??_M. ??_M is sometimes interpreted as the spatial average of stress change over the ruptured area, denoted here as ??_A, and used accordingly, for example, to discuss the relation between recurrence interval and the healing of the frictional surface in a system with one degree of freedom [e.g., Marone, 1998]. ??_M is also used to estimate available energy (defined as the strain energy change computed using the final stress state as the reference one) and radiation efficiency [e.g., Venkataraman and Kanamori, 2004]. In this work, we define a stress drop measure, ??_E, that would enter the exact computation of available energy and radiation efficiency. The three stress drop measures - ??_M that can be estimated from observations, ??_A, and ??_E - are equal if the static stress change is spatially uniform, and that motivates substituting ??_M for the other two quantities in applications. However, finite source inversions suggest that the stress change is heterogeneous in natural earthquakes [e.g., Bouchon, 1997]. Since ??_M is the average of stress change weighted by slip distribution due to a uniform stress drop [Madariaga, 1979], ??_E is the average of stress change weighted by actual slip distribution in the event (this work), and ??_A is the simple spatial average of stress change, the three measures should, in general, be different. Here, we investigate the effect of heterogeneity aiming to understand how to use the seismological estimates of stress drop appropriately. We create heterogeneous slip distributions for both circular and rectangular planar ruptures using the approach motivated by Liu-Zeng et al. [2005] and Lavalle et al [2005]. We find that, indeed, the three stress drop measures differ in our scenarios. In particular, heterogeneity increases ??_E and thus the available energy when the seismic moment (and hence ??_M) is preserved. So using ??_M instead of ??_E would underestimate available energy and hence overestimate radiation efficiency. For a range of parameters, ??_E is well-approximated by the seismic estimate ??_M if the latter is computed using a modified (decreased) rupture area that excludes low-slipped regions; a qualitatively similar procedure is already being used in practice [Somerville et al, 1999].
On the Weighted Mean Value Theorem for Integrals
ERIC Educational Resources Information Center
Polezzi, M.
2006-01-01
The Mean Value Theorem for Integrals is a powerful tool, which can be used to prove the Fundamental Theorem of Calculus, and to obtain the average value of a function on an interval. On the other hand, its weighted version is very useful for evaluating inequalities for definite integrals. This article shows the solutions on applying the weighted
Evaluation of a Viscosity-Molecular Weight Relationship.
ERIC Educational Resources Information Center
Mathias, Lon J.
1983-01-01
Background information, procedures, and results are provided for a series of graduate/undergraduate polymer experiments. These include synthesis of poly(methylmethacrylate), viscosity experiment (indicating large effect even small amounts of a polymer may have on solution properties), and measurement of weight-average molecular weight by light
76 FR 19275 - Passenger Weight and Inspected Vessel Stability Requirements
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-07
...On December 14, 2010, the Coast Guard amended its regulations governing the maximum weight and number of passengers that may safely be permitted on board a vessel and other stability regulations, including increasing the Assumed Average Weight per Person (AAWPP) to 185 lb. The amendment triggered new information collection requirements affecting documentation needed from certain inspected......
Evaluation of a Viscosity-Molecular Weight Relationship.
ERIC Educational Resources Information Center
Mathias, Lon J.
1983-01-01
Background information, procedures, and results are provided for a series of graduate/undergraduate polymer experiments. These include synthesis of poly(methylmethacrylate), viscosity experiment (indicating large effect even small amounts of a polymer may have on solution properties), and measurement of weight-average molecular weight by light…
Parents' Reactions to Finding Out That Their Children Have Average or above Average IQ Scores.
ERIC Educational Resources Information Center
Dirks, Jean; And Others
1983-01-01
Parents of 41 children who had been given an individually-administered intelligence test were contacted 19 months after testing. Parents of average IQ children were less accurate in their memory of test results. Children with above average IQ experienced extremely low frequencies of sibling rivalry, conceit or pressure. (Author/HLM)
Estimates of Random Error in Satellite Rainfall Averages
NASA Technical Reports Server (NTRS)
Bell, Thomas L.; Kundu, Prasun K.
2003-01-01
Satellite rain estimates are most accurate when obtained with microwave instruments on low earth-orbiting satellites. Estimation of daily or monthly total areal rainfall, typically of interest to hydrologists and climate researchers, is made difficult, however, by the relatively poor coverage generally available from such satellites. Intermittent coverage by the satellites leads to random "sampling error" in the satellite products. The inexact information about hydrometeors inferred from microwave data also leads to random "retrieval errors" in the rain estimates. In this talk we will review approaches to quantitative estimation of the sampling error in area/time averages of satellite rain retrievals using ground-based observations, and methods of estimating rms random error, both sampling and retrieval, in averages using satellite measurements themselves.
Hydraulic Conductivity Estimation using Bayesian Model Averaging and Generalized Parameterization
NASA Astrophysics Data System (ADS)
Tsai, F. T.; Li, X.
2006-12-01
Non-uniqueness in parameterization scheme is an inherent problem in groundwater inverse modeling due to limited data. To cope with the non-uniqueness problem of parameterization, we introduce a Bayesian Model Averaging (BMA) method to integrate a set of selected parameterization methods. The estimation uncertainty in BMA includes the uncertainty in individual parameterization methods as the within-parameterization variance and the uncertainty from using different parameterization methods as the between-parameterization variance. Moreover, the generalized parameterization (GP) method is considered in the geostatistical framework in this study. The GP method aims at increasing the flexibility of parameterization through the combination of a zonation structure and an interpolation method. The use of BMP with GP avoids over-confidence in a single parameterization method. A normalized least-squares estimation (NLSE) is adopted to calculate the posterior probability for each GP. We employee the adjoint state method for the sensitivity analysis on the weighting coefficients in the GP method. The adjoint state method is also applied to the NLSE problem. The proposed methodology is implemented to the Alamitos Barrier Project (ABP) in California, where the spatially distributed hydraulic conductivity is estimated. The optimal weighting coefficients embedded in GP are identified through the maximum likelihood estimation (MLE) where the misfits between the observed and calculated groundwater heads are minimized. The conditional mean and conditional variance of the estimated hydraulic conductivity distribution using BMA are obtained to assess the estimation uncertainty.
Weight Loss Nutritional Supplements
NASA Astrophysics Data System (ADS)
Eckerson, Joan M.
Obesity has reached what may be considered epidemic proportions in the United States, not only for adults but for children. Because of the medical implications and health care costs associated with obesity, as well as the negative social and psychological impacts, many individuals turn to nonprescription nutritional weight loss supplements hoping for a quick fix, and the weight loss industry has responded by offering a variety of products that generates billions of dollars each year in sales. Most nutritional weight loss supplements are purported to work by increasing energy expenditure, modulating carbohydrate or fat metabolism, increasing satiety, inducing diuresis, or blocking fat absorption. To review the literally hundreds of nutritional weight loss supplements available on the market today is well beyond the scope of this chapter. Therefore, several of the most commonly used supplements were selected for critical review, and practical recommendations are provided based on the findings of well controlled, randomized clinical trials that examined their efficacy. In most cases, the nutritional supplements reviewed either elicited no meaningful effect or resulted in changes in body weight and composition that are similar to what occurs through a restricted diet and exercise program. Although there is some evidence to suggest that herbal forms of ephedrine, such as ma huang, combined with caffeine or caffeine and aspirin (i.e., ECA stack) is effective for inducing moderate weight loss in overweight adults, because of the recent ban on ephedra manufacturers must now use ephedra-free ingredients, such as bitter orange, which do not appear to be as effective. The dietary fiber, glucomannan, also appears to hold some promise as a possible treatment for weight loss, but other related forms of dietary fiber, including guar gum and psyllium, are ineffective.
Auto-exploratory average reward reinforcement learning
Ok, DoKyeong; Tadepalli, P.
1996-12-31
We introduce a model-based average reward Reinforcement Learning method called H-learning and compare it with its discounted counterpart, Adaptive Real-Time Dynamic Programming, in a simulated robot scheduling task. We also introduce an extension to H-learning, which automatically explores the unexplored parts of the state space, while always choosing greedy actions with respect to the current value function. We show that this {open_quotes}Auto-exploratory H-learning{close_quotes} performs better than the original H-learning under previously studied exploration methods such as random, recency-based, or counter-based exploration.
Averaging spherically symmetric spacetimes in general relativity
Coley, A. A.; Pelavas, N.
2006-10-15
We discuss the averaging problem in general relativity, using the form of the macroscopic gravity equations in the case of spherical symmetry in volume preserving coordinates. In particular, we calculate the form of the correlation tensor under some reasonable assumptions on the form for the inhomogeneous gravitational field and matter distribution. On cosmological scales, the correlation tensor in a Friedmann-Lemaitre-Robertson-Walker (FLRW) background is found to be of the form of a spatial curvature. On astrophysical scales the correlation tensor can be interpreted as the sum of a spatial curvature and an anisotropic fluid. We briefly discuss the physical implications of these results.
Quetelet, the average man and medical knowledge.
Caponi, Sandra
2013-01-01
Using two books by Adolphe Quetelet, I analyze his theory of the 'average man', which associates biological and social normality with the frequency with which certain characteristics appear in a population. The books are Sur l'homme et le dveloppement de ses facults and Du systeme social et des lois qui le rgissent. Both reveal that Quetelet's ideas are permeated by explanatory strategies drawn from physics and astronomy, and also by discursive strategies drawn from theology and religion. The stability of the mean as opposed to the dispersion of individual characteristics and events provided the basis for the use of statistics in social sciences and medicine. PMID:23970171
... this figure were adjusted to today’s costs, the price might be an average of less than a ... rich ice cream or chocolate chip cookies, for special occasions and serve appropriate portion sizes when you ...
Predictive RANS simulations via Bayesian Model-Scenario Averaging
Edeling, W.N.; Cinnella, P.; Dwight, R.P.
2014-10-15
The turbulence closure model is the dominant source of error in most Reynolds-Averaged Navier–Stokes simulations, yet no reliable estimators for this error component currently exist. Here we develop a stochastic, a posteriori error estimate, calibrated to specific classes of flow. It is based on variability in model closure coefficients across multiple flow scenarios, for multiple closure models. The variability is estimated using Bayesian calibration against experimental data for each scenario, and Bayesian Model-Scenario Averaging (BMSA) is used to collate the resulting posteriors, to obtain a stochastic estimate of a Quantity of Interest (QoI) in an unmeasured (prediction) scenario. The scenario probabilities in BMSA are chosen using a sensor which automatically weights those scenarios in the calibration set which are similar to the prediction scenario. The methodology is applied to the class of turbulent boundary-layers subject to various pressure gradients. For all considered prediction scenarios the standard-deviation of the stochastic estimate is consistent with the measurement ground truth. Furthermore, the mean of the estimate is more consistently accurate than the individual model predictions.
Averaged null energy condition and quantum inequalities in curved spacetime
NASA Astrophysics Data System (ADS)
Kontou, Eleni-Alexandra
The Averaged Null Energy Condition (ANEC) states that the integral along a complete null geodesic of the projection of the stress-energy tensor onto the tangent vector to the geodesic cannot be negative. ANEC can be used to rule out spacetimes with exotic phenomena, such as closed timelike curves, superluminal travel and wormholes. We prove that ANEC is obeyed by a minimally-coupled, free quantum scalar field on any achronal null geodesic (not two points can be connected with a timelike curve) surrounded by a tubular neighborhood whose curvature is produced by a classical source. To prove ANEC we use a null-projected quantum inequality, which provides constraints on how negative the weighted average of the renormalized stress energy tensor of a quantum field can be. Starting with a general result of Fewster and Smith, we first derive a timelike projected quantum inequality for a minimally-coupled scalar field on flat spacetime with a background potential. Using that result we proceed to find the bound of a quantum inequality on a geodesic in a spacetime with small curvature, working to first order in the Ricci tensor and its derivatives. The last step is to derive a bound for the null-projected quantum inequality on a general timelike path. Finally we use that result to prove achronal ANEC in spacetimes with small curvature.
Women's work. Maintaining a healthy body weight.
Welch, Nicky; Hunter, Wendy; Butera, Karina; Willis, Karen; Cleland, Verity; Crawford, David; Ball, Kylie
2009-08-01
This study describes women's perceptions of the supports and barriers to maintaining a healthy weight among currently healthy weight women from urban and rural socio-economically disadvantaged areas. Using focus groups and interviews, we asked women about their experiences of maintaining a healthy weight. Overwhelmingly, women described their healthy weight practices in terms of concepts related to work and management. The theme of 'managing health' comprised issues of managing multiple responsibilities, time, and emotions associated with healthy practices. Rural women faced particular difficulties in accessing supports at a practical level (for example, lack of childcare) and due to the gendered roles they enacted in caring for others. Family background (in particular, mothers' attitudes to food and weight) also appeared to influence perceptions about healthy weight maintenance. In the context of global increases in the prevalence of obesity, the value of initiatives aimed at supporting healthy weight women to maintain their weight should not be under-estimated. Such initiatives need to work within the social and personal constraints that women face in maintaining good health. PMID:19446587
Analysis of averaged multichannel delay times
NASA Astrophysics Data System (ADS)
Kelkar, N. G.; Nowakowski, M.
2008-07-01
The physical significances and the pros and cons involved in the usage of different time-delay formalisms are discussed. The delay-time matrix introduced by Eisenbud, where only s waves participate in a reaction, is in general related to the definition of an angular time delay which is shown not to be equivalent to the so-called phase time delay of Eisenbud and Wigner even for single channel scattering. Whereas the expression due to Smith which is derived from a time-delayed radial wave packet is consistent with a lifetime matrix which is Hermitian, this is not true for any Eisenbud-type lifetime matrix which violates time-reversal invariance. Extending the angular time delay of Nussenzveig to multiple channels, we show that if one performs an average over the directions and subtracts the forward angle contribution containing an interference of the incident and scattered waves, the multichannel angle-dependent average time delay reduces to the one given by Smith. The present work also rectifies a recently misinterpreted misnomer of the relation due to Smith.
Forecast of the Decadal Average Sunspot Number
NASA Astrophysics Data System (ADS)
Volobuev, D. M.; Makarenko, N. G.
2008-05-01
The forecast of the decadal average sunspot number (SN) becomes possible with an extension of telescopic observations based on proxy reconstructions using the tree ring radiocarbon data during the Holocene. These decadal numbers (SNRC) provide a powerful statistic to verify the forecasting methods. Complicated dynamics of long-term solar activity and noise of proxy-based reconstruction make the one-step-ahead forecast challenging for any forecasting method. Here we construct a continuous data set of SNRC which extends the group sunspot number and the international sunspot number. The known technique of nonlinear forecast, the local linear approximation, is adapted to estimate the coming SN. Both the method and the continuous data set were tested and tuned to obtain the minimum of a normalized average prediction error ( E) during the last millennium using several past millennia as a training data set. E=0.58 ? D is achieved to forecast the SN successive differences whose standard deviation is ? D=7.39 for the period of training. This corresponds to the correlation ( r=0.97) between true and forecasted SN. This error is significantly smaller than the prediction error when the surrogate data were used for the training data set, and proves the nonlinearity in the decadal SN. The estimated coming SN is smaller than the previous one.
Adaptive common average filtering for myocontrol applications.
Rehbaum, Hubertus; Farina, Dario
2015-02-01
The use of electromyography (EMG) for the control of upper-limb prostheses has received great interest in neurorehabilitation engineering since decades. Important advances have been performed in the development of machine learning algorithms for myocontrol. This paper describes a novel adaptive filter for EMG preprocessing to be applied as conditioning stage for optimal subsequent information extraction. The aim of this filter is to improve both the quality (signal-to-noise ratio) and the selectivity of the EMG recordings. The filter is based on the classic common average reference (CAR), often used in EEG processing. However, while CAR is stationary, the proposed filter, which is referred to as adaptive common average reference (ACAR), is signal-dependent and its spatial transfer function is adapted over time. The ACAR filter is evaluated in this study for noise reduction and selectivity. Furthermore, it is proven that its application improves the performance of both pattern recognition and regression methods for myoelectric control. It is concluded that the proposed novel filter for EMG conditioning is a useful preprocessing tool in myocontrol applications. PMID:25388778
Global atmospheric circulation statistics: Four year averages
NASA Technical Reports Server (NTRS)
Wu, M. F.; Geller, M. A.; Nash, E. R.; Gelman, M. E.
1987-01-01
Four year averages of the monthly mean global structure of the general circulation of the atmosphere are presented in the form of latitude-altitude, time-altitude, and time-latitude cross sections. The numerical values are given in tables. Basic parameters utilized include daily global maps of temperature and geopotential height for 18 pressure levels between 1000 and 0.4 mb for the period December 1, 1978 through November 30, 1982 supplied by NOAA/NMC. Geopotential heights and geostrophic winds are constructed using hydrostatic and geostrophic formulae. Meridional and vertical velocities are calculated using thermodynamic and continuity equations. Fields presented in this report are zonally averaged temperature, zonal, meridional, and vertical winds, and amplitude of the planetary waves in geopotential height with zonal wave numbers 1-3. The northward fluxes of sensible heat and eastward momentum by the standing and transient eddies along with their wavenumber decomposition and Eliassen-Palm flux propagation vectors and divergences by the standing and transient eddies along with their wavenumber decomposition are also given. Large interhemispheric differences and year-to-year variations are found to originate in the changes in the planetary wave activity.
Average observational quantities in the timescape cosmology
Wiltshire, David L.
2009-12-15
We examine the properties of a recently proposed observationally viable alternative to homogeneous cosmology with smooth dark energy, the timescape cosmology. In the timescape model cosmic acceleration is realized as an apparent effect related to the calibration of clocks and rods of observers in bound systems relative to volume-average observers in an inhomogeneous geometry in ordinary general relativity. The model is based on an exact solution to a Buchert average of the Einstein equations with backreaction. The present paper examines a number of observational tests which will enable the timescape model to be distinguished from homogeneous cosmologies with a cosmological constant or other smooth dark energy, in current and future generations of dark energy experiments. Predictions are presented for comoving distance measures; H(z); the equivalent of the dark energy equation of state, w(z); the Om(z) measure of Sahni, Shafieloo, and Starobinsky; the Alcock-Paczynski test; the baryon acoustic oscillation measure, D{sub V}; the inhomogeneity test of Clarkson, Bassett, and Lu; and the time drift of cosmological redshifts. Where possible, the predictions are compared to recent independent studies of similar measures in homogeneous cosmologies with dark energy. Three separate tests with indications of results in possible tension with the {lambda}CDM model are found to be consistent with the expectations of the timescape cosmology.
Generalized constructive tree weights
Rivasseau, Vincent E-mail: adrian.tanasa@ens-lyon.org; Tanasa, Adrian E-mail: adrian.tanasa@ens-lyon.org
2014-04-15
The Loop Vertex Expansion (LVE) is a quantum field theory (QFT) method which explicitly computes the Borel sum of Feynman perturbation series. This LVE relies in a crucial way on symmetric tree weights which define a measure on the set of spanning trees of any connected graph. In this paper we generalize this method by defining new tree weights. They depend on the choice of a partition of a set of vertices of the graph, and when the partition is non-trivial, they are no longer symmetric under permutation of vertices. Nevertheless we prove they have the required positivity property to lead to a convergent LVE; in fact we formulate this positivity property precisely for the first time. Our generalized tree weights are inspired by the Brydges-Battle-Federbush work on cluster expansions and could be particularly suited to the computation of connected functions in QFT. Several concrete examples are explicitly given.
Generalized constructive tree weights
NASA Astrophysics Data System (ADS)
Rivasseau, Vincent; Tanasa, Adrian
2014-04-01
The Loop Vertex Expansion (LVE) is a quantum field theory (QFT) method which explicitly computes the Borel sum of Feynman perturbation series. This LVE relies in a crucial way on symmetric tree weights which define a measure on the set of spanning trees of any connected graph. In this paper we generalize this method by defining new tree weights. They depend on the choice of a partition of a set of vertices of the graph, and when the partition is non-trivial, they are no longer symmetric under permutation of vertices. Nevertheless we prove they have the required positivity property to lead to a convergent LVE; in fact we formulate this positivity property precisely for the first time. Our generalized tree weights are inspired by the Brydges-Battle-Federbush work on cluster expansions and could be particularly suited to the computation of connected functions in QFT. Several concrete examples are explicitly given.
Light weight phosphate cements
Wagh, Arun S. (Naperville, IL); Natarajan, Ramkumar, (Woodridge, IL); Kahn, David (Miami, FL)
2010-03-09
A sealant having a specific gravity in the range of from about 0.7 to about 1.6 for heavy oil and/or coal bed methane fields is disclosed. The sealant has a binder including an oxide or hydroxide of Al or of Fe and a phosphoric acid solution. The binder may have MgO or an oxide of Fe and/or an acid phosphate. The binder is present from about 20 to about 50% by weight of the sealant with a lightweight additive present in the range of from about 1 to about 10% by weight of said sealant, a filler, and water sufficient to provide chemically bound water present in the range of from about 9 to about 36% by weight of the sealant when set. A porous ceramic is also disclosed.
Average Strength Parameters of Reactivated Mudstone Landslide for Countermeasure Works
NASA Astrophysics Data System (ADS)
Nakamura, Shinya; Kimura, Sho; Buddhi Vithana, Shriwantha
2015-04-01
Among many approaches to landslide stability analysis, in several landslide-related studies, shear strength parameters obtained from laboratory shear tests have been used with the limit equilibrium method. In most of them, it concluded that the average strength parameters, i.e. average cohesion (c'avg) and average angle of shearing resistance (φ'avg), calculated from back analysis were in agreement with the residual shear strength parameters measured by torsional ring-shear tests on undisturbed and remolded samples. However, disagreement with this contention can be found elsewhere that the residual shear strength measured using a torsional ring-shear apparatus were found to be lower than the average strength calculated by back analysis. One of the reasons why the singular application of residual shear strength in stability analysis causes an underestimation of the safety factor is the fact that the condition of the slip surface of a landslide can be heterogeneous. It may consist of portions that have already reached residual conditions along with other portions that have not on the slip surface. With a view of accommodating such possible differences of slip surface conditions of a landslide, it is worthy to first grasp an appropriate perception of the heterogeneous nature of the actual slip-surface to ensure a more suitable selection of measured shear strength values for stability calculation of landslides. For the present study, the determination procedure of the average strength parameters acting along the slip surface has been presented through the stability calculations of reactivated landslides in the area of Shimajiri-mudstone, Okinawa, Japan. The average strength parameters along slip surfaces of landslides have been estimated using the results of laboratory shear tests of the slip surface/zone soils accompanying a rational way of accessing the actual, heterogeneous slip surface conditions. The results tend to show that the shear strength acting along the slip surface of imperfectly-reactivated landslides cannot always be considered equal to its laboratory-measured residual strength. The engineers should rediscover the fact that it is reasonable to apply different strength parameters to the stability analysis depending on the actual conditions of the slip surface that are visible on the boring core samples. In that context, we suggest to show that it is more appropriate to consider average strength parameters for imperfectly-reactivated landslides, for which purpose the use of 'residual shear strength' in combination with other categories of shear strength is recommended. This way, the outcome of the stability analysis will be much more inclusive and representative of the non-slickensided portions of a slip surface as well.
Weighted Uncertainty Relations
Xiao, Yunlong; Jing, Naihuan; Li-Jost, Xianqing; Fei, Shao-Ming
2016-01-01
Recently, Maccone and Pati have given two stronger uncertainty relations based on the sum of variances and one of them is nontrivial when the quantum state is not an eigenstate of the sum of the observables. We derive a family of weighted uncertainty relations to provide an optimal lower bound for all situations and remove the restriction on the quantum state. Generalization to multi-observable cases is also given and an optimal lower bound for the weighted sum of the variances is obtained in general quantum situation. PMID:26984295
NASA Astrophysics Data System (ADS)
Davis, C. V.; Hill, T. M.; Moffitt, S. E.
2013-12-01
Foraminiferal shell weight can be impacted by environmental factors both during initial shell formation and as the result of post mortem preservation. An improved understanding of what determines this relationship can lead to both an understanding of foraminiferal calcite production in modern oceans and proxy development for past environmental conditions. Significantly, foraminiferal shell weight has been linked to carbonate ion concentration in both laboratory culture (of both planktic and benthic species) and in the modern and fossil record (in planktic foraminifera). This study explores the relationship between shell weight and changes in oxygenation and carbonate saturation in fossil benthic foraminifera from a high-resolution sedimentary record (MV0811-15JC; 3436.930' N, 11912.920' W; 418m water depth; 16.1-3.4 ka; sedimentation rate 44-100 cm kyr-1) from Santa Barbara Basin, CA (SBB). Ongoing work in SBB has described rapid biotic reorganization through the recent deglaciation in response to changes in dissolved oxygen concentrations, which are used here to create a semi quantitative oxygenation history for site MV0811-15JC. In modern Oxygen Minimum Zones, decreases in oxygen closely covary with increases in Total Carbon (with a corresponding decrease in the carbonate saturation state). We interpret that records from SBB of the average size-normalized test weight of Uvigerinid and Bolivinid foraminifera show that shell weight responds to these changes in oxygenation and saturation state. Multiple metrics of 'size normalization' including by length, geometric estimation of surface area and volume, and tracing of individual silhouettes are tested. Regardless of method utilized, the size normalized shell weight of all species fluctuates with abrupt changes in oxygenation and saturation state. Although all species respond to large-scale environmental changes, the weight records of Bolivinids and Uvigerinids reveal distinct differences, indicating that processes governing shell weight may vary across taxonomic groups.
Li, Yongfu; Shen, Hongwei; Lyons, John W; Sammler, Robert L; Brackhagen, Meinolf; Meunier, David M
2016-03-15
Size-exclusion chromatography (SEC) coupled with multi-angle laser light scattering (MALLS) and differential refractive index (DRI) detectors was employed for determination of the molecular weight distributions (MWD) of methylcellulose ethers (MC) and hydroxypropyl methylcellulose ethers (HPMC) having weight-average molecular weights (Mw) ranging from 20 to more than 1,000kg/mol. In comparison to previous work involving right-angle light scattering (RALS) and a viscometer for MWD characterization of MC and HPMC, MALLS yields more reliable molecular weight for materials having weight-average molecular weights (Mw) exceeding about 300kg/mol. A non-ideal SEC separation was observed for cellulose ethers with Mw>800kg/mol, and was manifested by upward divergence of logM vs. elution volume (EV) at larger elution volume at typical SEC flow rate such as 1.0mL/min. As such, the number-average molecular weight (Mn) determined for the sample was erroneously large and polydispersity (Mw/Mn) was erroneously small. This non-ideality resulting in the late elution of high molecular weight chains could be due to the elongation of polymer chains when experimental conditions yield Deborah numbers (De) exceeding 0.5. Non-idealities were eliminated when sufficiently low flow rates were used. Thus, using carefully selected experimental conditions, SEC coupled with MALLS and DRI can provide reliable MWD characterization of MC and HPMC covering the entire ranges of compositions and molecular weights of commercial interest. PMID:26794765
Impact of Field of Study, College and Year on Calculation of Cumulative Grade Point Average
ERIC Educational Resources Information Center
Trail, Carla; Reiter, Harold I.; Bridge, Michelle; Stefanowska, Patricia; Schmuck, Marylou; Norman, Geoff
2008-01-01
A consistent finding from many reviews is that undergraduate Grade Point Average (uGPA) is a key predictor of academic success in medical school. Curiously, while uGPA has established predictive validity, little is known about its reliability. For a variety of reasons, medical schools use different weighting schemas to combine years of study.…
Model averaging methods to merge statistical and dynamic seasonal streamflow forecasts in Australia
NASA Astrophysics Data System (ADS)
Schepen, A.; Wang, Q. J.
2014-12-01
The Australian Bureau of Meteorology operates a statistical seasonal streamflow forecasting service. It has also developed a dynamic seasonal streamflow forecasting approach. The two approaches produce similarly reliable forecasts in terms of ensemble spread but can differ in forecast skill depending on catchment and season. Therefore, it may be possible to augment the skill of the existing service by objectively weighting and merging the forecasts. Bayesian model averaging (BMA) is first applied to merge statistical and dynamic forecasts for 12 locations using leave-five-years-out cross-validation. It is seen that the BMA merged forecasts can sometimes be too uncertain, as shown by ensemble spreads that are unrealistically wide and even bi-modal. The BMA method applies averaging to forecast probability densities (and thus cumulative probabilities) for a given forecast variable value. An alternative approach is quantile model averaging (QMA), whereby forecast variable values (quantiles) are averaged for a given cumulative probability (quantile fraction). For the 12 locations, QMA is compared to BMA. BMA and QMA perform similarly in terms of forecast accuracy skill scores and reliability in terms of ensemble spread. Both methods improve forecast skill across catchments and seasons by combining the different strengths of the statistical and dynamic approaches. A major advantage of QMA over BMA is that it always produces reasonably well defined forecast distributions, even in the special cases where BMA does not. Optimally estimated QMA weights and BMA weights are similar; however, BMA weights are more efficiently estimated.
Orthopedic stretcher with average-sized person can pass through 18-inch opening
NASA Technical Reports Server (NTRS)
Lothschuetz, F. X.
1966-01-01
Modified Robinson stretcher for vertical lifting and carrying, will pass through an opening 18 inches in diameter, while containing a person of average height and weight. A subject 6 feet tall and weighing 200 pounds was lowered and raised out of an 18 inch diameter opening in a tank to test the stretcher.
Impact of Field of Study, College and Year on Calculation of Cumulative Grade Point Average
ERIC Educational Resources Information Center
Trail, Carla; Reiter, Harold I.; Bridge, Michelle; Stefanowska, Patricia; Schmuck, Marylou; Norman, Geoff
2008-01-01
A consistent finding from many reviews is that undergraduate Grade Point Average (uGPA) is a key predictor of academic success in medical school. Curiously, while uGPA has established predictive validity, little is known about its reliability. For a variety of reasons, medical schools use different weighting schemas to combine years of study.
Cleaning Physical Education Areas.
ERIC Educational Resources Information Center
Griffin, William R.
1999-01-01
Discusses techniques to help create clean and inviting school locker rooms. Daily, weekly or monthly, biannual, and annual cleaning strategies for locker room showers are highlighted as are the specialized maintenance needs for aerobic and dance areas, running tracks, and weight training areas. (GR)
Average kinetic energy of the superconducting state
NASA Astrophysics Data System (ADS)
Doria, Mauro M.; Salem-Sugui, S.; de Oliveira, I. G.; Ghivelder, L.; Brandt, E. H.
2002-04-01
Isothermal magnetization curves are plotted as the magnetization times the magnetic induction, 4?M.B, versus the applied field H. We show here that this curve is the average kinetic energy of the superconducting state versus the applied field, for type-II superconductors with a high Ginzburg-Landau parameter ?. The maximum of 4?M.B occurs at a field H*, directly related to the upper critical field Hc2, suggesting that Hc2(T) may be extracted from such plots even in cases when it is too high for direct measurement. We obtain these plots both theoretically, from the Ginzburg-Landau theory, and experimentally, using a Niobium sample with Tc=8.5 K, and compare them.
Simplified optical path-averaged rain gauge.
Wang, T I; Earnshaw, K B; Lawrence, R S
1978-02-01
It has previously been shown that the scintillations produced by raindrops falling through a collimated laser beam can be used to measure the drop-size distribution and the rainfall rate, both averaged over the path. We now present a theoretical analysis, verified by observation, showing that the variance of the scintillations detected by a line-detector measured at frequencies near 1 kHz is closely related to rain rate and is nearly independent of drop-size distribution. If only rain rate is desired, the variance type of optical rain gauge has several advantages over the earlier model. It could use a diverging beam, thus eliminating the practical difficulties of maintaining adjustment and pointing of a collimated beam. Furthermore, it is less sensitive to the presence of updrafts and downdrafts along the beam and can thus be used over rough terrain. PMID:20174419
Average deployments versus missile and defender parameters
Canavan, G.H.
1991-03-01
This report evaluates the average number of reentry vehicles (RVs) that could be deployed successfully as a function of missile burn time, RV deployment times, and the number of space-based interceptors (SBIs) in defensive constellations. Leakage estimates of boost-phase kinetic-energy defenses as functions of launch parameters and defensive constellation size agree with integral predictions of near-exact calculations for constellation sizing. The calculations discussed here test more detailed aspects of the interaction. They indicate that SBIs can efficiently remove about 50% of the RVs from a heavy missile attack. The next 30% can removed with two-fold less effectiveness. The next 10% could double constellation sizes. 5 refs., 7 figs.
High average power linear induction accelerator development
Bayless, J.R.; Adler, R.J.
1987-07-01
There is increasing interest in linear induction accelerators (LIAs) for applications including free electron lasers, high power microwave generators and other types of radiation sources. Lawrence Livermore National Laboratory has developed LIA technology in combination with magnetic pulse compression techniques to achieve very impressive performance levels. In this paper we will briefly discuss the LIA concept and describe our development program. Our goals are to improve the reliability and reduce the cost of LIA systems. An accelerator is presently under construction to demonstrate these improvements at an energy of 1.6 MeV in 2 kA, 65 ns beam pulses at an average beam power of approximately 30 kW. The unique features of this system are a low cost accelerator design and an SCR-switched, magnetically compressed, pulse power system. 4 refs., 7 figs.
Weight and Diabetes (For Parents)
... Cerebral Palsy: Caring for Your Child Weight and Diabetes KidsHealth > For Parents > Weight and Diabetes Print A ... or type 2 diabetes. Weight and Type 1 Diabetes Undiagnosed or untreated, type 1 diabetes can make ...
Brief report: Weight dissatisfaction, weight status, and weight loss in Mexican-American children
Technology Transfer Automated Retrieval System (TEKTRAN)
The study objectives were to assess the association between weight dissatisfaction, weight status, and weight loss in Mexican-American children participating in a weight management program. Participants included 265 Mexican American children recruited for a school-based weight management program. Al...
... to get us off the hook. How Our Minds Influence Weight Loss Why is it so hard to stick with a healthy eating or exercise plan? Why do we slip up or go back to our old ways? A lot of it has to do with the habits we've developed over time. We humans are creatures of habit. We tend to get ...
Menichetti, Giulia; Remondini, Daniel; Panzarasa, Pietro; Mondragón, Raúl J.; Bianconi, Ginestra
2014-01-01
One of the most important challenges in network science is to quantify the information encoded in complex network structures. Disentangling randomness from organizational principles is even more demanding when networks have a multiplex nature. Multiplex networks are multilayer systems of nodes that can be linked in multiple interacting and co-evolving layers. In these networks, relevant information might not be captured if the single layers were analyzed separately. Here we demonstrate that such partial analysis of layers fails to capture significant correlations between weights and topology of complex multiplex networks. To this end, we study two weighted multiplex co-authorship and citation networks involving the authors included in the American Physical Society. We show that in these networks weights are strongly correlated with multiplex structure, and provide empirical evidence in favor of the advantage of studying weighted measures of multiplex networks, such as multistrength and the inverse multiparticipation ratio. Finally, we introduce a theoretical framework based on the entropy of multiplex ensembles to quantify the information stored in multiplex networks that would remain undetected if the single layers were analyzed in isolation. PMID:24906003
ERIC Educational Resources Information Center
Sherman, Rachel M.
1997-01-01
Examines ways of giving an existing weight-training room new life without spending a lot of time and money. Tips include adding rubber floor coverings; using indirect lighting; adding windows, art work, or mirrors to open up the room; using more aesthetically pleasing ceiling tiles; upgrading ventilation; repadding or painting the equipment; and
ERIC Educational Resources Information Center
Nutter, June
1995-01-01
Secondary level physical education teachers can have their students use math concepts while working out on the weight-room equipment. The article explains how students can reinforce math skills while weightlifting by estimating their strength, estimating their power, or calculating other formulas. (SM)
Racial/ethnic differences in impact of gestational weight gain on interconception weight change.
Sackoff, Judith E; Yunzal-Butler, Cristina
2015-06-01
Describe the association between gestational weight gain (GWG) and interconception weight change within race/ethnic groups, and differences across them. Data are from linked New York City birth certificates 1994-2004. The sample comprised nulliparous women ages ?18 with two consecutive singleton births (N = 115,651). The dependent variable was interconception weight change. Adjusted analyses were from ordinary least squares regression model fully interacted by race/ethnic group, controlling for covariates. Within race/ethnic groups, adjusted interconception weight change was calculated for each GWG level compared with GWG 20-24 pounds; across groups, weight change was calculated for each group compared with white non-Hispanics. GWG ?40 pounds was 18 % for Asian Pacific Islanders, and 27-29 % for other race/ethnic groups. Interconception weight change >15 pounds was highest for black non-Hispanics (34 %) and lowest for Asian Pacific Islanders (17 %). In the multivariable analysis, mean interconception weight change increased with increasing GWG in all race/ethnic groups, an average of 1.5-1.6 pounds for each 5-pound GWG interval. Compared with white non-Hispanics, adjusted mean interconception weight change was higher at every GWG level for black non-Hispanics (3.5-5.1 pounds), and at every level except <15 pounds for Hispanics (1.6-3.0 pounds). GWG ?40 pounds was prevalent in all groups. GWG contributes to long-term interconception weight change, and non-Hispanic blacks and Hispanics are at risk of greater weight change. Interventions at many levels, during and after pregnancy, are needed to support women to achieve healthy GWG and postpartum weight loss. PMID:25424453
Efficiency of transportation on weighted extended Koch networks
NASA Astrophysics Data System (ADS)
Wu, Zikai; Hou, Baoyu; Zhang, Hongjuan
2013-10-01
In this paper, we propose a family of weighted extended Koch networks based on a class of extended Koch networks. They originate from a r-complete graph, and each node in each r-complete graph of current generation produces mr-complete graphs whose weighted edges are scaled by factor h in subsequent evolutionary step. We study the structural properties of these networks and random walks on them. In more detail, we calculate exactly the average weighted shortest path length (AWSP), average receiving time (ART) and average sending time (AST). Besides, the technique of resistor network is employed to uncover the relationship between ART and AST on networks with unit weight. In the infinite network order limit, the average weighted shortest path lengths stay bounded with growing network order (0 < h < 1). The closed form expression of ART shows that it exhibits a sub-linear dependence (0 < h < 1) or linear dependence ( h = 1) on network order. On the contrary, the AST behaves super-linearly with the network order. Collectively, all the obtained results show that the efficiency of message transportation on weighted extended Koch networks has close relation to the network parameters h, m and r. All these findings could shed light on the structure and random walks of general weighted networks.
Ensemble Weight Enumerators for Protograph LDPC Codes
NASA Technical Reports Server (NTRS)
Divsalar, Dariush
2006-01-01
Recently LDPC codes with projected graph, or protograph structures have been proposed. In this paper, finite length ensemble weight enumerators for LDPC codes with protograph structures are obtained. Asymptotic results are derived as the block size goes to infinity. In particular we are interested in obtaining ensemble average weight enumerators for protograph LDPC codes which have minimum distance that grows linearly with block size. As with irregular ensembles, linear minimum distance property is sensitive to the proportion of degree-2 variable nodes. In this paper the derived results on ensemble weight enumerators show that linear minimum distance condition on degree distribution of unstructured irregular LDPC codes is a sufficient but not a necessary condition for protograph LDPC codes.
High Average Power, High Energy Short Pulse Fiber Laser System
Messerly, M J
2007-11-13
Recently continuous wave fiber laser systems with output powers in excess of 500W with good beam quality have been demonstrated [1]. High energy, ultrafast, chirped pulsed fiber laser systems have achieved record output energies of 1mJ [2]. However, these high-energy systems have not been scaled beyond a few watts of average output power. Fiber laser systems are attractive for many applications because they offer the promise of high efficiency, compact, robust systems that are turn key. Applications such as cutting, drilling and materials processing, front end systems for high energy pulsed lasers (such as petawatts) and laser based sources of high spatial coherence, high flux x-rays all require high energy short pulses and two of the three of these applications also require high average power. The challenge in creating a high energy chirped pulse fiber laser system is to find a way to scale the output energy while avoiding nonlinear effects and maintaining good beam quality in the amplifier fiber. To this end, our 3-year LDRD program sought to demonstrate a high energy, high average power fiber laser system. This work included exploring designs of large mode area optical fiber amplifiers for high energy systems as well as understanding the issues associated chirped pulse amplification in optical fiber amplifier systems.
Andersson, Neil; Mitchell, Steven
2006-01-01
Evaluation of mine risk education in Afghanistan used population weighted raster maps as an evaluation tool to assess mine education performance, coverage and costs. A stratified last-stage random cluster sample produced representative data on mine risk and exposure to education. Clusters were weighted by the population they represented, rather than the land area. A "friction surface" hooked the population weight into interpolation of cluster-specific indicators. The resulting population weighted raster contours offer a model of the population effects of landmine risks and risk education. Five indicator levels ordered the evidence from simple description of the population-weighted indicators (level 0), through risk analysis (levels 1-3) to modelling programme investment and local variations (level 4). Using graphic overlay techniques, it was possible to metamorphose the map, portraying the prediction of what might happen over time, based on the causality models developed in the epidemiological analysis. Based on a lattice of local site-specific predictions, each cluster being a small universe, the "average" prediction was immediately interpretable without losing the spatial complexity. PMID:16390549
Andersson, Neil; Mitchell, Steven
2006-01-01
Evaluation of mine risk education in Afghanistan used population weighted raster maps as an evaluation tool to assess mine education performance, coverage and costs. A stratified last-stage random cluster sample produced representative data on mine risk and exposure to education. Clusters were weighted by the population they represented, rather than the land area. A "friction surface" hooked the population weight into interpolation of cluster-specific indicators. The resulting population weighted raster contours offer a model of the population effects of landmine risks and risk education. Five indicator levels ordered the evidence from simple description of the population-weighted indicators (level 0), through risk analysis (levels 13) to modelling programme investment and local variations (level 4). Using graphic overlay techniques, it was possible to metamorphose the map, portraying the prediction of what might happen over time, based on the causality models developed in the epidemiological analysis. Based on a lattice of local site-specific predictions, each cluster being a small universe, the "average" prediction was immediately interpretable without losing the spatial complexity. PMID:16390549
Association between Dietary Carbohydrates and Body Weight
Ma, Yunsheng; Olendzki, Barbara; Chiriboga, David; Hebert, James R.; Li, Youfu; Li, Wenjun; Campbell, MaryJane; Gendreau, Katherine; Ockene, Ira S.
2005-01-01
The role of dietary carbohydrates in weight loss has received considerable attention in light of the current obesity epidemic. The authors investigated the association of body mass index (weight (kg)/height (m)2) with dietary intake of carbohydrates and with measures of the induced glycemic response, using data from an observational study of 572 healthy adults in central Massachusetts. Anthropometric measurements, 7-day dietary recalls, and physical activity recalls were collected quarterly from each subject throughout a 1-year study period. Data were collected between 1994 and 1998. Longitudinal analyses were conducted, and results were adjusted for other factors related to body habitus. Average body mass index was 27.4 kg/m2 (standard deviation, 5.5), while the average percentage of calories from carbohydrates was 44.9 (standard deviation, 9.6). Mean daily dietary glycemic index was 81.7 (standard deviation, 5.5), and glycemic load was 197.8 (standard deviation, 105.2). Body mass index was found to be positively associated with glycemic index, a measure of the glycemic response associated with ingesting different types of carbohydrates, but not with daily carbohydrate intake, percentage of calories from carbohydrates, or glycemic load. Results suggest that the type of carbohydrate may be related to body weight. However, further research is required to elucidate this association and its implications for weight management. PMID:15692080
Representation of object weight in human ventral visual cortex.
Gallivan, Jason P; Cant, Jonathan S; Goodale, Melvyn A; Flanagan, J Randall
2014-08-18
Skilled manipulation requires the ability to predict the weights of viewed objects based on learned associations linking object weight to object visual appearance. However, the neural mechanisms involved in extracting weight information from viewed object properties are unknown. Given that ventral visual pathway areas represent a wide variety of object features, one intriguing but as yet untested possibility is that these areas also represent object weight, a nonvisual motor-relevant object property. Here, using event-related fMRI and pattern classification techniques, we tested the novel hypothesis that object-sensitive regions in occipitotemporal cortex (OTC), in addition to traditional motor-related brain areas, represent object weight when preparing to lift that object. In two studies, the same participants prepared and then executed lifting actions with objects of varying weight. In the first study, we show that when lifting visually identical objects, where predicted weight is based solely on sensorimotor memory, weight is represented in object-sensitive OTC. In the second study, we show that when object weight is associated with a particular surface texture, that texture-sensitive OTC areas also come to represent object weight. Notably, these texture-sensitive areas failed to carry information about weight in the first study, when object surface properties did not specify weight. Our results indicate that the integration of visual and motor-relevant object information occurs at the level of single OTC areas and provide evidence that the ventral visual pathway is actively and flexibly engaged in processing object weight, an object property critical for action planning and control. PMID:25065755
A hierarchical Bayesian model averaging framework for groundwater prediction under uncertainty.
Chitsazan, Nima; Tsai, Frank T-C
2015-01-01
Groundwater prediction models are subjected to various sources of uncertainty. This study introduces a hierarchical Bayesian model averaging (HBMA) method to segregate and prioritize sources of uncertainty in a hierarchical structure and conduct BMA for concentration prediction. A BMA tree of models is developed to understand the impact of individual sources of uncertainty and uncertainty propagation to model predictions. HBMA evaluates the relative importance of different modeling propositions at each level in the BMA tree of model weights. The HBMA method is applied to chloride concentration prediction for the "1,500-foot" sand of the Baton Rouge area, Louisiana from 2005 to 2029. The groundwater head data from 1990 to 2004 is used for model calibration. Four sources of uncertainty are considered and resulted in 180 flow and transport models for concentration prediction. The results show that prediction variances of concentration from uncertain model elements are much higher than the prediction variance from uncertain model parameters. The HBMA method is able to quantify the contributions of individual sources of uncertainty to the total uncertainty. PMID:24890644
A theoretical account of cue averaging in the rodent head direction system
Page, Hector J. I.; Walters, Daniel M.; Knight, Rebecca; Piette, Caitlin E.; Jeffery, Kathryn J.; Stringer, Simon M.
2014-01-01
Head direction (HD) cell responses are thought to be derived from a combination of internal (or idiothetic) and external (or allothetic) sources of information. Recent work from the Jeffery laboratory shows that the relative influence of visual versus vestibular inputs upon the HD cell response depends on the disparity between these sources. In this paper, we present simulation results from a model designed to explain these observations. The model accurately replicates the Knight et al. data. We suggest that cue conflict resolution is critically dependent on plastic remapping of visual information onto the HD cell layer. This remap results in a shift in preferred directions of a subset of HD cells, which is then inherited by the rest of the cells during path integration. Thus, we demonstrate how, over a period of several minutes, a visual landmark may gain cue control. Furthermore, simulation results show that weaker visual landmarks fail to gain cue control as readily. We therefore suggest a second longer term plasticity in visual projections onto HD cell areas, through which landmarks with an inconsistent relationship to idiothetic information are made less salient, significantly hindering their ability to gain cue control. Our results provide a mechanism for reliability-weighted cue averaging that may pertain to other neural systems in addition to the HD system. PMID:24366143
A theoretical account of cue averaging in the rodent head direction system.
Page, Hector J I; Walters, Daniel M; Knight, Rebecca; Piette, Caitlin E; Jeffery, Kathryn J; Stringer, Simon M
2014-02-01
Head direction (HD) cell responses are thought to be derived from a combination of internal (or idiothetic) and external (or allothetic) sources of information. Recent work from the Jeffery laboratory shows that the relative influence of visual versus vestibular inputs upon the HD cell response depends on the disparity between these sources. In this paper, we present simulation results from a model designed to explain these observations. The model accurately replicates the Knight et al. data. We suggest that cue conflict resolution is critically dependent on plastic remapping of visual information onto the HD cell layer. This remap results in a shift in preferred directions of a subset of HD cells, which is then inherited by the rest of the cells during path integration. Thus, we demonstrate how, over a period of several minutes, a visual landmark may gain cue control. Furthermore, simulation results show that weaker visual landmarks fail to gain cue control as readily. We therefore suggest a second longer term plasticity in visual projections onto HD cell areas, through which landmarks with an inconsistent relationship to idiothetic information are made less salient, significantly hindering their ability to gain cue control. Our results provide a mechanism for reliability-weighted cue averaging that may pertain to other neural systems in addition to the HD system. PMID:24366143
Microstructural effects on the average properties in porous battery electrodes
NASA Astrophysics Data System (ADS)
García-García, Ramiro; García, R. Edwin
2016-03-01
A theoretical framework is formulated to analytically quantify the effects of the microstructure on the average properties of porous electrodes, including reactive area density and the through-thickness tortuosity as observed in experimentally-determined tomographic sections. The proposed formulation includes the microstructural non-idealities but also captures the well-known perfectly spherical limit. Results demonstrate that in the absence of any particle alignment, the through-thickness Bruggeman exponent α, reaches an asymptotic value of α ∼ 2 / 3 as the shape of the particles become increasingly prolate (needle- or fiber-like). In contrast, the Bruggeman exponent diverges as the shape of the particles become increasingly oblate, regardless of the degree of particle alignment. For aligned particles, tortuosity can be dramatically suppressed, e.g., α → 1 / 10 for ra → 1 / 10 and MRD ∼ 40 . Particle size polydispersity impacts the porosity-tortuosity relation when the average particle size is comparable to the thickness of the electrode layers. Electrode reactivity density can be arbitrarily increased as the particles become increasingly oblate, but asymptotically reach a minimum value as the particles become increasingly prolate. In the limit of a porous electrode comprised of fiber-like particles, the area density decreases by 24% , with respect to a distribution of perfectly spherical particles.
Tracheal cross-sectional area in children: CT determination
Effmann, E.L.; Fram, E.K.; Vock, P.; Kirks, D.R.
1983-10-01
A computer method that calculates tracheal cross-sectional area by compensating for partial volume averaging was developed and validated in a study with phantoms. The program was then used to determine the tracheal cross-sectional area of 30 normal children who ranged in age from four months to 18 years. CT-derived cross sections were correlated with age, height, weight, and body-surface area, and they were compared with findings of published clinical and post-mortem studies. CT cross-sectional areas ranged from 20 to 275 mm/sup 2/, varied by as much as 22% at the three different tracheal levels studied, and appeared to correlate most closely with body height. CT-derived tracheal cross-sectional areas are quite similar to those in published reports of postmortem and clinical studies. Measurements of tracheal cross section by CT may prove useful in quantitating tracheal compromise by intrinsic or extrinsic causes.
Paper area density measurement from forward transmitted scattered light
Koo, Jackson C.
2001-01-01
A method whereby the average paper fiber area density (weight per unit area) can be directly calculated from the intensity of transmitted, scattered light at two different wavelengths, one being a non-absorpted wavelength. Also, the method makes it possible to derive the water percentage per fiber area density from a two-wavelength measurement. In the optical measuring technique optical transmitted intensity, for example, at 2.1 microns cellulose absorption line is measured and compared with another scattered, optical transmitted intensity reference in the nearby spectrum region, such as 1.68 microns, where there is no absorption. From the ratio of these two intensities, one can calculate the scattering absorption coefficient at 2.1 microns. This absorption coefficient at this wavelength is, then, experimentally correlated to the paper fiber area density. The water percentage per fiber area density can be derived from this two-wavelength measurement approach.
Optimizing Average Precision Using Weakly Supervised Data.
Behl, Aseem; Mohapatra, Pritish; Jawahar, C V; Kumar, M Pawan
2015-12-01
Many tasks in computer vision, such as action classification and object detection, require us to rank a set of samples according to their relevance to a particular visual category. The performance of such tasks is often measured in terms of the average precision (ap). Yet it is common practice to employ the support vector machine ( svm) classifier, which optimizes a surrogate 0-1 loss. The popularity of svmcan be attributed to its empirical performance. Specifically, in fully supervised settings, svm tends to provide similar accuracy to ap-svm, which directly optimizes an ap-based loss. However, we hypothesize that in the significantly more challenging and practically useful setting of weakly supervised learning, it becomes crucial to optimize the right accuracy measure. In order to test this hypothesis, we propose a novel latent ap-svm that minimizes a carefully designed upper bound on the ap-based loss function over weakly supervised samples. Using publicly available datasets, we demonstrate the advantage of our approach over standard loss-based learning frameworks on three challenging problems: action classification, character recognition and object detection. PMID:26539857
Average oxidation state of carbon in proteins
Dick, Jeffrey M.
2014-01-01
The formal oxidation state of carbon atoms in organic molecules depends on the covalent structure. In proteins, the average oxidation state of carbon (ZC) can be calculated as an elemental ratio from the chemical formula. To investigate oxidation–reduction (redox) patterns, groups of proteins from different subcellular locations and phylogenetic groups were selected for comparison. Extracellular proteins of yeast have a relatively high oxidation state of carbon, corresponding with oxidizing conditions outside of the cell. However, an inverse relationship between ZC and redox potential occurs between the endoplasmic reticulum and cytoplasm. This trend provides support for the hypothesis that protein transport and turnover are ultimately coupled to the maintenance of different glutathione redox potentials in subcellular compartments. There are broad changes in ZC in whole-genome protein compositions in microbes from different environments, and in Rubisco homologues, lower ZC tends to occur in organisms with higher optimal growth temperature. Energetic costs calculated from thermodynamic models are consistent with the notion that thermophilic organisms exhibit molecular adaptation to not only high temperature but also the reducing nature of many hydrothermal fluids. Further characterization of the material requirements of protein metabolism in terms of the chemical conditions of cells and environments may help to reveal other linkages among biochemical processes with implications for changes on evolutionary time scales. PMID:25165594
Calculating Free Energies Using Average Force
NASA Technical Reports Server (NTRS)
Darve, Eric; Pohorille, Andrew; DeVincenzi, Donald L. (Technical Monitor)
2001-01-01
A new, general formula that connects the derivatives of the free energy along the selected, generalized coordinates of the system with the instantaneous force acting on these coordinates is derived. The instantaneous force is defined as the force acting on the coordinate of interest so that when it is subtracted from the equations of motion the acceleration along this coordinate is zero. The formula applies to simulations in which the selected coordinates are either unconstrained or constrained to fixed values. It is shown that in the latter case the formula reduces to the expression previously derived by den Otter and Briels. If simulations are carried out without constraining the coordinates of interest, the formula leads to a new method for calculating the free energy changes along these coordinates. This method is tested in two examples - rotation around the C-C bond of 1,2-dichloroethane immersed in water and transfer of fluoromethane across the water-hexane interface. The calculated free energies are compared with those obtained by two commonly used methods. One of them relies on determining the probability density function of finding the system at different values of the selected coordinate and the other requires calculating the average force at discrete locations along this coordinate in a series of constrained simulations. The free energies calculated by these three methods are in excellent agreement. The relative advantages of each method are discussed.
Average structure of incommensurately modulated monoclinic lazurite
Bolotina, N. B.; Rastsvetaeva, R. K. Sapozhnikov, A. N.
2006-07-15
The average structure of the monoclinic modification of lazurite Ca{sub 1.26}Na{sub 6.63}K{sub 0.04}[Al{sub 6}Si{sub 6}O{sub 24}](SO{sub 4}){sub 1.53}S{sub 0.99}Cl{sub 0.05} (discovered in the Lake Baikal region) incommensurately modulated along the c axis is solved by the single-crystal X-ray diffraction method. The unit-cell parameters are a = 9.069(1) A, b = 12.868(1) A, c = 12.872(1) A, {gamma} = 90.19(1) deg., sp. gr. Pa, R = 6.9%, 2057 reflections. The AlO{sub 4} and SiO{sub 4} tetrahedra form a partially ordered framework. The positions in the cavities of the framework are split and randomly occupied by Na and Ca atoms and the SO{sub 4}, S{sub 2}, S{sub 3}, and SO{sub 2} anionic groups. The structure of the mineral is compared with the superstructure of triclinic lazurite. Conclusions are drawn about the causes of the incommensurate modulation in monoclinic and other lazurites.
Models for predicting objective function weights in prostate cancer IMRT
Boutilier, Justin J. Lee, Taewoo; Craig, Tim; Sharpe, Michael B.; Chan, Timothy C. Y.
2015-04-15
Purpose: To develop and evaluate the clinical applicability of advanced machine learning models that simultaneously predict multiple optimization objective function weights from patient geometry for intensity-modulated radiation therapy of prostate cancer. Methods: A previously developed inverse optimization method was applied retrospectively to determine optimal objective function weights for 315 treated patients. The authors used an overlap volume ratio (OV) of bladder and rectum for different PTV expansions and overlap volume histogram slopes (OVSR and OVSB for the rectum and bladder, respectively) as explanatory variables that quantify patient geometry. Using the optimal weights as ground truth, the authors trained and applied three prediction models: logistic regression (LR), multinomial logistic regression (MLR), and weighted K-nearest neighbor (KNN). The population average of the optimal objective function weights was also calculated. Results: The OV at 0.4 cm and OVSR at 0.1 cm features were found to be the most predictive of the weights. The authors observed comparable performance (i.e., no statistically significant difference) between LR, MLR, and KNN methodologies, with LR appearing to perform the best. All three machine learning models outperformed the population average by a statistically significant amount over a range of clinical metrics including bladder/rectum V53Gy, bladder/rectum V70Gy, and dose to the bladder, rectum, CTV, and PTV. When comparing the weights directly, the LR model predicted bladder and rectum weights that had, on average, a 73% and 74% relative improvement over the population average weights, respectively. The treatment plans resulting from the LR weights had, on average, a rectum V70Gy that was 35% closer to the clinical plan and a bladder V70Gy that was 29% closer, compared to the population average weights. Similar results were observed for all other clinical metrics. Conclusions: The authors demonstrated that the KNN and MLR weight prediction methodologies perform comparably to the LR model and can produce clinical quality treatment plans by simultaneously predicting multiple weights that capture trade-offs associated with sparing multiple OARs.
Particle sizing by weighted measurements of scattered light
NASA Technical Reports Server (NTRS)
Buchele, Donald R.
1988-01-01
A description is given of a measurement method, applicable to a poly-dispersion of particles, in which the intensity of scattered light at any angle is weighted by a factor proportional to that angle. Determination is then made of four angles at which the weighted intensity is four fractions of the maximum intensity. These yield four characteristic diameters, i.e., the diameters of the volume/area mean (D sub 32 the Sauter mean) and the volume/diameter mean (D sub 31); the diameters at cumulative volume fractions of 0.5 (D sub v0.5 the volume median) and 0.75 (D sub v0.75). They also yield the volume dispersion of diameters. Mie scattering computations show that an average diameter less than three micrometers cannot be accurately measured. The results are relatively insensitive to extraneous background light and to the nature of the diameter distribution. Also described is an experimental method of verifying the conclusions by using two microscopic slides coated with polystyrene microspheres to simulate the particles and the background.
Basso, Olga
2008-03-01
Birth weight is associated not just with infant morbidity and mortality, but with outcomes occurring much later in life, including adult mortality, as reported by a paper by Baker and colleagues in this issue of Epidemiology. While these associations are tantalizing per se, the truly interesting question concerns the mechanisms that underlie these links. The prevailing hypothesis suggests a "fetal origin" of diseases resulting from alterations in fetal nutrition that permanently program organ function. The most commonly proposed alternative is that factors, mainly genetic, that affect both fetal growth and disease risk are responsible for the observed associations. Although both mechanisms are intellectually attractive-and may well coexist-we should be cautious to not focus excessively on fetal growth. Doing this may lead us in the wrong direction, as has likely happened in the case of birth weight in relation to infant survival. PMID:18277158
Morasiewicz, Piotr; Dragan, Szymon
2013-01-01
Distortion of the axis and shortening of the limbs result in multiple musculoskeletal pathologies. Rotation disorders should also be included among the disorders of the axis of the lower limb. In the case of rotational distortion, only derotation osteotomy can effectively correct torsion-associated deformations. Rotational distortion correction is accompanied by translational displacement and torsion, which results in more complex biomechanics. Using the pedobarographic platform, it is possible to evaluate static and dynamic posture and gait, percentage of body weight distribution on the lower limbs, and balance. Physiological gait and distribution of weight on the lower extremities are symmetrical. Balance is one of the determinants of proper biomechanics of the musculoskeletal system. An important aspect of treatment evaluation is pedobarographic assessment of balance and body weight distribution on the lower extremities ratio. The aim of this work was to evaluate the pedobarographic assessment of body weight distribution on the lower limbs and balance in patients with derotation corticotomies using the Ilizarov method. The study examined a group of 56 patients, who underwent derotation corticotomy using the Illizarov method between 1996 and 2012 at the Clinic of Orthopaedics and Traumatology of the Musculoskeletal System in Wrocław. The control group consisted of 54 patients, who were treated with correctional derotation-free corticotomy using the Ilizarov. Distribution of body weight on the lower limbs and balance were assessed with the pedobarographic platform. Following derotation corticotomy, the amount of body weight placed on the operated limb by subjects from the study group averaged 47.81%, 52.19% in the case of the healthy limb. These differences were not statistically significant. The difference between the average percentage of body weight placed on the diseased and healthy limb in the study group and the controls were not found to be statistically significant. There were no statistical differences in the average length of the gravity line or in the average surface area of the center of gravity position between the study and control groups. Balanced distribution of body weight on the lower limbs was achieved following derotation corticotomies using the Ilizarov method. Derotation corticotomies performed with the Ilizarov method allow for achieving normalization of body weight distribution on the lower limbs and balance, with values similar to those resulting from Ilizarov method derotation-free osteotomy. PMID:23952018
Kirkland, L.; Anderson, R.
1993-01-01
Only 5% of patients dieting to achieve permanent weight loss will be successful and reap the associated health benefits. Ninety-five percent will be unsuccessful. The health implications of failed dieting attempts are numerous and include negative effects on both physical and psychological well-being. Better alternatives to dieting help patients take small, positive, enjoyable steps toward healthy eating, active living, and a positive self-image. PMID:8435552
High School Weight Training: A Comprehensive Program.
ERIC Educational Resources Information Center
Viscounte, Roger; Long, Ken
1989-01-01
Describes a weight training program, suitable for the general student population and the student-athlete, which is designed to produce improvement in specific, measurable areas including bench press (upper body), leg press (lower body), vertical jump (explosiveness); and 40-yard dash (speed). Two detailed charts are included, with notes on their
High School Weight Training: A Comprehensive Program.
ERIC Educational Resources Information Center
Viscounte, Roger; Long, Ken
1989-01-01
Describes a weight training program, suitable for the general student population and the student-athlete, which is designed to produce improvement in specific, measurable areas including bench press (upper body), leg press (lower body), vertical jump (explosiveness); and 40-yard dash (speed). Two detailed charts are included, with notes on their…
Raboisson, Didier; Dervill, Marie; Herman, Nicolas; Cahuzac, Eric; Sans, Pierre; Allaire, Gilles
2012-08-01
Mastitis is a multifactorial disease and the most costly dairy production issue. In spite of extensive literature on udder-health risk factors, effects of metabolic diseases, farmers' competencies and livestock farming system on somatic cells count (SCC) are sparsely described. Herd-level or territorial-level factors affecting monthly composite milk weighted mean cow SCC (CMSCC) were analysed with a linear mixed effect model. The average CMSCC was 266,000 cells/ml. Half of the herds had CMSCC >300,000 cells/ml for 2-6 months a year, and 15% of herds for more than 7 months a year. CMSCC was positively associated with the number of cows, having a beef or fattening herd in addition to the dairy herd, the monthly average days in milk, the yearly age at first calving, the yearly proportion of purchased cows and the yearly culling rate. Moreover, a positive association is reported between CMSCC and the monthly proportion of cows probably with subacute ruminal acidosis (fat percentage minus protein percentage ?030%, for Holstein) and negative energy balance (protein to fat ratio ?066, for Holstein), the yearly average calving interval, having at least one dead cow and the mean monthly temperature. The association was negative for a predominant breed other than Holstein, the monthly milk production, the yearly dry-off period length, the monthly first calving cow proportion, having an autumn calving peak, being a Good Breeding Practices member, the monthly number of days with rain, the altitude and the territorial cattle density. CMSCC varied widely among the 11 dairy production areas. In conclusion, this study showed the average CMSCC for the French dairy cows, compared with international results. Moreover, it quantified the contribution of several factors to CMSCC, in particular metabolic diseases and the farm environment. PMID:22687283
Correlation Weights in Multiple Regression
ERIC Educational Resources Information Center
Waller, Niels G.; Jones, Jeff A.
2010-01-01
A general theory on the use of correlation weights in linear prediction has yet to be proposed. In this paper we take initial steps in developing such a theory by describing the conditions under which correlation weights perform well in population regression models. Using OLS weights as a comparison, we define cases in which the two weighting
Social embeddedness in an online weight management programme is linked to greater weight loss.
Poncela-Casasnovas, Julia; Spring, Bonnie; McClary, Daniel; Moller, Arlen C; Mukogo, Rufaro; Pellegrini, Christine A; Coons, Michael J; Davidson, Miriam; Mukherjee, Satyam; Nunes Amaral, Luis A
2015-03-01
The obesity epidemic is heightening chronic disease risk globally. Online weight management (OWM) communities could potentially promote weight loss among large numbers of people at low cost. Because little is known about the impact of these online communities, we examined the relationship between individual and social network variables, and weight loss in a large, international OWM programme. We studied the online activity and weight change of 22,419 members of an OWM system during a six-month period, focusing especially on the 2033 members with at least one friend within the community. Using Heckman's sample-selection procedure to account for potential selection bias and data censoring, we found that initial body mass index, adherence to self-monitoring and social networking were significantly correlated with weight loss. Remarkably, greater embeddedness in the network was the variable with the highest statistical significance in our model for weight loss. Average per cent weight loss at six months increased in a graded manner from 4.1% for non-networked members, to 5.2% for those with a few (two to nine) friends, to 6.8% for those connected to the giant component of the network, to 8.3% for those with high social embeddedness. Social networking within an OWM community, and particularly when highly embedded, may offer a potent, scalable way to curb the obesity epidemic and other disorders that could benefit from behavioural changes. PMID:25631561
Longitudinal Study of Body Weight Changes in Children: Who is Gaining and Who is Losing Weight
Williamson, D. A.; Han, H.; Johnson, W.D.; Stewart, T.M.; Harsha, D.
2010-01-01
Cross-sectional studies have reported significant temporal increases in prevalence of childhood obesity in both genders and various racial groups, but recently the rise has subsided. Childhood obesity prevention trials suggest that, on average, overweight/obese children lose body weight and non-overweight children gain weight. This investigation tested the hypothesis that overweight children lose body weight/fat and non-overweight children gain body weight/fat using a longitudinal research design that did not include an obesity prevention program. The participants were 451 children in 4th to 6th grades at baseline. Height, weight, and body fat were measured at Month 0 and Month 28. Each child’s body mass index (BMI) percentile score was calculated specific for their age, gender and height. Higher BMI percentile scores and percent body fat at baseline were associated with larger decreases in BMI and percent body fat after 28 months. The BMI percentile mean for African-American girls increased whereas BMI percentile means for white boys and girls and African-American boys were stable over the 28 month study period. Estimates of obesity and overweight prevalence were stable because incidence and remission were similar. These findings support the hypothesis that overweight children tend to lose body weight and non-overweight children tend to gain body weight. PMID:20885393
Physiological weight loss in the breastfed neonate: a systematic review
Noel-Weiss, Joy; Courant, Genevieve; Woodend, A Kirsten
2008-01-01
Background Healthy, full-term, exclusively breastfed infants are expected to lose weight in the first days following birth. There are conflicting opinions about what constitutes a normal neonatal weight loss, and about when interventions such as supplemental feedings should be considered. Objective To establish the reference weight loss for the first 2 weeks following birth by conducting a systematic review of studies reporting birth weights of exclusively breastfed neonates. Methods We searched 5 electronic databases from June 2006 to June 2007: the Cochrane Database of Systematic Reviews; MEDLINE (from 1950); CINAHL (from 1982); EMBASE (from 1980); and Ovid HealthSTAR (from 1999). We included primary research studies with weight loss data for healthy, full-term, exclusively breastfed neonates in the first 2 weeks following birth. Results Eleven studies met the inclusion criteria. Definitions, types of measurements, and reporting styles varied among studies. In most studies, daily weights were not measured and measurements did not continue for 2 weeks. Mean weight loss ranged from 5.7% to 6.6%, with standard deviations around 2%. Median percentage weight loss ranged from 3.2 to 8.3, with the majority around 6%. The majority of infants in these 11 studies regained their birth weight within the first 2 weeks postpartum. The second and third days following birth appear to be the days of maximum weight loss. Discussion Methods used to report weight loss were inconsistent, using either an average of single lowest weights or a combination of weight losses. The 7% maximum allowable weight loss recommended in 4 clinical practice guidelines appears to be based on mean weight loss and does not account for standard deviation. Further research is needed to understand the causes of neonatal weight loss and its implications for morbidity and mortality. PMID:21602959
Tonsillitis, tonsillectomy and weight disturbance.
Conlon, B J; Donnelly, M J; McShane, D P
1997-10-18
To determine the relationship between tonsillitis, tonsillectomy and abnormalities in body weight, we have analyzed pre- and post-operative weights in a population of 55 children who underwent adenotonsillectomy in our department. Pre-operative mean weight was 9.8% heavier than the standard mean normal weight for age and post-operative mean weight was 22% greater than standard mean weight for age. The mean weight gain during the follow-up period was 12% greater than that which would be normally expected (p < 0.001). This study suggests that children undergoing tonsillectomy are slightly heavier than their peers and that following the procedure this discrepancy increases. PMID:9477349
Relationship between smoking, weight and attitudes to weight in adolescent schoolgirls.
Halek, C.; Kerry, S.; Humphrey, H.; Crisp, A. H.; Hughes, J. M.
1993-01-01
A total of 1,932 schoolgirls aged 11-18 from seven schools in the South London area were surveyed using questionnaires which addressed eating patterns, body weight history, attitudes to body weight and shape, menstrual history and current smoking behaviour. They were also weighed and their height was measured. Twelve per cent of the girls were regular smokers and 10% smoked seven or more cigarettes over a 4 day period. Amongst girls aged 14 and over, 15% smoked regularly and a further 9% occasionally. A significant relationship was found between smoking and weight. Smokers were more likely to be moderately overweight in relation to their peers and to have been worried about their weight at some stage. There were differences between girls in state schools and those in independent schools with regard to smoking behaviour and weight. The findings have implications for anti-smoking strategies and health education generally. PMID:8506187
Low birth weight and residential proximity to PCB-contaminated waste sites.
Baibergenova, Akerke; Kudyakov, Rustam; Zdeb, Michael; Carpenter, David O
2003-01-01
Previous investigations have shown that women exposed to polychlorinated biphenyls (PCBs) are at increased risk of giving birth to an infant with low birth weight (< 2,500 g), and that this relationship is stronger for male than for female infants. We have tested the hypothesis that residents in a zip code that contains a PCB hazardous waste site or abuts a body of water contaminated with PCBs are at increased risk of giving birth to a low-birth-weight baby. We used the birth registry of the New York State Vital Statistics to identify all births between 1994 and 2000 in New York State except for New York City. This registry provides information on the infant, mother, and father together with the zip code of the mother's residence. The 865 state Superfund sites, the 86 National Priority List sites, and the six Areas of Concern in New York were characterized regarding whether or not they contain PCBs as a major contaminant. We identified 187 zip codes containing or abutting PCB-contaminated sites, and these zip codes were the residences of 24.5% of the 945,077 births. The birth weight in the PCB zip codes was on average 21.6 g less than in other zip codes (p < 0.001). Because there are many other risk factors for low birth weight, we have adjusted for these using a logistic regression model for these confounders. After adjusting for sex of the infant, mother's age, race, weight, height, education, income, marital status, and smoking, there was still a statistically significant 6% increased risk of giving birth to a male infant of low birth weight. These observations support the hypothesis that living in a zip code near a PCB-contaminated site poses a risk of exposure and giving birth to an infant of low birth weight. PMID:12896858
Upper Limit of Weights in TAI Computation
NASA Technical Reports Server (NTRS)
Thomas, Claudine; Azoubib, Jacques
1996-01-01
The international reference time scale International Atomic Time (TAI) computed by the Bureau International des Poids et Mesures (BIPM) relies on a weighted average of data from a large number of atomic clocks. In it, the weight attributed to a given clock depends on its long-term stability. In this paper the TAI algorithm is used as the basis for a discussion of how to implement an upper limit of weight for clocks contributing to the ensemble time. This problem is approached through the comparison of two different techniques. In one case, a maximum relative weight is fixed: no individual clock can contribute more than a given fraction to the resulting time scale. The weight of each clock is then adjusted according to the qualities of the whole set of contributing elements. In the other case, a parameter characteristic of frequency stability is chosen: no individual clock can appear more stable than the stated limit. This is equivalent to choosing an absolute limit of weight and attributing this to to the most stable clocks independently of the other elements of the ensemble. The first technique is more robust than the second and automatically optimizes the stability of the resulting time scale, but leads to a more complicated computatio. The second technique has been used in the TAI algorithm since the very beginning. Careful analysis of tests on real clock data shows that improvement of the stability of the time scale requires revision from time to time of the fixed value chosen for the upper limit of absolute weight. In particular, we present results which confirm the decision of the CCDS Working Group on TAI to increase the absolute upper limit by a factor of 2.5. We also show that the use of an upper relative contribution further helps to improve the stability and may be a useful step towards better use of the massive ensemble of HP 507IA clocks now contributing to TAI.
Optimal estimation of the diffusion coefficient from non-averaged and averaged noisy magnitude data
NASA Astrophysics Data System (ADS)
Kristoffersen, Anders
2007-08-01
The magnitude operation changes the signal distribution in MRI images from Gaussian to Rician. This introduces a bias that must be taken into account when estimating the apparent diffusion coefficient. Several estimators are known in the literature. In the present paper, two novel schemes are proposed. Both are based on simple least squares fitting of the measured signal, either to the median (MD) or to the maximum probability (MP) value of the Probability Density Function (PDF). Fitting to the mean (MN) or a high signal-to-noise ratio approximation to the mean (HS) is also possible. Special attention is paid to the case of averaged magnitude images. The PDF, which cannot be expressed in closed form, is analyzed numerically. A scheme for performing maximum likelihood (ML) estimation from averaged magnitude images is proposed. The performance of several estimators is evaluated by Monte Carlo (MC) simulations. We focus on typical clinical situations, where the number of acquisitions is limited. For non-averaged data the optimal choice is found to be MP or HS, whereas uncorrected schemes and the power image (PI) method should be avoided. For averaged data MD and ML perform equally well, whereas uncorrected schemes and HS are inadequate. MD provides easier implementation and higher computational efficiency than ML. Unbiased estimation of the diffusion coefficient allows high resolution diffusion tensor imaging (DTI) and may therefore help solving the problem of crossing fibers encountered in white matter tractography.
Weight misperception amongst youth of a developing country: Pakistan -a cross-sectional study
2013-01-01
Background Weight misperception is the discordance between an individual’s actual weight status and the perception of his/her weight. It is a common problem in the youth population as enumerated by many international studies. However data from Pakistan in this area is deficient. Methods A multi-center cross-sectional survey was carried out in undergraduate university students of Karachi between the ages of 15–24. Participants were questioned regarding their perception of being thin, normal or fat and it was compared with their Body Mass Index (BMI). Measurements of height and weight were taken for this purpose and BMI was categorized using Asian cut offs. Weight misperception was identified when the self-perceived weight (average, fat, thin) did not match the calculated BMI distribution. Chi square tests and logistic regression tests were applied to show associations of misperception and types of misperception (overestimation, underestimation) with independent variables like age, gender, type of university and faculties. P-value of <0.05 was taken as statistically significant. Results 42.4% of the total participants i.e. 43.3% males and 41% females misperceived their weight. Amongst those who misperceived 38.2% had overestimated and 61.8% had underestimated their weight. Greatest misperception of was observed in the overweight category (91%), specifically amongst overweight males (95%). Females of the underweight category overestimated their weight and males of the overweight category underestimated their weight. Amongst the total participants, females overestimated 8 times more than males (OR 8.054, 95% CI 5.34-12.13). Misperception increased with the age of the participants (OR 1.114, 95% CI 1.041-1.191). Odds of misperception were greater in students of private sector universities as compared to public (OR 1.861, 95% CI: 1.29-2.67). Odds of misperception were less in students of medical sciences (OR 0.693, 95% CI 0.491-0.977), engineering (OR 0.586, 95% CI 0.364-0.941) and business administration (OR 0.439, 95% CI 0.290-0.662) as compared to general faculty universities. Conclusion There was marked discrepancy between the calculated BMI and the self-perceived weight in the youth of Karachi. Better awareness campaigns need to be implemented to reverse these trends. PMID:23915180
NASA Astrophysics Data System (ADS)
Bostan, P. A.; Heuvelink, G. B. M.; Akyurek, S. Z.
2012-10-01
Accurate mapping of the spatial distribution of annual precipitation is important for many applications in hydrology, climatology, agronomy, ecology and other environmental sciences. In this study, we compared five different statistical methods to predict spatially the average annual precipitation of Turkey using point observations of annual precipitation at meteorological stations and spatially exhaustive covariate data (i.e. elevation, aspect, surface roughness, distance to coast, land use and eco-region). The methods compared were multiple linear regression (MLR), ordinary kriging (OK), regression kriging (RK), universal kriging (UK), and geographically weighted regression (GWR). Average annual precipitation of Turkey from 1970 to 2006 was measured at 225 meteorological stations that are fairly uniformly distributed across the country, with a somewhat higher spatial density along the coastline. The observed annual precipitation varied between 255 mm and 2209 mm with an average of 628 mm. The annual precipitation was highest along the southern and northern coasts and low in the centre of the country, except for the area near the Van Lake, Keban and Ataturk Dams. To compare the performance of the interpolation techniques the total dataset was first randomly split in ten equally sized test datasets. Next, for each test data set the remaining 90% of the data comprised the training dataset. Each training dataset was then used to calibrate and apply the spatial prediction model. Predictions at the test dataset locations were compared with the observed test data. Validation was done by calculating the Root Mean Squared Error (RMSE), R-square and Standardized MSE (SMSE) values. According to these criteria, universal kriging is the most accurate with an RMSE of 178 mm, an R-square of 0.61 and an SMSE of 1.06, whilst multiple linear regression performed worst (RMSE of 222 mm, R-square of 0.39, and SMSE of 1.44). Ordinary kriging, UK using only elevation and geographically weighted regression are intermediate with RMSE values of 201 mm, 212 mm and 211 mm, and an R-square of 0.50, 0.44 and 0.45, respectively. The RK results are close to those of UK with an RMSE of 186 mm and R-square of 0.57. The spatial extrapolation performance of each method was also evaluated. This was done by predicting the annual precipitation in the eastern part of Turkey using observations from the western part. Results showed that MLR, GWR and RK performed best with little differences between these methods. The large prediction error variances confirmed that extrapolation is more difficult than interpolation. Whilst spatial extrapolation benefits most from covariate information as shown by an RMSE reduction of about 60 mm, in this study covariate information was also valuable for spatial interpolation because it reduced the RMSE with on average 30 mm.
Interpreting Sky-Averaged 21-cm Measurements
NASA Astrophysics Data System (ADS)
Mirocha, Jordan
2015-01-01
Within the first ~billion years after the Big Bang, the intergalactic medium (IGM) underwent a remarkable transformation, from a uniform sea of cold neutral hydrogen gas to a fully ionized, metal-enriched plasma. Three milestones during this epoch of reionization -- the emergence of the first stars, black holes (BHs), and full-fledged galaxies -- are expected to manifest themselves as extrema in sky-averaged ("global") measurements of the redshifted 21-cm background. However, interpreting these measurements will be complicated by the presence of strong foregrounds and non-trivialities in the radiative transfer (RT) modeling required to make robust predictions.I have developed numerical models that efficiently solve the frequency-dependent radiative transfer equation, which has led to two advances in studies of the global 21-cm signal. First, frequency-dependent solutions facilitate studies of how the global 21-cm signal may be used to constrain the detailed spectral properties of the first stars, BHs, and galaxies, rather than just the timing of their formation. And second, the speed of these calculations allows one to search vast expanses of a currently unconstrained parameter space, while simultaneously characterizing the degeneracies between parameters of interest. I find principally that (1) physical properties of the IGM, such as its temperature and ionization state, can be constrained robustly from observations of the global 21-cm signal without invoking models for the astrophysical sources themselves, (2) translating IGM properties to galaxy properties is challenging, in large part due to frequency-dependent effects. For instance, evolution in the characteristic spectrum of accreting BHs can modify the 21-cm absorption signal at levels accessible to first generation instruments, but could easily be confused with evolution in the X-ray luminosity star-formation rate relation. Finally, (3) the independent constraints most likely to aide in the interpretation of global 21-cm signal measurements are detections of Lyman Alpha Emitters at high redshifts and constraints on the midpoint of reionization, both of which are among the primary science objectives of ongoing or near-future experiments.
Uncertainty of GHz-band Whole-body Average SARs in Infants based on their Kaup Indices
NASA Astrophysics Data System (ADS)
Miwa, Hironobu; Hirata, Akimasa; Fujiwara, Osamu; Nagaoka, Tomoaki; Watanabe, Soichi
We previously showed that a strong correlation exists between the absorption cross section and the body surface area of a human for 0.3-2GHz far field exposure, and proposed a formula for estimating whole-body-average specific absorption rates (WBA-SARs) in terms of height and weight. In this study, to evaluate variability in the WBA-SARs in infants based on their physique, we derived a new formula including Kaup indices of infants, which are being used to check their growth, and thereby estimated the WBA-SARs in infants with respect to their age from 0 month to three years. As a result, we found that under the same height/weight, the smaller the Kaup indices are, the larger the WBA-SARs become, and that the variability in the WBA-SARs is around 15% at the same age. To validate these findings, using the FDTD method, we simulated the GHz-band WBA-SARs in numerical human models corresponding to infants with age of 0, 1, 3, 6 and 9 months, which were obtained by scaling down the anatomically based Japanese three-year child model developed by NICT (National Institute of Information and Communications Technology). Results show that the FDTD-simulated WBA-SARs are smaller by 20% compared to those estimated for infants having the median height and the Kaup index of 0.5 percentiles, which provide conservative WBA-SARs.
Code of Federal Regulations, 2014 CFR
2014-07-01
... average and corporate pool average sulfur level determined? 80.205 Section 80.205 Protection of... ADDITIVES Gasoline Sulfur Gasoline Sulfur Standards 80.205 How is the annual refinery or importer average and corporate pool average sulfur level determined? (a) The annual refinery or importer average...
Code of Federal Regulations, 2011 CFR
2011-07-01
... average and corporate pool average sulfur level determined? 80.205 Section 80.205 Protection of... ADDITIVES Gasoline Sulfur Gasoline Sulfur Standards 80.205 How is the annual refinery or importer average and corporate pool average sulfur level determined? (a) The annual refinery or importer average...
Code of Federal Regulations, 2013 CFR
2013-07-01
... average and corporate pool average sulfur level determined? 80.205 Section 80.205 Protection of... ADDITIVES Gasoline Sulfur Gasoline Sulfur Standards 80.205 How is the annual refinery or importer average and corporate pool average sulfur level determined? (a) The annual refinery or importer average...
Code of Federal Regulations, 2010 CFR
2010-07-01
... average and corporate pool average sulfur level determined? 80.205 Section 80.205 Protection of... ADDITIVES Gasoline Sulfur Gasoline Sulfur Standards 80.205 How is the annual refinery or importer average and corporate pool average sulfur level determined? (a) The annual refinery or importer average...
Weighting by Inverse Variance or by Sample Size in Random-Effects Meta-Analysis
ERIC Educational Resources Information Center
Marin-Martinez, Fulgencio; Sanchez-Meca, Julio
2010-01-01
Most of the statistical procedures in meta-analysis are based on the estimation of average effect sizes from a set of primary studies. The optimal weight for averaging a set of independent effect sizes is the inverse variance of each effect size, but in practice these weights have to be estimated, being affected by sampling error. When assuming a…
Weighting by Inverse Variance or by Sample Size in Random-Effects Meta-Analysis
ERIC Educational Resources Information Center
Marin-Martinez, Fulgencio; Sanchez-Meca, Julio
2010-01-01
Most of the statistical procedures in meta-analysis are based on the estimation of average effect sizes from a set of primary studies. The optimal weight for averaging a set of independent effect sizes is the inverse variance of each effect size, but in practice these weights have to be estimated, being affected by sampling error. When assuming a
Social aspects of low birth weight.
Dunn, H G
1984-05-01
The categories of low birth weigth infants, social vs. racial factors, factors increasing the risk of low birth weight, prevention of low birth weight, social factors in the development of low birth weight children, the influence of social factors vs. other variables, and implications for management are reviewed. In 1948 the World Health Assembly designated children who were born weighing 2500 g or less as "immature" and further stated that a liveborn infant with a period of gestation of less than 37 weeks or specified as "premature" may be considered as the equivalent of an immature event. In 1961 it was recommended that babies weighing 2500 g or less should no longer be referred to as being "premature" and that the concept of "prematurity" in the definition should give way to that of "low birth weight." Intrauterine growth curves for liveborn males and females were devised from data on birth weight and gestational age. Infants born prior to 37 completed weeks of gestation whose weight lies between the 10th and 90th percentiles on such curves may be called preterm with a weight appropriate for gestational age (AGA), whereas infants born after any length of gestation whose birth weight is at or below the 10th percentile may be named hypotrophic or small for gestational age (SGA). On a worldwide scale it has been estimated that about 22 million low birth weight babies, representing roughly 1/6 of all births, are born alive each day. Only about 1 million of them (mostly preterm) are born in developed countries; of the 21 million born in developing areas, roughly 16 million are SGA full-term and not preterm babies. Socioeconomic status appears as 1 of the most important dterminants of the ultimate level of brain function in children of low birth weight, and this is true with respect to neurologic, psychologic, and educational outcome. Social class also has an indirect effect through birth weight, frequency of perinatal brain injury, and other biological variables as well as maternal habits, the quality of nutrition and health care for mother and child, and other "culture factors." The following seem important facets of the management of low birth weight children: optimal obstetric and perinatal care; "bonding" by parents visiting the intensive care nursery and handling the infant; anticipatory guidance; regular pediatric follow-up for at-risk infants; infant stimulation; early correction of refractive errors, strabismus, other visual defects, hearing defects and orthopedic deformities; and developmental assessments and school readiness tests. PMID:6713335
Local feature weighting in nearest prototype classification.
Fernandez, Fernando; Isasi, Pedro
2008-01-01
The distance metric is the corner stone of nearest neighbor (NN)-based methods, and therefore, of nearest prototype (NP) algorithms. That is because they classify depending on the similarity of the data. When the data is characterized by a set of features which may contribute to the classification task in different levels, feature weighting or selection is required, sometimes in a local sense. However, local weighting is typically restricted to NN approaches. In this paper, we introduce local feature weighting (LFW) in NP classification. LFW provides each prototype its own weight vector, opposite to typical global weighting methods found in the NP literature, where all the prototypes share the same one. Providing each prototype its own weight vector has a novel effect in the borders of the Voronoi regions generated: They become nonlinear. We have integrated LFW with a previously developed evolutionary nearest prototype classifier (ENPC). The experiments performed both in artificial and real data sets demonstrate that the resulting algorithm that we call LFW in nearest prototype classification (LFW-NPC) avoids overfitting on training data in domains where the features may have different contribution to the classification task in different areas of the feature space. This generalization capability is also reflected in automatically obtaining an accurate and reduced set of prototypes. PMID:18269937
The averaging process in permeability estimation from well-test data
Oliver, D.S. )
1990-09-01
Permeability estimates from the pressure derivative or the slope of the semilog plot usually are considered to be averages of some large ill-defined reservoir volume. This paper presents results of a study of the averaging process, including identification of the region of the reservoir that influences permeability estimates, and a specification of the relative contribution of the permeability of various regions to the estimate of average permeability. The diffusion equation for the pressure response of a well situated in an infinite reservoir where permeability is an arbitrary function of position was solved for the case of small variations from a mean value. Permeability estimates from the slope of the plot of pressure vs. the logarithm of drawdown time are shown to be weighted averages of the permeabilities within an inner and outer radius of investigation.
Biodegradation of high molecular weight polylactic acid
NASA Astrophysics Data System (ADS)
Stloukal, Petr; Koutny, Marek; Sedlarik, Vladimir; Kucharczyk, Pavel
2012-07-01
Polylactid acid seems to be an appropriate replacement of conventional non-biodegradable synthetic polymer primarily due to comparable mechanical, thermal and processing properties in its high molecular weight form. Biodegradation of high molecular PLA was studied in compost for various forms differing in their specific surface area. The material proved its good biodegradability under composting conditions and all investigated forms showed to be acceptable for industrial composting. Despite expectations, no significant differences in resulting mineralizations were observed for fiber, film and powder sample forms with different specific surface areas. The clearly faster biodegradation was detected only for the thin coating on porous material with high specific surface area.
Understanding coastal morphodynamic patterns from depth-averaged sediment concentration
NASA Astrophysics Data System (ADS)
Ribas, F.; Falqus, A.; de Swart, H. E.; Dodd, N.; Garnier, R.; Calvete, D.
2015-06-01
This review highlights the important role of the depth-averaged sediment concentration (DASC) to understand the formation of a number of coastal morphodynamic features that have an alongshore rhythmic pattern: beach cusps, surf zone transverse and crescentic bars, and shoreface-connected sand ridges. We present a formulation and methodology, based on the knowledge of the DASC (which equals the sediment load divided by the water depth), that has been successfully used to understand the characteristics of these features. These sand bodies, relevant for coastal engineering and other disciplines, are located in different parts of the coastal zone and are characterized by different spatial and temporal scales, but the same technique can be used to understand them. Since the sand bodies occur in the presence of depth-averaged currents, the sediment transport approximately equals a sediment load times the current. Moreover, it is assumed that waves essentially mobilize the sediment, and the current increases this mobilization and advects the sediment. In such conditions, knowing the spatial distribution of the DASC and the depth-averaged currents induced by the forcing (waves, wind, and pressure gradients) over the patterns allows inferring the convergence/divergence of sediment transport. Deposition (erosion) occurs where the current flows from areas of high to low (low to high) values of DASC. The formulation and methodology are especially useful to understand the positive feedback mechanisms between flow and morphology leading to the formation of those morphological features, but the physical mechanisms for their migration, their finite-amplitude behavior and their decay can also be explored.
Weighted triangulation adjustment
Anderson, Walter L.
1969-01-01
The variation of coordinates method is employed to perform a weighted least squares adjustment of horizontal survey networks. Geodetic coordinates are required for each fixed and adjustable station. A preliminary inverse geodetic position computation is made for each observed line. Weights associated with each observed equation for direction, azimuth, and distance are applied in the formation of the normal equations in-the least squares adjustment. The number of normal equations that may be solved is twice the number of new stations and less than 150. When the normal equations are solved, shifts are produced at adjustable stations. Previously computed correction factors are applied to the shifts and a most probable geodetic position is found for each adjustable station. Pinal azimuths and distances are computed. These may be written onto magnetic tape for subsequent computation of state plane or grid coordinates. Input consists of punch cards containing project identification, program options, and position and observation information. Results listed include preliminary and final positions, residuals, observation equations, solution of the normal equations showing magnitudes of shifts, and a plot of each adjusted and fixed station. During processing, data sets containing irrecoverable errors are rejected and the type of error is listed. The computer resumes processing of additional data sets.. Other conditions cause warning-errors to be issued, and processing continues with the current data set.
Marital status and body weight, weight perception, and weight management among U.S. adults.
Klos, Lori A; Sobal, Jeffery
2013-12-01
Married individuals often have higher body weights than unmarried individuals, but it is unclear how marital roles affect body weight-related perceptions, desires, and behaviors. This study analyzed cross-sectional data for 4,089 adult men and 3,989 adult women using multinomial logistic regression to examine associations between marital status, perceived body weight, desired body weight, and weight management approach. Controlling for demographics and current weight, married or cohabiting women and divorced or separated women more often perceived themselves as overweight and desired to weigh less than women who had never married. Marital status was unrelated to men's weight perception and desired weight change. Marital status was also generally unrelated to weight management approach, except that divorced or separated women were more likely to have intentionally lost weight within the past year compared to never married women. Additionally, never married men were more likely to be attempting to prevent weight gain than married or cohabiting men and widowed men. Overall, married and formerly married women more often perceived themselves as overweight and desired a lower weight. Men's marital status was generally unassociated with weight-related perceptions, desires, and behaviors. Women's but not men's marital roles appear to influence their perceived and desired weight, suggesting that weight management interventions should be sensitive to both marital status and gender differences. PMID:24183145
A Light-Weight Instrumentation System Design
Kidner, Ronald
1999-06-02
To meet challenging constraints on telemetry system weight and volume, a custom Light-Weight Instrumentation System was developed to collect vehicle environment and dynamics on a short-duration exo-atmospheric flight test vehicle. The total telemetry system, including electronics, sensors, batteries, and a 1 watt transmitter weighs about 1 kg. Over 80 channels of measurement, housekeeping, and telemetry system diagnostic data are transmitted at 128 kbps. The microcontroller-based design uses the automotive industry standard Controller Area Network to interface with and support in-flight control fimctions. Operational parameters are downloaded via a standard asynchronous serial communications intefiace. The basic design philosophy and functionality is described here.