High-order scheme for the source-sink term in a one-dimensional water temperature model
Jing, Zheng; Kang, Ling
2017-01-01
The source-sink term in water temperature models represents the net heat absorbed or released by a water system. This term is very important because it accounts for solar radiation that can significantly affect water temperature, especially in lakes. However, existing numerical methods for discretizing the source-sink term are very simplistic, causing significant deviations between simulation results and measured data. To address this problem, we present a numerical method specific to the source-sink term. A vertical one-dimensional heat conduction equation was chosen to describe water temperature changes. A two-step operator-splitting method was adopted as the numerical solution. In the first step, using the undetermined coefficient method, a high-order scheme was adopted for discretizing the source-sink term. In the second step, the diffusion term was discretized using the Crank-Nicolson scheme. The effectiveness and capability of the numerical method was assessed by performing numerical tests. Then, the proposed numerical method was applied to a simulation of Guozheng Lake (located in central China). The modeling results were in an excellent agreement with measured data. PMID:28264005
High-order scheme for the source-sink term in a one-dimensional water temperature model.
Jing, Zheng; Kang, Ling
2017-01-01
The source-sink term in water temperature models represents the net heat absorbed or released by a water system. This term is very important because it accounts for solar radiation that can significantly affect water temperature, especially in lakes. However, existing numerical methods for discretizing the source-sink term are very simplistic, causing significant deviations between simulation results and measured data. To address this problem, we present a numerical method specific to the source-sink term. A vertical one-dimensional heat conduction equation was chosen to describe water temperature changes. A two-step operator-splitting method was adopted as the numerical solution. In the first step, using the undetermined coefficient method, a high-order scheme was adopted for discretizing the source-sink term. In the second step, the diffusion term was discretized using the Crank-Nicolson scheme. The effectiveness and capability of the numerical method was assessed by performing numerical tests. Then, the proposed numerical method was applied to a simulation of Guozheng Lake (located in central China). The modeling results were in an excellent agreement with measured data.
Observation-based source terms in the third-generation wave model WAVEWATCH
NASA Astrophysics Data System (ADS)
Zieger, Stefan; Babanin, Alexander V.; Erick Rogers, W.; Young, Ian R.
2015-12-01
Measurements collected during the AUSWEX field campaign, at Lake George (Australia), resulted in new insights into the processes of wind wave interaction and whitecapping dissipation, and consequently new parameterizations of the input and dissipation source terms. The new nonlinear wind input term developed accounts for dependence of the growth on wave steepness, airflow separation, and for negative growth rate under adverse winds. The new dissipation terms feature the inherent breaking term, a cumulative dissipation term and a term due to production of turbulence by waves, which is particularly relevant for decaying seas and for swell. The latter is consistent with the observed decay rate of ocean swell. This paper describes these source terms implemented in WAVEWATCH III ®and evaluates the performance against existing source terms in academic duration-limited tests, against buoy measurements for windsea-dominated conditions, under conditions of extreme wind forcing (Hurricane Katrina), and against altimeter data in global hindcasts. Results show agreement by means of growth curves as well as integral and spectral parameters in the simulations and hindcast.
Bayesian source term determination with unknown covariance of measurements
NASA Astrophysics Data System (ADS)
Belal, Alkomiet; Tichý, Ondřej; Šmídl, Václav
2017-04-01
Determination of a source term of release of a hazardous material into the atmosphere is a very important task for emergency response. We are concerned with the problem of estimation of the source term in the conventional linear inverse problem, y = Mx, where the relationship between the vector of observations y is described using the source-receptor-sensitivity (SRS) matrix M and the unknown source term x. Since the system is typically ill-conditioned, the problem is recast as an optimization problem minR,B(y - Mx)TR-1(y - Mx) + xTB-1x. The first term minimizes the error of the measurements with covariance matrix R, and the second term is a regularization of the source term. There are different types of regularization arising for different choices of matrices R and B, for example, Tikhonov regularization assumes covariance matrix B as the identity matrix multiplied by scalar parameter. In this contribution, we adopt a Bayesian approach to make inference on the unknown source term x as well as unknown R and B. We assume prior on x to be a Gaussian with zero mean and unknown diagonal covariance matrix B. The covariance matrix of the likelihood R is also unknown. We consider two potential choices of the structure of the matrix R. First is the diagonal matrix and the second is a locally correlated structure using information on topology of the measuring network. Since the inference of the model is intractable, iterative variational Bayes algorithm is used for simultaneous estimation of all model parameters. The practical usefulness of our contribution is demonstrated on an application of the resulting algorithm to real data from the European Tracer Experiment (ETEX). This research is supported by EEA/Norwegian Financial Mechanism under project MSMT-28477/2014 Source-Term Determination of Radionuclide Releases by Inverse Atmospheric Dispersion Modelling (STRADI).
Interlaboratory study of the ion source memory effect in 36Cl accelerator mass spectrometry
NASA Astrophysics Data System (ADS)
Pavetich, Stefan; Akhmadaliev, Shavkat; Arnold, Maurice; Aumaître, Georges; Bourlès, Didier; Buchriegler, Josef; Golser, Robin; Keddadouche, Karim; Martschini, Martin; Merchel, Silke; Rugel, Georg; Steier, Peter
2014-06-01
Understanding and minimization of contaminations in the ion source due to cross-contamination and long-term memory effect is one of the key issues for accurate accelerator mass spectrometry (AMS) measurements of volatile elements. The focus of this work is on the investigation of the long-term memory effect for the volatile element chlorine, and the minimization of this effect in the ion source of the Dresden accelerator mass spectrometry facility (DREAMS). For this purpose, one of the two original HVE ion sources at the DREAMS facility was modified, allowing the use of larger sample holders having individual target apertures. Additionally, a more open geometry was used to improve the vacuum level. To evaluate this improvement in comparison to other up-to-date ion sources, an interlaboratory comparison had been initiated. The long-term memory effect of the four Cs sputter ion sources at DREAMS (two sources: original and modified), ASTER (Accélérateur pour les Sciences de la Terre, Environnement, Risques) and VERA (Vienna Environmental Research Accelerator) had been investigated by measuring samples of natural 35Cl/37Cl-ratio and samples highly-enriched in 35Cl (35Cl/37Cl ∼ 999). Besides investigating and comparing the individual levels of long-term memory, recovery time constants could be calculated. The tests show that all four sources suffer from long-term memory, but the modified DREAMS ion source showed the lowest level of contamination. The recovery times of the four ion sources were widely spread between 61 and 1390 s, where the modified DREAMS ion source with values between 156 and 262 s showed the fastest recovery in 80% of the measurements.
NASA Astrophysics Data System (ADS)
Winiarek, Victor; Bocquet, Marc; Duhanyan, Nora; Roustan, Yelva; Saunier, Olivier; Mathieu, Anne
2014-01-01
Inverse modelling techniques can be used to estimate the amount of radionuclides and the temporal profile of the source term released in the atmosphere during the accident of the Fukushima Daiichi nuclear power plant in March 2011. In Winiarek et al. (2012b), the lower bounds of the caesium-137 and iodine-131 source terms were estimated with such techniques, using activity concentration measurements. The importance of an objective assessment of prior errors (the observation errors and the background errors) was emphasised for a reliable inversion. In such critical context where the meteorological conditions can make the source term partly unobservable and where only a few observations are available, such prior estimation techniques are mandatory, the retrieved source term being very sensitive to this estimation. We propose to extend the use of these techniques to the estimation of prior errors when assimilating observations from several data sets. The aim is to compute an estimate of the caesium-137 source term jointly using all available data about this radionuclide, such as activity concentrations in the air, but also daily fallout measurements and total cumulated fallout measurements. It is crucial to properly and simultaneously estimate the background errors and the prior errors relative to each data set. A proper estimation of prior errors is also a necessary condition to reliably estimate the a posteriori uncertainty of the estimated source term. Using such techniques, we retrieve a total released quantity of caesium-137 in the interval 11.6-19.3 PBq with an estimated standard deviation range of 15-20% depending on the method and the data sets. The “blind” time intervals of the source term have also been strongly mitigated compared to the first estimations with only activity concentration data.
Next Generation Emission Measurements for Fugitive, Area Source, and Fence Line Applications?
Next generation emissions measurements (NGEM) is an EPA term for the rapidly advancing field of air pollutant sensor technologies, data integration concepts, and associated geospatial modeling strategies for source emissions measurements. Ranging from low coat sensors to satelli...
ERIC Educational Resources Information Center
Littlejohn, Emily
2018-01-01
"Adaptation" originally began as a scientific term, but from 1860 to today it most often refers to an altered version of a text, film, or other literary source. When this term was first analyzed, humanities scholars often measured adaptations against their source texts, frequently privileging "original" texts. However, this…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Eslinger, Paul W.; Biegalski, S.; Bowyer, Ted W.
2014-01-01
Systems designed to monitor airborne radionuclides released from underground nuclear explosions detected radioactive fallout from the Fukushima Daiichi nuclear accident in March 2011. Atmospheric transport modeling (ATM) of plumes of noble gases and particulates were performed soon after the accident to determine plausible detection locations of any radioactive releases to the atmosphere. We combine sampling data from multiple International Modeling System (IMS) locations in a new way to estimate the magnitude and time sequence of the releases. Dilution factors from the modeled plume at five different detection locations were combined with 57 atmospheric concentration measurements of 133-Xe taken from Marchmore » 18 to March 23 to estimate the source term. This approach estimates that 59% of the 1.24×1019 Bq of 133-Xe present in the reactors at the time of the earthquake was released to the atmosphere over a three day period. Source term estimates from combinations of detection sites have lower spread than estimates based on measurements at single detection sites. Sensitivity cases based on data from four or more detection locations bound the source term between 35% and 255% of available xenon inventory.« less
NASA Astrophysics Data System (ADS)
Saunier, Olivier; Mathieu, Anne; Didier, Damien; Tombette, Marilyne; Quélo, Denis; Winiarek, Victor; Bocquet, Marc
2013-04-01
The Chernobyl nuclear accident and more recently the Fukushima accident highlighted that the largest source of error on consequences assessment is the source term estimation including the time evolution of the release rate and its distribution between radioisotopes. Inverse modelling methods have proved to be efficient to assess the source term due to accidental situation (Gudiksen, 1989, Krysta and Bocquet, 2007, Stohl et al 2011, Winiarek et al 2012). These methods combine environmental measurements and atmospheric dispersion models. They have been recently applied to the Fukushima accident. Most existing approaches are designed to use air sampling measurements (Winiarek et al, 2012) and some of them use also deposition measurements (Stohl et al, 2012, Winiarek et al, 2013). During the Fukushima accident, such measurements are far less numerous and not as well distributed within Japan than the dose rate measurements. To efficiently document the evolution of the contamination, gamma dose rate measurements were numerous, well distributed within Japan and they offered a high temporal frequency. However, dose rate data are not as easy to use as air sampling measurements and until now they were not used in inverse modelling approach. Indeed, dose rate data results from all the gamma emitters present in the ground and in the atmosphere in the vicinity of the receptor. They do not allow one to determine the isotopic composition or to distinguish the plume contribution from wet deposition. The presented approach proposes a way to use dose rate measurement in inverse modeling approach without the need of a-priori information on emissions. The method proved to be efficient and reliable when applied on the Fukushima accident. The emissions for the 8 main isotopes Xe-133, Cs-134, Cs-136, Cs-137, Ba-137m, I-131, I-132 and Te-132 have been assessed. The Daiichi power plant events (such as ventings, explosions…) known to have caused atmospheric releases are well identified in the retrieved source term, except for unit 3 explosion where no measurement was available. The comparisons between the simulations of atmospheric dispersion and deposition of the retrieved source term show a good agreement with environmental observations. Moreover, an important outcome of this study is that the method proved to be perfectly suited to crisis management and should contribute to improve our response in case of a nuclear accident.
Low birth weight and air pollution in California: Which sources and components drive the risk?
Laurent, Olivier; Hu, Jianlin; Li, Lianfa; Kleeman, Michael J; Bartell, Scott M; Cockburn, Myles; Escobedo, Loraine; Wu, Jun
2016-01-01
Intrauterine growth restriction has been associated with exposure to air pollution, but there is a need to clarify which sources and components are most likely responsible. This study investigated the associations between low birth weight (LBW, <2500g) in term born infants (≥37 gestational weeks) and air pollution by source and composition in California, over the period 2001-2008. Complementary exposure models were used: an empirical Bayesian kriging model for the interpolation of ambient pollutant measurements, a source-oriented chemical transport model (using California emission inventories) that estimated fine and ultrafine particulate matter (PM2.5 and PM0.1, respectively) mass concentrations (4km×4km) by source and composition, a line-source roadway dispersion model at fine resolution, and traffic index estimates. Birth weight was obtained from California birth certificate records. A case-cohort design was used. Five controls per term LBW case were randomly selected (without covariate matching or stratification) from among term births. The resulting datasets were analyzed by logistic regression with a random effect by hospital, using generalized additive mixed models adjusted for race/ethnicity, education, maternal age and household income. In total 72,632 singleton term LBW cases were included. Term LBW was positively and significantly associated with interpolated measurements of ozone but not total fine PM or nitrogen dioxide. No significant association was observed between term LBW and primary PM from all sources grouped together. A positive significant association was observed for secondary organic aerosols. Exposure to elemental carbon (EC), nitrates and ammonium were also positively and significantly associated with term LBW, but only for exposure during the third trimester of pregnancy. Significant positive associations were observed between term LBW risk and primary PM emitted by on-road gasoline and diesel or by commercial meat cooking sources. Primary PM from wood burning was inversely associated with term LBW. Significant positive associations were also observed between term LBW and ultrafine particle numbers modeled with the line-source roadway dispersion model, traffic density and proximity to roadways. This large study based on complementary exposure metrics suggests that not only primary pollution sources (traffic and commercial meat cooking) but also EC and secondary pollutants are risk factors for term LBW. Copyright © 2016 Elsevier Ltd. All rights reserved.
Directional Unfolded Source Term (DUST) for Compton Cameras.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mitchell, Dean J.; Horne, Steven M.; O'Brien, Sean
2018-03-01
A Directional Unfolded Source Term (DUST) algorithm was developed to enable improved spectral analysis capabilities using data collected by Compton cameras. Achieving this objective required modification of the detector response function in the Gamma Detector Response and Analysis Software (GADRAS). Experimental data that were collected in support of this work include measurements of calibration sources at a range of separation distances and cylindrical depleted uranium castings.
Long-term variability in bright hard X-ray sources: 5+ years of BATSE data
NASA Technical Reports Server (NTRS)
Robinson, C. R.; Harmon, B. A.; McCollough, M. L.; Paciesas, W. S.; Sahi, M.; Scott, D. M.; Wilson, C. A.; Zhang, S. N.; Deal, K. J.
1997-01-01
The operation of the Compton Gamma Ray Observatory (CGRO)/burst and transient source experiment (BATSE) continues to provide data for inclusion into a data base for the analysis of long term variability in bright, hard X-ray sources. The all-sky capability of BATSE provides up to 30 flux measurements/day for each source. The long baseline and the various rising and setting occultation flux measurements allow searches for periodic and quasi-periodic signals with periods of between several hours to hundreds of days to be conducted. The preliminary results from an analysis of the hard X-ray variability in 24 of the brightest BATSE sources are presented. Power density spectra are computed for each source and profiles are presented of the hard X-ray orbital modulations in some X-ray binaries, together with amplitude modulations and variations in outburst durations and intensities in recurrent X-ray transients.
REVIEW OF METHODS FOR REMOTE SENSING OF ATMOSPHERIC EMISSIONS FROM STATIONARY SOURCES
The report reviews the commercially available and developing technologies for the application of remote sensing to the measurement of source emissions. The term 'remote sensing technology', as applied in the report, means the detection or concentration measurement of trace atmosp...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ford, S R; Dreger, D S; Phillips, W S
2008-07-16
Inversions for regional attenuation (1/Q) of Lg are performed in two different regions. The path attenuation component of the Lg spectrum is isolated using the coda-source normalization method, which corrects the Lg spectral amplitude for the source using the stable, coda-derived source spectra. Tomographic images of Northern California agree well with one-dimensional (1-D) Lg Q estimated from five different methods. We note there is some tendency for tomographic smoothing to increase Q relative to targeted 1-D methods. For example in the San Francisco Bay Area, which contains high attenuation relative to the rest of it's region, Q is over-estimated bymore » {approx}30. Coda-source normalized attenuation tomography is also carried out for the Yellow Sea/Korean Peninsula (YSKP) where output parameters (site, source, and path terms) are compared with those from the amplitude tomography method of Phillips et al. (2005) as well as a new method that ties the source term to the MDAC formulation (Walter and Taylor, 2001). The source terms show similar scatter between coda-source corrected and MDAC source perturbation methods, whereas the amplitude method has the greatest correlation with estimated true source magnitude. The coda-source better represents the source spectra compared to the estimated magnitude and could be the cause of the scatter. The similarity in the source terms between the coda-source and MDAC-linked methods shows that the latter method may approximate the effect of the former, and therefore could be useful in regions without coda-derived sources. The site terms from the MDAC-linked method correlate slightly with global Vs30 measurements. While the coda-source and amplitude ratio methods do not correlate with Vs30 measurements, they do correlate with one another, which provides confidence that the two methods are consistent. The path Q{sup -1} values are very similar between the coda-source and amplitude ratio methods except for small differences in the Da-xin-anling Mountains, in the northern YSKP. However there is one large difference between the MDAC-linked method and the others in the region near stations TJN and INCN, which point to site-effect as the cause for the difference.« less
Watershed nitrogen and phosphorus balance: The upper Potomac River basin
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jaworski, N.A.; Groffman, P.M.; Keller, A.A.
1992-01-01
Nitrogen and phosphorus mass balances were estimated for the portion of the Potomac River basin watershed located above Washington, D.C. The total nitrogen (N) balance included seven input source terms, six sinks, and one 'change-in-storage' term, but was simplified to five input terms and three output terms. The phosphorus (P) baance had four input and three output terms. The estimated balances are based on watershed data from seven information sources. Major sources of nitrogen are animal waste and atmospheric deposition. The major sources of phosphorus are animal waste and fertilizer. The major sink for nitrogen is combined denitrification, volatilization, andmore » change-in-storage. The major sink for phosphorus is change-in-storage. River exports of N and P were 17% and 8%, respectively, of the total N and P inputs. Over 60% of the N and P were volatilized or stored. The major input and output terms on the budget are estimated from direct measurements, but the change-in-storage term is calculated by difference. The factors regulating retention and storage processes are discussed and research needs are identified.« less
Henry, S B; Holzemer, W L; Reilly, C A; Campbell, K E
1994-01-01
OBJECTIVE: To analyze the terms used by nurses in a variety of data sources and to test the feasibility of using SNOMED III to represent nursing terms. DESIGN: Prospective research design with manual matching of terms to the SNOMED III vocabulary. MEASUREMENTS: The terms used by nurses to describe patient problems during 485 episodes of care for 201 patients hospitalized for Pneumocystis carinii pneumonia were identified. Problems from four data sources (nurse interview, intershift report, nursing care plan, and nurse progress note/flowsheet) were classified based on the substantive area of the problem and on the terminology used to describe the problem. A test subset of the 25 most frequently used terms from the two written data sources (nursing care plan and nurse progress note/flowsheet) were manually matched to SNOMED III terms to test the feasibility of using that existing vocabulary to represent nursing terms. RESULTS: Nurses most frequently described patient problems as signs/symptoms in the verbal nurse interview and intershift report. In the written data sources, problems were recorded as North American Nursing Diagnosis Association (NANDA) terms and signs/symptoms with similar frequencies. Of the nursing terms in the test subset, 69% were represented using one or more SNOMED III terms. PMID:7719788
Common Calibration Source for Monitoring Long-term Ozone Trends
NASA Technical Reports Server (NTRS)
Kowalewski, Matthew
2004-01-01
Accurate long-term satellite measurements are crucial for monitoring the recovery of the ozone layer. The slow pace of the recovery and limited lifetimes of satellite monitoring instruments demands that datasets from multiple observation systems be combined to provide the long-term accuracy needed. A fundamental component of accurately monitoring long-term trends is the calibration of these various instruments. NASA s Radiometric Calibration and Development Facility at the Goddard Space Flight Center has provided resources to minimize calibration biases between multiple instruments through the use of a common calibration source and standardized procedures traceable to national standards. The Facility s 50 cm barium sulfate integrating sphere has been used as a common calibration source for both US and international satellite instruments, including the Total Ozone Mapping Spectrometer (TOMS), Solar Backscatter Ultraviolet 2 (SBUV/2) instruments, Shuttle SBUV (SSBUV), Ozone Mapping Instrument (OMI), Global Ozone Monitoring Experiment (GOME) (ESA), Scanning Imaging SpectroMeter for Atmospheric ChartographY (SCIAMACHY) (ESA), and others. We will discuss the advantages of using a common calibration source and its effects on long-term ozone data sets. In addition, sphere calibration results from various instruments will be presented to demonstrate the accuracy of the long-term characterization of the source itself.
Bayesian estimation of a source term of radiation release with approximately known nuclide ratios
NASA Astrophysics Data System (ADS)
Tichý, Ondřej; Šmídl, Václav; Hofman, Radek
2016-04-01
We are concerned with estimation of a source term in case of an accidental release from a known location, e.g. a power plant. Usually, the source term of an accidental release of radiation comprises of a mixture of nuclide. The gamma dose rate measurements do not provide a direct information on the source term composition. However, physical properties of respective nuclide (deposition properties, decay half-life) can be used when uncertain information on nuclide ratios is available, e.g. from known reactor inventory. The proposed method is based on linear inverse model where the observation vector y arise as a linear combination y = Mx of a source-receptor-sensitivity (SRS) matrix M and the source term x. The task is to estimate the unknown source term x. The problem is ill-conditioned and further regularization is needed to obtain a reasonable solution. In this contribution, we assume that nuclide ratios of the release is known with some degree of uncertainty. This knowledge is used to form the prior covariance matrix of the source term x. Due to uncertainty in the ratios the diagonal elements of the covariance matrix are considered to be unknown. Positivity of the source term estimate is guaranteed by using multivariate truncated Gaussian distribution. Following Bayesian approach, we estimate all parameters of the model from the data so that y, M, and known ratios are the only inputs of the method. Since the inference of the model is intractable, we follow the Variational Bayes method yielding an iterative algorithm for estimation of all model parameters. Performance of the method is studied on simulated 6 hour power plant release where 3 nuclide are released and 2 nuclide ratios are approximately known. The comparison with method with unknown nuclide ratios will be given to prove the usefulness of the proposed approach. This research is supported by EEA/Norwegian Financial Mechanism under project MSMT-28477/2014 Source-Term Determination of Radionuclide Releases by Inverse Atmospheric Dispersion Modelling (STRADI).
A microchip laser source with stable intensity and frequency used for self-mixing interferometry.
Zhang, Shaohui; Zhang, Shulian; Tan, Yidong; Sun, Liqun
2016-05-01
We present a stable 40 × 40 × 30 mm(3) Laser-diode (LD)-pumped-microchip laser (ML) laser source used for self-mixing interferometry which can measure non-cooperative targets. We simplify the coupling process of pump light in order to make its polarization and intensity robust against environmental disturbance. Thermal frequency stabilization technology is used to stabilize the laser frequency of both LD and ML. Frequency stability of about 1 × 10(-7) and short-term intensity fluctuation of 0.1% are achieved. The theoretical long-term displacement accuracy limited by frequency and intensity fluctuation is about 10 nm when the measuring range is 0.1 m. The line-width of this laser is about 25 kHz corresponding to 12 km coherent length and 6 km measurement range for self-mixing interference. The laser source has been equipped to a self-mixing interferometer, and it works very well.
NASA Astrophysics Data System (ADS)
Karamehmedović, Mirza; Kirkeby, Adrian; Knudsen, Kim
2018-06-01
We consider the multi-frequency inverse source problem for the scalar Helmholtz equation in the plane. The goal is to reconstruct the source term in the equation from measurements of the solution on a surface outside the support of the source. We study the problem in a certain finite dimensional setting: from measurements made at a finite set of frequencies we uniquely determine and reconstruct sources in a subspace spanned by finitely many Fourier–Bessel functions. Further, we obtain a constructive criterion for identifying a minimal set of measurement frequencies sufficient for reconstruction, and under an additional, mild assumption, the reconstruction method is shown to be stable. Our analysis is based on a singular value decomposition of the source-to-measurement forward operators and the distribution of positive zeros of the Bessel functions of the first kind. The reconstruction method is implemented numerically and our theoretical findings are supported by numerical experiments.
Three-Dimensional Model Synthesis of the Global Methane Cycle
NASA Technical Reports Server (NTRS)
Fung, I.; Prather, M.; John, J.; Lerner, J.; Matthews, E.
1991-01-01
A synthesis of the global methane cycle is presented to attempt to generate an accurate global methane budget. Methane-flux measurements, energy data, and agricultural statistics are merged with databases of land-surface characteristics and anthropogenic activities. The sources and sinks of methane are estimated based on atmospheric methane composition and variations, and a global 3D transport model simulates the corresponding atmospheric responses. The geographic and seasonal variations of candidate budgets are compared with observational data, and the available observations are used to constrain the plausible methane budgets. The preferred budget includes annual destruction rates and annual emissions for various sources. The lack of direct flux measurements in the regions of many of these fluxes makes the unique determination of each term impossible. OH oxidation is found to be the largest single term, although more measurements of this and other terms are recommended.
NASA Astrophysics Data System (ADS)
Taha, M. P. M.; Drew, G. H.; Longhurst, P. J.; Smith, R.; Pollard, S. J. T.
The passive and active release of bioaerosols during green waste composting, measured at source is reported for a commercial composting facility in South East (SE) England as part of a research programme focused on improving risk assessments at composting facilities. Aspergillus fumigatus and actinomycetes concentrations of 9.8-36.8×10 6 and 18.9-36.0×10 6 cfu m -3, respectively, measured during the active turning of green waste compost, were typically 3-log higher than previously reported concentrations from static compost windrows. Source depletion curves constructed for A. fumigatus during compost turning and modelled using SCREEN3 suggest that bioaerosol concentrations could reduce to background concentrations of 10 3 cfu m -3 within 100 m of this site. Authentic source term data produced from this study will help to refine the risk assessment methodologies that support improved permitting of compost facilities.
Performance evaluation of WAVEWATCH III model in the Persian Gulf using different wind resources
NASA Astrophysics Data System (ADS)
Kazeminezhad, Mohammad Hossein; Siadatmousavi, Seyed Mostafa
2017-07-01
The third-generation wave model, WAVEWATCH III, was employed to simulate bulk wave parameters in the Persian Gulf using three different wind sources: ERA-Interim, CCMP, and GFS-Analysis. Different formulations for whitecapping term and the energy transfer from wind to wave were used, namely the Tolman and Chalikov (J Phys Oceanogr 26:497-518, 1996), WAM cycle 4 (BJA and WAM4), and Ardhuin et al. (J Phys Oceanogr 40(9):1917-1941, 2010) (TEST405 and TEST451 parameterizations) source term packages. The obtained results from numerical simulations were compared to altimeter-derived significant wave heights and measured wave parameters at two stations in the northern part of the Persian Gulf through statistical indicators and the Taylor diagram. Comparison of the bulk wave parameters with measured values showed underestimation of wave height using all wind sources. However, the performance of the model was best when GFS-Analysis wind data were used. In general, when wind veering from southeast to northwest occurred, and wind speed was high during the rotation, the model underestimation of wave height was severe. Except for the Tolman and Chalikov (J Phys Oceanogr 26:497-518, 1996) source term package, which severely underestimated the bulk wave parameters during stormy condition, the performances of other formulations were practically similar. However, in terms of statistics, the Ardhuin et al. (J Phys Oceanogr 40(9):1917-1941, 2010) source terms with TEST405 parameterization were the most successful formulation in the Persian Gulf when compared to in situ and altimeter-derived observations.
Source Credibility in Tobacco Control Messaging
Schmidt, Allison M.; Ranney, Leah M.; Pepper, Jessica K.; Goldstein, Adam O.
2016-01-01
Objectives Perceived credibility of a message’s source can affect persuasion. This paper reviews how beliefs about the source of tobacco control messages may encourage attitude and behavior change. Methods We conducted a series of searches of the peer-reviewed literature using terms from communication and public health fields. We reviewed research on source credibility, its underlying concepts, and its relation to the persuasiveness of tobacco control messages. Results We recommend an agenda for future research to bridge the gaps between communication literature on source credibility and tobacco control research. Our recommendations are to study the impact of source credibility on persuasion with long-term behavior change outcomes, in different populations and demographic groups, by developing new credibility measures that are topic- and organization-specific, by measuring how credibility operates across media platforms, and by identifying factors that enhance credibility and persuasion. Conclusions This manuscript reviews the state of research on source credibility and identifies gaps that are maximally relevant to tobacco control communication. Knowing first whether a source is perceived as credible, and second, how to enhance perceived credibility, can inform the development of future tobacco control campaigns and regulatory communications. PMID:27525298
He, Xiaowei; Liang, Jimin; Wang, Xiaorui; Yu, Jingjing; Qu, Xiaochao; Wang, Xiaodong; Hou, Yanbin; Chen, Duofang; Liu, Fang; Tian, Jie
2010-11-22
In this paper, we present an incomplete variables truncated conjugate gradient (IVTCG) method for bioluminescence tomography (BLT). Considering the sparse characteristic of the light source and insufficient surface measurement in the BLT scenarios, we combine a sparseness-inducing (ℓ1 norm) regularization term with a quadratic error term in the IVTCG-based framework for solving the inverse problem. By limiting the number of variables updated at each iterative and combining a variable splitting strategy to find the search direction more efficiently, it obtains fast and stable source reconstruction, even without a priori information of the permissible source region and multispectral measurements. Numerical experiments on a mouse atlas validate the effectiveness of the method. In vivo mouse experimental results further indicate its potential for a practical BLT system.
NASA Astrophysics Data System (ADS)
Roustan, Yelva; Duhanyan, Nora; Bocquet, Marc; Winiarek, Victor
2013-04-01
A sensitivity study of the numerical model, as well as, an inverse modelling approach applied to the atmospheric dispersion issues after the Chernobyl disaster are both presented in this paper. On the one hand, the robustness of the source term reconstruction through advanced data assimilation techniques was tested. On the other hand, the classical approaches for sensitivity analysis were enhanced by the use of an optimised forcing field which otherwise is known to be strongly uncertain. The POLYPHEMUS air quality system was used to perform the simulations of radionuclide dispersion. Activity concentrations in air and deposited to the ground of iodine-131, caesium-137 and caesium-134 were considered. The impact of the implemented parameterizations of the physical processes (dry and wet depositions, vertical turbulent diffusion), of the forcing fields (meteorology and source terms) and of the numerical configuration (horizontal resolution) were investigated for the sensitivity study of the model. A four dimensional variational scheme (4D-Var) based on the approximate adjoint of the chemistry transport model was used to invert the source term. The data assimilation is performed with measurements of activity concentrations in air extracted from the Radioactivity Environmental Monitoring (REM) database. For most of the investigated configurations (sensitivity study), the statistics to compare the model results to the field measurements as regards the concentrations in air are clearly improved while using a reconstructed source term. As regards the ground deposited concentrations, an improvement can only be seen in case of satisfactorily modelled episode. Through these studies, the source term and the meteorological fields are proved to have a major impact on the activity concentrations in air. These studies also reinforce the use of reconstructed source term instead of the usual estimated one. A more detailed parameterization of the deposition process seems also to be able to improve the simulation results. For deposited activities the results are more complex probably due to a strong sensitivity to some of the meteorological fields which remain quite uncertain.
NASA Astrophysics Data System (ADS)
Evangeliou, Nikolaos; Hamburger, Thomas; Cozic, Anne; Balkanski, Yves; Stohl, Andreas
2017-07-01
This paper describes the results of an inverse modeling study for the determination of the source term of the radionuclides 134Cs, 137Cs and 131I released after the Chernobyl accident. The accident occurred on 26 April 1986 in the Former Soviet Union and released about 1019 Bq of radioactive materials that were transported as far away as the USA and Japan. Thereafter, several attempts to assess the magnitude of the emissions were made that were based on the knowledge of the core inventory and the levels of the spent fuel. More recently, when modeling tools were further developed, inverse modeling techniques were applied to the Chernobyl case for source term quantification. However, because radioactivity is a sensitive topic for the public and attracts a lot of attention, high-quality measurements, which are essential for inverse modeling, were not made available except for a few sparse activity concentration measurements far from the source and far from the main direction of the radioactive fallout. For the first time, we apply Bayesian inversion of the Chernobyl source term using not only activity concentrations but also deposition measurements from the most recent public data set. These observations refer to a data rescue attempt that started more than 10 years ago, with a final goal to provide available measurements to anyone interested. In regards to our inverse modeling results, emissions of 134Cs were estimated to be 80 PBq or 30-50 % higher than what was previously published. From the released amount of 134Cs, about 70 PBq were deposited all over Europe. Similar to 134Cs, emissions of 137Cs were estimated as 86 PBq, on the same order as previously reported results. Finally, 131I emissions of 1365 PBq were found, which are about 10 % less than the prior total releases. The inversion pushes the injection heights of the three radionuclides to higher altitudes (up to about 3 km) than previously assumed (≈ 2.2 km) in order to better match both concentration and deposition observations over Europe. The results of the present inversion were confirmed using an independent Eulerian model, for which deposition patterns were also improved when using the estimated posterior releases. Although the independent model tends to underestimate deposition in countries that are not in the main direction of the plume, it reproduces country levels of deposition very efficiently. The results were also tested for robustness against different setups of the inversion through sensitivity runs. The source term data from this study are publicly available.
Fiber-coupled pillar array as a highly pure and stable single-photon source
NASA Astrophysics Data System (ADS)
Odashima, S.; Sasakura, H.; Nakajima, H.; Kumano, H.
2017-12-01
A highly pure and stable single-photon source is prepared that comprises a well-designed pillar array, in which each pillar contains only a few InAs quantum dots. A nano-pillar in this array is in direct contact with a fiber end surface and cooled in a liquid-He bath. Auto-correlation measurements show that this source provides an average g(2)(0) value of 0.0174 in the measured excitation-power range. This photon source and fiber coupling are quite rigid against external disturbances such as cooling-heating cycles and vibration, with long-term stability.
NASA Technical Reports Server (NTRS)
Gnanalingam, S.; Kane, J. A.
1973-01-01
An extensive set of ground-based measurements of the diurnal variation of medium frequency radio wave adsorption and virtual height is analyzed in terms of current understanding of the D- and lower E-region ion production and loss process. When this is done a gross discrepancy arises, the source of which is not known.
Galactic water vapor emission: further observations of variability.
Knowles, S H; Mayer, C H; Sullivan, W T; Cheung, A C
1969-10-10
Recent observations of the 1.35-centimeter line emission of water vapor from galactic sources show short-term variability in the spectra of several sources. Two additional sources, Cygnus 1 and NGC 6334N, have been observed, and the spectra of W49 and VY Canis Majoris were measured over a wider range of radial velocity.
NASA Astrophysics Data System (ADS)
Delpueyo, D.; Balandraud, X.; Grédiac, M.
2013-09-01
The aim of this paper is to present a post-processing technique based on a derivative Gaussian filter to reconstruct heat source fields from temperature fields measured by infrared thermography. Heat sources can be deduced from temperature variations thanks to the heat diffusion equation. Filtering and differentiating are key-issues which are closely related here because the temperature fields which are processed are unavoidably noisy. We focus here only on the diffusion term because it is the most difficult term to estimate in the procedure, the reason being that it involves spatial second derivatives (a Laplacian for isotropic materials). This quantity can be reasonably estimated using a convolution of the temperature variation fields with second derivatives of a Gaussian function. The study is first based on synthetic temperature variation fields corrupted by added noise. The filter is optimised in order to reconstruct at best the heat source fields. The influence of both the dimension and the level of a localised heat source is discussed. Obtained results are also compared with another type of processing based on an averaging filter. The second part of this study presents an application to experimental temperature fields measured with an infrared camera on a thin plate in aluminium alloy. Heat sources are generated with an electric heating patch glued on the specimen surface. Heat source fields reconstructed from measured temperature fields are compared with the imposed heat sources. Obtained results illustrate the relevancy of the derivative Gaussian filter to reliably extract heat sources from noisy temperature fields for the experimental thermomechanics of materials.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cabrera-Palmer, Belkis
Predicting the performance of radiation detection systems at field sites based on measured performance acquired under controlled conditions at test locations, e.g., the Nevada National Security Site (NNSS), remains an unsolved and standing issue within DNDO’s testing methodology. Detector performance can be defined in terms of the system’s ability to detect and/or identify a given source or set of sources, and depends on the signal generated by the detector for the given measurement configuration (i.e., source strength, distance, time, surrounding materials, etc.) and on the quality of the detection algorithm. Detector performance is usually evaluated in the performance and operationalmore » testing phases, where the measurement configurations are selected to represent radiation source and background configurations of interest to security applications.« less
Phenomenological Modeling of Infrared Sources: Recent Advances
NASA Technical Reports Server (NTRS)
Leung, Chun Ming; Kwok, Sun (Editor)
1993-01-01
Infrared observations from planned space facilities (e.g., ISO (Infrared Space Observatory), SIRTF (Space Infrared Telescope Facility)) will yield a large and uniform sample of high-quality data from both photometric and spectroscopic measurements. To maximize the scientific returns of these space missions, complementary theoretical studies must be undertaken to interpret these observations. A crucial step in such studies is the construction of phenomenological models in which we parameterize the observed radiation characteristics in terms of the physical source properties. In the last decade, models with increasing degree of physical realism (in terms of grain properties, physical processes, and source geometry) have been constructed for infrared sources. Here we review current capabilities available in the phenomenological modeling of infrared sources and discuss briefly directions for future research in this area.
NASA Astrophysics Data System (ADS)
Braban, Christine; Tang, Sim; Bealey, Bill; Roberts, Elin; Stephens, Amy; Galloway, Megan; Greenwood, Sarah; Sutton, Mark; Nemitz, Eiko; Leaver, David
2017-04-01
Ambient ammonia measurements have been undertaken both in the atmosphere to understand sources, concentrations at background and vulnerable ecosystems and for long term monitoring of concentrations. As a pollutant which is projected to increase concentration in the coming decades with significant policy challenges to implementing mitigation strategies it is useful to assess what has been measured, where and why. In this study a review of the literature, has shown that ammonia measurements are frequently not publically reported and in general not reposited in the open data centres, available for research. The specific sectors where measurements have been undertaken are: agricultural point source assessments, agricultural surface exchange measurements, sensitive ecosystem monitoring, landscape/regional studies and governmental long term monitoring. Less frequently ammonia is measured as part of an intensive atmospheric chemistry field campaign. Technology is developing which means a shift from chemical denuder methods to spectroscopic techniques may be possible, however chemical denuding techniques with off-line laboratory analysis will likely be an economical approach for some time to come. This paper reviews existing datasets from the different sectors of research and integrates them for a global picture to allow both a long term understanding and facilitate comparison with future measurements.
Remotely measuring populations during a crisis by overlaying two data sources
Bharti, Nita; Lu, Xin; Bengtsson, Linus; Wetter, Erik; Tatem, Andrew J.
2015-01-01
Background Societal instability and crises can cause rapid, large-scale movements. These movements are poorly understood and difficult to measure but strongly impact health. Data on these movements are important for planning response efforts. We retrospectively analyzed movement patterns surrounding a 2010 humanitarian crisis caused by internal political conflict in Côte d'Ivoire using two different methods. Methods We used two remote measures, nighttime lights satellite imagery and anonymized mobile phone call detail records, to assess average population sizes as well as dynamic population changes. These data sources detect movements across different spatial and temporal scales. Results The two data sources showed strong agreement in average measures of population sizes. Because the spatiotemporal resolution of the data sources differed, we were able to obtain measurements on long- and short-term dynamic elements of populations at different points throughout the crisis. Conclusions Using complementary, remote data sources to measure movement shows promise for future use in humanitarian crises. We conclude with challenges of remotely measuring movement and provide suggestions for future research and methodological developments. PMID:25733558
Standardization of terminology in field of ionizing radiations and their measurements
NASA Astrophysics Data System (ADS)
Yudin, M. F.; Karaveyev, F. M.
1984-03-01
A new standard terminology was introduced on 1 January 1982 by the Scientific-Technical Commission on All-Union State Standards to cover ionizing radiations and their measurements. It is based on earlier standards such as GOST 15484-74/81, 18445-70/73, 19849-74, 22490-77 as well as the latest recommendations by international committees. One hundred eighty-six terms and definitions in 14 paragraphs are contained. Fundamental concepts, sources and forms of ionizing radiations, characteristics and parameters of ionizing radiations, and methods of measuring their characteristics and parameters are covered. New terms have been added to existing ones. The equivalent English, French, and German terms are also given. The terms measurement of ionizing radiation and transfer of ionizing particles (equivalent of particle fluence of energy fluence) are still under discussion.
Coast of California Storm and Tidal Waves Study. Southern California Coastal Processes Data Summary,
1986-02-01
distribution of tracers injected on the beach. The suspended load was obtained from in situ measurements of the water column in the surf zone (Zampol and...wind waves. 3.2.2 Wave Climate There are relatively few in situ long-term measurements of the deep ocean (i.e. unaffected by the channel islands and...climate parameters and were not intended for that purpose. In the literature reviewed, the principal source of long-term in situ measurements is the
Inverse modelling of radionuclide release rates using gamma dose rate observations
NASA Astrophysics Data System (ADS)
Hamburger, Thomas; Stohl, Andreas; von Haustein, Christoph; Thummerer, Severin; Wallner, Christian
2014-05-01
Severe accidents in nuclear power plants such as the historical accident in Chernobyl 1986 or the more recent disaster in the Fukushima Dai-ichi nuclear power plant in 2011 have drastic impacts on the population and environment. The hazardous consequences reach out on a national and continental scale. Environmental measurements and methods to model the transport and dispersion of the released radionuclides serve as a platform to assess the regional impact of nuclear accidents - both, for research purposes and, more important, to determine the immediate threat to the population. However, the assessments of the regional radionuclide activity concentrations and the individual exposure to radiation dose underlie several uncertainties. For example, the accurate model representation of wet and dry deposition. One of the most significant uncertainty, however, results from the estimation of the source term. That is, the time dependent quantification of the released spectrum of radionuclides during the course of the nuclear accident. The quantification of the source terms of severe nuclear accidents may either remain uncertain (e.g. Chernobyl, Devell et al., 1995) or rely on rather rough estimates of released key radionuclides given by the operators. Precise measurements are mostly missing due to practical limitations during the accident. Inverse modelling can be used to realise a feasible estimation of the source term (Davoine and Bocquet, 2007). Existing point measurements of radionuclide activity concentrations are therefore combined with atmospheric transport models. The release rates of radionuclides at the accident site are then obtained by improving the agreement between the modelled and observed concentrations (Stohl et al., 2012). The accuracy of the method and hence of the resulting source term depends amongst others on the availability, reliability and the resolution in time and space of the observations. Radionuclide activity concentrations are observed on a relatively sparse grid and the temporal resolution of available data may be low within the order of hours or a day. Gamma dose rates on the other hand are observed routinely on a much denser grid and higher temporal resolution. Gamma dose rate measurements contain no explicit information on the observed spectrum of radionuclides and have to be interpreted carefully. Nevertheless, they provide valuable information for the inverse evaluation of the source term due to their availability (Saunier et al., 2013). We present a new inversion approach combining an atmospheric dispersion model and observations of radionuclide activity concentrations and gamma dose rates to obtain the source term of radionuclides. We use the Lagrangian particle dispersion model FLEXPART (Stohl et al., 1998; Stohl et al., 2005) to model the atmospheric transport of the released radionuclides. The gamma dose rates are calculated from the modelled activity concentrations. The inversion method uses a Bayesian formulation considering uncertainties for the a priori source term and the observations (Eckhardt et al., 2008). The a priori information on the source term is a first guess. The gamma dose rate observations will be used with inverse modelling to improve this first guess and to retrieve a reliable source term. The details of this method will be presented at the conference. This work is funded by the Bundesamt für Strahlenschutz BfS, Forschungsvorhaben 3612S60026. References Davoine, X. and Bocquet, M., Atmos. Chem. Phys., 7, 1549-1564, 2007. Devell, L., et al., OCDE/GD(96)12, 1995. Eckhardt, S., et al., Atmos. Chem. Phys., 8, 3881-3897, 2008. Saunier, O., et al., Atmos. Chem. Phys., 13, 11403-11421, 2013. Stohl, A., et al., Atmos. Environ., 32, 4245-4264, 1998. Stohl, A., et al., Atmos. Chem. Phys., 5, 2461-2474, 2005. Stohl, A., et al., Atmos. Chem. Phys., 12, 2313-2343, 2012.
Horowitz, A.J.; Elrick, K.A.; Smith, J.J.
2005-01-01
In cooperation with the City of Atlanta, Georgia, the US Geological Survey has designed and implemented a water-quantity and quality monitoring network that measures a variety of biological and chemical constituents in water and suspended sediment. The network consists of 20 long-term monitoring sites and is intended to assess water-quality trends in response to planned infrastructural improvements. Initial results from the network indicate that nonpoint-source contributions may be more significant than point-source contributions for selected sediment associated trace elements and nutrients. There also are indications of short-term discontinuous point-source contributions of these same constituents during baseflow.
McDonald, Brian C; Goldstein, Allen H; Harley, Robert A
2015-04-21
A fuel-based approach is used to assess long-term trends (1970-2010) in mobile source emissions of black carbon (BC) and organic aerosol (OA, including both primary emissions and secondary formation). The main focus of this analysis is the Los Angeles Basin, where a long record of measurements is available to infer trends in ambient concentrations of BC and organic carbon (OC), with OC used here as a proxy for OA. Mobile source emissions and ambient concentrations have decreased similarly, reflecting the importance of on- and off-road engines as sources of BC and OA in urban areas. In 1970, the on-road sector accounted for ∼90% of total mobile source emissions of BC and OA (primary + secondary). Over time, as on-road engine emissions have been controlled, the relative importance of off-road sources has grown. By 2010, off-road engines were estimated to account for 37 ± 20% and 45 ± 16% of total mobile source contributions to BC and OA, respectively, in the Los Angeles area. This study highlights both the success of efforts to control on-road emission sources, and the importance of considering off-road engine and other VOC source contributions when assessing long-term emission and ambient air quality trends.
NASA Astrophysics Data System (ADS)
Phillips-Smith, Catherine; Jeong, Cheol-Heon; Healy, Robert M.; Dabek-Zlotorzynska, Ewa; Celo, Valbona; Brook, Jeffrey R.; Evans, Greg
2017-08-01
The province of Alberta, Canada, is home to three oil sands regions which, combined, contain the third largest deposit of oil in the world. Of these, the Athabasca oil sands region is the largest. As part of Environment and Climate Change Canada's program in support of the Joint Canada-Alberta Implementation Plan for Oil Sands Monitoring program, concentrations of trace elements in PM2. 5 (particulate matter smaller than 2.5 µm in diameter) were measured through two campaigns that involved different methodologies: a long-term filter campaign and a short-term intensive campaign. In the long-term campaign, 24 h filter samples were collected once every 6 days over a 2-year period (December 2010-November 2012) at three air monitoring stations in the regional municipality of Wood Buffalo. For the intensive campaign (August 2013), hourly measurements were made with an online instrument at one air monitoring station; daily filter samples were also collected. The hourly and 24 h filter data were analyzed individually using positive matrix factorization. Seven emission sources of PM2. 5 trace elements were thereby identified: two types of upgrader emissions, soil, haul road dust, biomass burning, and two sources of mixed origin. The upgrader emissions, soil, and haul road dust sources were identified through both the methodologies and both methodologies identified a mixed source, but these exhibited more differences than similarities. The second upgrader emissions and biomass burning sources were only resolved by the hourly and filter methodologies, respectively. The similarity of the receptor modeling results from the two methodologies provided reassurance as to the identity of the sources. Overall, much of the PM2. 5-related trace elements were found to be anthropogenic, or at least to be aerosolized through anthropogenic activities. These emissions may in part explain the previously reported higher levels of trace elements in snow, water, and biota samples collected near the oil sands operations.
Analysis of jet-airfoil interaction noise sources by using a microphone array technique
NASA Astrophysics Data System (ADS)
Fleury, Vincent; Davy, Renaud
2016-03-01
The paper is concerned with the characterization of jet noise sources and jet-airfoil interaction sources by using microphone array data. The measurements were carried-out in the anechoic open test section wind tunnel of Onera, Cepra19. The microphone array technique relies on the convected, Lighthill's and Ffowcs-Williams and Hawkings' acoustic analogy equation. The cross-spectrum of the source term of the analogy equation is sought. It is defined as the optimal solution to a minimal error equation using the measured microphone cross-spectra as reference. This inverse problem is ill-posed yet. A penalty term based on a localization operator is therefore added to improve the recovery of jet noise sources. The analysis of isolated jet noise data in subsonic regime shows the contribution of the conventional mixing noise source in the low frequency range, as expected, and of uniformly distributed, uncorrelated noise sources in the jet flow at higher frequencies. In underexpanded supersonic regime, a shock-associated noise source is clearly identified, too. An additional source is detected in the vicinity of the nozzle exit both in supersonic and subsonic regimes. In the presence of the airfoil, the distribution of the noise sources is deeply modified. In particular, a strong noise source is localized on the flap. For high Strouhal numbers, higher than about 2 (based on the jet mixing velocity and diameter), a significant contribution from the shear-layer near the flap is observed, too. Indications of acoustic reflections on the airfoil are also discerned.
ERROR IN ANNUAL AVERAGE DUE TO USE OF LESS THAN EVERYDAY MEASUREMENTS
Long term averages of the concentration of PM mass and components are of interest for determining compliance with annual averages, for developing exposure surrogated for cross-sectional epidemiologic studies of the long-term of PM, and for determination of aerosol sources by chem...
A Cross-Lingual Similarity Measure for Detecting Biomedical Term Translations
Bollegala, Danushka; Kontonatsios, Georgios; Ananiadou, Sophia
2015-01-01
Bilingual dictionaries for technical terms such as biomedical terms are an important resource for machine translation systems as well as for humans who would like to understand a concept described in a foreign language. Often a biomedical term is first proposed in English and later it is manually translated to other languages. Despite the fact that there are large monolingual lexicons of biomedical terms, only a fraction of those term lexicons are translated to other languages. Manually compiling large-scale bilingual dictionaries for technical domains is a challenging task because it is difficult to find a sufficiently large number of bilingual experts. We propose a cross-lingual similarity measure for detecting most similar translation candidates for a biomedical term specified in one language (source) from another language (target). Specifically, a biomedical term in a language is represented using two types of features: (a) intrinsic features that consist of character n-grams extracted from the term under consideration, and (b) extrinsic features that consist of unigrams and bigrams extracted from the contextual windows surrounding the term under consideration. We propose a cross-lingual similarity measure using each of those feature types. First, to reduce the dimensionality of the feature space in each language, we propose prototype vector projection (PVP)—a non-negative lower-dimensional vector projection method. Second, we propose a method to learn a mapping between the feature spaces in the source and target language using partial least squares regression (PLSR). The proposed method requires only a small number of training instances to learn a cross-lingual similarity measure. The proposed PVP method outperforms popular dimensionality reduction methods such as the singular value decomposition (SVD) and non-negative matrix factorization (NMF) in a nearest neighbor prediction task. Moreover, our experimental results covering several language pairs such as English–French, English–Spanish, English–Greek, and English–Japanese show that the proposed method outperforms several other feature projection methods in biomedical term translation prediction tasks. PMID:26030738
NASA Astrophysics Data System (ADS)
Diapouli, E.; Manousakas, M.; Vratolis, S.; Vasilatou, V.; Maggos, Th; Saraga, D.; Grigoratos, Th; Argyropoulos, G.; Voutsa, D.; Samara, C.; Eleftheriadis, K.
2017-09-01
Metropolitan Urban areas in Greece have been known to suffer from poor air quality, due to variety of emission sources, topography and climatic conditions favouring the accumulation of pollution. While a number of control measures have been implemented since the 1990s, resulting in reductions of atmospheric pollution and changes in emission source contributions, the financial crisis which started in 2009 has significantly altered this picture. The present study is the first effort to assess the contribution of emission sources to PM10 and PM2.5 concentration levels and their long-term variability (over 5-10 years), in the two largest metropolitan urban areas in Greece (Athens and Thessaloniki). Intensive measurement campaigns were conducted during 2011-2012 at suburban, urban background and urban traffic sites in these two cities. In addition, available datasets from previous measurements in Athens and Thessaloniki were used in order to assess the long-term variability of concentrations and sources. Chemical composition analysis of the 2011-2012 samples showed that carbonaceous matter was the most abundant component for both PM size fractions. Significant increase of carbonaceous particle concentrations and of OC/EC ratio during the cold period, especially in the residential urban background sites, pointed towards domestic heating and more particularly wood (biomass) burning as a significant source. PMF analysis further supported this finding. Biomass burning was the largest contributing source at the two urban background sites (with mean contributions for the two size fractions in the range of 24-46%). Secondary aerosol formation (sulphate, nitrate & organics) was also a major contributing source for both size fractions at the suburban and urban background sites. At the urban traffic site, vehicular traffic (exhaust and non-exhaust emissions) was the source with the highest contributions, accounting for 44% of PM10 and 37% of PM2.5, respectively. The long-term variability of emission sources in the two cities (over 5-10 years), assessed through a harmonized application of the PMF technique on recent and past year data, clearly demonstrates the effective reduction in emissions during the last decade due to control measures and technological development; however, it also reflects the effects of the financial crisis in Greece during these years, which has led to decreased economic activities and the adoption of more polluting practices by the local population in an effort to reduce living costs.
NASA Technical Reports Server (NTRS)
Greenwood, Eric, II; Schmitz, Fredric H.
2010-01-01
A new physics-based parameter identification method for rotor harmonic noise sources is developed using an acoustic inverse simulation technique. This new method allows for the identification of individual rotor harmonic noise sources and allows them to be characterized in terms of their individual non-dimensional governing parameters. This new method is applied to both wind tunnel measurements and ground noise measurements of two-bladed rotors. The method is shown to match the parametric trends of main rotor Blade-Vortex Interaction (BVI) noise, allowing accurate estimates of BVI noise to be made for operating conditions based on a small number of measurements taken at different operating conditions.
Fundamental Rotorcraft Acoustic Modeling From Experiments (FRAME)
NASA Technical Reports Server (NTRS)
Greenwood, Eric
2011-01-01
A new methodology is developed for the construction of helicopter source noise models for use in mission planning tools from experimental measurements of helicopter external noise radiation. The models are constructed by employing a parameter identification method to an assumed analytical model of the rotor harmonic noise sources. This new method allows for the identification of individual rotor harmonic noise sources and allows them to be characterized in terms of their individual non-dimensional governing parameters. The method is applied to both wind tunnel measurements and ground noise measurements of two-bladed rotors. The method is shown to match the parametric trends of main rotor harmonic noise, allowing accurate estimates of the dominant rotorcraft noise sources to be made for operating conditions based on a small number of measurements taken at different operating conditions. The ability of this method to estimate changes in noise radiation due to changes in ambient conditions is also demonstrated.
Disentangling the effects of CO2 and short-lived climate forcer mitigation.
Rogelj, Joeri; Schaeffer, Michiel; Meinshausen, Malte; Shindell, Drew T; Hare, William; Klimont, Zbigniew; Velders, Guus J M; Amann, Markus; Schellnhuber, Hans Joachim
2014-11-18
Anthropogenic global warming is driven by emissions of a wide variety of radiative forcers ranging from very short-lived climate forcers (SLCFs), like black carbon, to very long-lived, like CO2. These species are often released from common sources and are therefore intricately linked. However, for reasons of simplification, this CO2-SLCF linkage was often disregarded in long-term projections of earlier studies. Here we explicitly account for CO2-SLCF linkages and show that the short- and long-term climate effects of many SLCF measures consistently become smaller in scenarios that keep warming to below 2 °C relative to preindustrial levels. Although long-term mitigation of methane and hydrofluorocarbons are integral parts of 2 °C scenarios, early action on these species mainly influences near-term temperatures and brings small benefits for limiting maximum warming relative to comparable reductions taking place later. Furthermore, we find that maximum 21st-century warming in 2 °C-consistent scenarios is largely unaffected by additional black-carbon-related measures because key emission sources are already phased-out through CO2 mitigation. Our study demonstrates the importance of coherently considering CO2-SLCF coevolutions. Failing to do so leads to strongly and consistently overestimating the effect of SLCF measures in climate stabilization scenarios. Our results reinforce that SLCF measures are to be considered complementary rather than a substitute for early and stringent CO2 mitigation. Near-term SLCF measures do not allow for more time for CO2 mitigation. We disentangle and resolve the distinct benefits across different species and therewith facilitate an integrated strategy for mitigating both short and long-term climate change.
Disentangling the effects of CO2 and short-lived climate forcer mitigation
Rogelj, Joeri; Schaeffer, Michiel; Meinshausen, Malte; Shindell, Drew T.; Hare, William; Klimont, Zbigniew; Amann, Markus; Schellnhuber, Hans Joachim
2014-01-01
Anthropogenic global warming is driven by emissions of a wide variety of radiative forcers ranging from very short-lived climate forcers (SLCFs), like black carbon, to very long-lived, like CO2. These species are often released from common sources and are therefore intricately linked. However, for reasons of simplification, this CO2–SLCF linkage was often disregarded in long-term projections of earlier studies. Here we explicitly account for CO2–SLCF linkages and show that the short- and long-term climate effects of many SLCF measures consistently become smaller in scenarios that keep warming to below 2 °C relative to preindustrial levels. Although long-term mitigation of methane and hydrofluorocarbons are integral parts of 2 °C scenarios, early action on these species mainly influences near-term temperatures and brings small benefits for limiting maximum warming relative to comparable reductions taking place later. Furthermore, we find that maximum 21st-century warming in 2 °C-consistent scenarios is largely unaffected by additional black-carbon-related measures because key emission sources are already phased-out through CO2 mitigation. Our study demonstrates the importance of coherently considering CO2–SLCF coevolutions. Failing to do so leads to strongly and consistently overestimating the effect of SLCF measures in climate stabilization scenarios. Our results reinforce that SLCF measures are to be considered complementary rather than a substitute for early and stringent CO2 mitigation. Near-term SLCF measures do not allow for more time for CO2 mitigation. We disentangle and resolve the distinct benefits across different species and therewith facilitate an integrated strategy for mitigating both short and long-term climate change. PMID:25368182
The Funding of Long-Term Care in Canada: What Do We Know, What Should We Know?
Grignon, Michel; Spencer, Byron G
2018-06-01
ABSTRACTLong-term care is a growing component of health care spending but how much is spent or who bears the cost is uncertain, and the measures vary depending on the source used. We drew on regularly published series and ad hoc publications to compile preferred estimates of the share of long-term care spending in total health care spending, the private share of long-term care spending, and the share of residential care within long-term care. For each series, we compared estimates obtainable from published sources (CIHI [Canadian Institute for Health Information] and OECD [Organization for Economic Cooperation and Development]) with our preferred estimates. We conclude that using published series without adjustment would lead to spurious conclusions on the level and evolution of spending on long-term care in Canada as well as on the distribution of costs between private and public funders and between residential and home care.
Using high frequency CDOM hyperspectral absorption to fingerprint river water sources
NASA Astrophysics Data System (ADS)
Beckler, J. S.; Kirkpatrick, G. J.; Dixon, L. K.; Milbrandt, E. C.
2016-12-01
Quantifying riverine carbon transfer from land to sea is complicated by variability in dissolved organic carbon (DOC), closely-related dissolved organic matter (DOM) and chromophoric dissolved organic matter (CDOM) concentrations, as well as in the composition of the freshwater end members of multiple drainage basins and seasons. Discrete measurements in estuaries have difficulty resolving convoluted upstream watershed dynamics. Optical measurements, however, can provide more continuous data regarding the molecular composition and concentration of the CDOM as it relates to river flow, tidal mixing, and salinity and may be used to fingerprint source waters. For the first time, long-term, hyperspectral CDOM measurements were obtained on filtered Caloosahatchee River estuarine waters using an in situ, long-pathlength spectrophotometric instrument, the Optical Phytoplankton Discriminator (OPD). Through a collaborative monitoring effort among partners within the Gulf of Mexico Coastal Ocean Observing System (GCOOS), ancillary measurements of fluorescent DOM (FDOM) and water quality parameters were also obtained from co-located instrumentation at high frequency. Optical properties demonstrated both short-term (hourly) tidal variations and long-term (daily - weekly) variations corresponding to changes in riverine flow and salinity. The optical properties of the river waters are demonstrated to be a dilution-adjusted linear combination of the optical properties of the source waters comprising the overall composition (e.g. Lake Okeechobee, watershed drainage basins, Gulf of Mexico). Overall, these techniques are promising as a tool to more accurately constrain the carbon flux to the ocean and to predict the optical quality of coastal waters.
Next Generation Air Measurements for Fugitive, Area Source, and Fence Line Applications
Next generation air measurements (NGAM) is an EPA term for the advancing field of air pollutant sensor technologies, data integration concepts, and geospatial modeling strategies. Ranging from personal sensors to satellite remote sensing, NGAM systems may provide revolutionary n...
Remotely measuring populations during a crisis by overlaying two data sources.
Bharti, Nita; Lu, Xin; Bengtsson, Linus; Wetter, Erik; Tatem, Andrew J
2015-03-01
Societal instability and crises can cause rapid, large-scale movements. These movements are poorly understood and difficult to measure but strongly impact health. Data on these movements are important for planning response efforts. We retrospectively analyzed movement patterns surrounding a 2010 humanitarian crisis caused by internal political conflict in Côte d'Ivoire using two different methods. We used two remote measures, nighttime lights satellite imagery and anonymized mobile phone call detail records, to assess average population sizes as well as dynamic population changes. These data sources detect movements across different spatial and temporal scales. The two data sources showed strong agreement in average measures of population sizes. Because the spatiotemporal resolution of the data sources differed, we were able to obtain measurements on long- and short-term dynamic elements of populations at different points throughout the crisis. Using complementary, remote data sources to measure movement shows promise for future use in humanitarian crises. We conclude with challenges of remotely measuring movement and provide suggestions for future research and methodological developments. © The Author 2015. Published by Oxford University Press on behalf of Royal Society of Tropical Medicine and Hygiene.
NASA Astrophysics Data System (ADS)
Yoshida, Satoshi
Applications of inductively coupled plasma mass spectrometry (ICP-MS) to the determination of long-lived radionuclides in environmental samples were summarized. In order to predict the long-term behavior of the radionuclides, related stable elements were also determined. Compared with radioactivity measurements, the ICP-MS method has advantages in terms of its simple analytical procedures, prompt measurement time, and capability of determining the isotope ratio such as240Pu/239Pu, which can not be separated by radiation. Concentration of U and Th in Japanese surface soils were determined in order to determine the background level of the natural radionuclides. The 235U/238U ratio was successfully used to detect the release of enriched U from reconversion facilities to the environment and to understand the source term. The 240Pu/239Pu ratios in environmental samples varied widely depending on the Pu sources. Applications of ICP-MS to the measurement of I and Tc isotopes were also described. The ratio between radiocesium and stable Cs is useful for judging the equilibrium of deposited radiocesium in a forest ecosystem.
A new sensor system for mobile and aerial emission sampling was developed for open area pollutant sources, such as prescribed forest burns. The sensor system, termed “Kolibri”, consists of multiple low-cost air quality sensors measuring CO2, CO, samplers for particulate matter wi...
A new sensor system for mobile and aerial emission sampling was developed for open area pollutant sources, such as prescribed forest burns. The sensor system, termed “Kolibri”, consists of multiple low-cost air quality sensors measuring CO2, CO, samplers for particulate matter wi...
A new sensor system for mobile and aerial emission sampling was developed for open area sources, such as open burning. The sensor system, termed “Kolibri”, consists of multiple low-cost air quality sensors measuring CO2, CO, and black carbon, samplers for particulate matter with ...
We will estimate health risks associated with short- and long-term exposure to individual air pollutants, sources and air pollution mixtures within the Framingham Offspring and Third Generation populations. We will address which individual and area-level factors, measuring vul...
Biotic Nitrogen Enrichment Regulates Calcium Sources to Forests
NASA Astrophysics Data System (ADS)
Pett-Ridge, J. C.; Perakis, S. S.; Hynicka, J. D.
2015-12-01
Calcium is an essential nutrient in forest ecosystems that is susceptible to leaching loss and depletion. Calcium depletion can affect plant and animal productivity, soil acid buffering capacity, and fluxes of carbon and water. Excess nitrogen supply and associated soil acidification are often implicated in short-term calcium loss from soils, but the long-term role of nitrogen enrichment on calcium sources and resupply is unknown. Here we use strontium isotopes (87Sr/86Sr) as a proxy for calcium to investigate how soil nitrogen enrichment from biological nitrogen fixation interacts with bedrock calcium to regulate both short-term available supplies and the long-term sources of calcium in montane conifer forests. Our study examines 22 sites in western Oregon, spanning a 20-fold range of bedrock calcium on sedimentary and basaltic lithologies. In contrast to previous studies emphasizing abiotic control of weathering as a determinant of long-term ecosystem calcium dynamics and sources (via bedrock fertility, climate, or topographic/tectonic controls) we find instead that that biotic nitrogen enrichment of soil can strongly regulate calcium sources and supplies in forest ecosystems. For forests on calcium-rich basaltic bedrock, increasing nitrogen enrichment causes calcium sources to shift from rock-weathering to atmospheric dominance, with minimal influence from other major soil forming factors, despite regionally high rates of tectonic uplift and erosion that can rejuvenate weathering supply of soil minerals. For forests on calcium-poor sedimentary bedrock, we find that atmospheric inputs dominate regardless of degree of nitrogen enrichment. Short-term measures of soil and ecosystem calcium fertility are decoupled from calcium source sustainability, with fundamental implications for understanding nitrogen impacts, both in natural ecosystems and in the context of global change. Our finding that long-term nitrogen enrichment increases forest reliance on atmospheric calcium helps explain reports of greater ecological calcium limitation in an increasingly nitrogen-rich world.
Plutonium isotopes and 241Am in the atmosphere of Lithuania: A comparison of different source terms
NASA Astrophysics Data System (ADS)
Lujanienė, G.; Valiulis, D.; Byčenkienė, S.; Šakalys, J.; Povinec, P. P.
2012-12-01
137Cs, 241Am and Pu isotopes collected in aerosol samples during 1994-2011 were analyzed with special emphasis on better understanding of Pu and Am behavior in the atmosphere. The results from long-term measurements of 240Pu/239Pu atom ratios showed a bimodal frequency distribution with median values of 0.195 and 0.253, indicating two main sources contributing to the Pu activities at the Vilnius sampling station. The low Pu atom ratio of 0.141 could be attributed to the weapon-grade plutonium derived from the nuclear weapon test sites. The frequency of air masses arriving from the North-West and North-East correlated with the Pu atom ratio indicating the input from the sources located in these regions (the Novaya Zemlya test site, Siberian nuclear plants), while no correlation with the Chernobyl region was observed. Measurements carried out during the Fukushima accident showed a negligible impact of this source with Pu activities by four orders of magnitude lower as compared to the Chernobyl accident. The activity concentration of actinides measured in the integrated sample collected in March-April, 2011 showed a small contribution of Pu with unusual activity and atom ratios indicating the presence of the spent fuel of different origin than that of the Chernobyl accident.
Effect of source location and listener location on ILD cues in a reverberant room
NASA Astrophysics Data System (ADS)
Ihlefeld, Antje; Shinn-Cunningham, Barbara G.
2004-05-01
Short-term interaural level differences (ILDs) were analyzed for simulations of the signals that would reach a listener in a reverberant room. White noise was convolved with manikin head-related impulse responses measured in a classroom to simulate different locations of the source relative to the manikin and different manikin positions in the room. The ILDs of the signals were computed within each third-octave band over a relatively short time window to investigate how reliably ILD cues encode source laterality. Overall, the mean of the ILD magnitude increases with lateral angle and decreases with distance, as expected. Increasing reverberation decreases the mean ILD magnitude and increases the variance of the short-term ILD, so that the spatial information carried by ILD cues is degraded by reverberation. These results suggest that the mean ILD is not a reliable cue for determining source laterality in a reverberant room. However, by taking into account both the mean and variance, the distribution of high-frequency short-term ILDs provides some spatial information. This analysis suggests that, in order to use ILDs to judge source direction in reverberant space, listeners must accumulate information about how the short-term ILD varies over time. [Work supported by NIDCD and AFOSR.
A quality assurance program for clinical PDT
NASA Astrophysics Data System (ADS)
Dimofte, Andreea; Finlay, Jarod; Ong, Yi Hong; Zhu, Timothy C.
2018-02-01
Successful outcome of Photodynamic therapy (PDT) depends on accurate delivery of prescribed light dose. A quality assurance program is necessary to ensure that light dosimetry is correctly measured. We have instituted a QA program that include examination of long term calibration uncertainty of isotropic detectors for light fluence rate, power meter head intercomparison for laser power, stability of the light-emitting diode (LED) light source integrating sphere as a light fluence standard, laser output and calibration of in-vivo reflective fluorescence and absorption spectrometers. We examined the long term calibration uncertainty of isotropic detector sensitivity, defined as fluence rate per voltage. We calibrate the detector using the known calibrated light fluence rate of the LED light source built into an internally baffled 4" integrating sphere. LED light sources were examined using a 1mm diameter isotropic detector calibrated in a collimated beam. Wavelengths varying from 632nm to 690nm were used. The internal LED method gives an overall calibration accuracy of +/- 4%. Intercomparison among power meters was performed to determine the consistency of laser power and light fluence rate measured among different power meters. Power and fluence readings were measured and compared among detectors. A comparison of power and fluence reading among several power heads shows long term consistency for power and light fluence rate calibration to within 3% regardless of wavelength. The standard LED light source is used to calibrate the transmission difference between different channels for the diffuse reflective absorption and fluorescence contact probe as well as isotropic detectors used in PDT dose dosimeter.
NASA Astrophysics Data System (ADS)
Xu, L.; Suresh, S.; Guo, H.; Weber, R. J.; Ng, N. L.
2015-04-01
We deployed a High-Resolution Time-of-Flight Aerosol Mass Spectrometer (HR-ToF-AMS) and an Aerosol Chemical Speciation Monitor (ACSM) to characterize the chemical composition of submicron non-refractory particles (NR-PM1) in the southeastern US. Measurements were performed in both rural and urban sites in the greater Atlanta area, GA and Centreville, AL for approximately one year, as part of Southeastern Center of Air Pollution and Epidemiology study (SCAPE) and Southern Oxidant and Aerosol Study (SOAS). Organic aerosol (OA) accounts for more than half of NR1 mass concentration regardless of sampling sites and seasons. Positive matrix factorization (PMF) analysis of HR-ToF-AMS measurements identified various OA sources, depending on location and season. Hydrocarbon-like OA (HOA) and cooking OA (COA) have important but not dominant contributions to total OA in urban sites. Biomass burning OA (BBOA) concentration shows a distinct seasonal variation with a larger enhancement in winter than summer. We find a good correlation between BBOA and brown carbon, indicating biomass burning is an important source for brown carbon, although an additional, unidentified brown carbon source is likely present at the rural Yorkville site. Isoprene-derived OA (Isoprene-OA) is only deconvolved in warmer months and contributes 18-36% of total OA. The presence of Isoprene-OA factor in urban sites is more likely from local production in the presence of NOx than transport from rural sites. More-oxidized and less-oxidized oxygenated organic aerosol (MO-OOA and LO-OOA, respectively) are dominant fractions (47-79%) of OA in all sites. MO-OOA correlates well with ozone in summer, but not in winter, indicating MO-OOA sources may vary with seasons. LO-OOA, which reaches a daily maximum at night, correlates better with estimated nitrate functionality from organic nitrates than total nitrates. Based on the HR-ToF-AMS measurements, we estimate that the nitrate functionality from organic nitrates contributes 63-100% of total measured nitrates in summer. Further, the contribution of organic nitrates to total OA is estimated to be 5-12% in summer, suggesting that organic nitrates are important components in the ambient aerosol in the southeastern US. The spatial distribution of OA is investigated by comparing simultaneous HR-ToF-AMS measurements with ACSM measurements at two different sampling sites. OA is found to be spatially homogeneous in summer, possibly due to stagnant air mass and a dominant amount of regional SOA in the southeastern US. The homogeneity is less in winter, which is likely due to spatial variation of primary emissions. We observed that the seasonality of OA concentration shows a clear urban/rural contrast. While OA exhibits weak seasonal variation in the urban sites, its concentration is higher in summer than winter for rural sites. This observation from our year-long measurements is consistent with 14 years of organic carbon (OC) data from the SouthEastern Aerosol Research and Characterization (SEARCH) network. The comparison between short-term measurements with advanced instruments and long-term measurements of basic air quality indicators not only tests the robustness of the short-term measurements but also provides insights in interpreting long-term measurements. We find that OA factors resolved from PMF analysis on HR-ToF-AMS measurements have distinctly different diurnal variations. The compensation of OA factors with different diurnal trends is one possible reason for the repeatedly observed, relatively flat OA diurnal profile in the southeastern US. In addition, analysis of long-term measurements shows that the correlation between OC and sulfate is substantially higher in summer than winter. This seasonality could be partly due to the effects of sulfate on isoprene SOA formation as revealed by the short-term, intensive measurements.
NASA Astrophysics Data System (ADS)
Xu, L.; Suresh, S.; Guo, H.; Weber, R. J.; Ng, N. L.
2015-07-01
We deployed a High-Resolution Time-of-Flight Aerosol Mass Spectrometer (HR-ToF-AMS) and an Aerosol Chemical Speciation Monitor (ACSM) to characterize the chemical composition of submicron non-refractory particulate matter (NR-PM1) in the southeastern USA. Measurements were performed in both rural and urban sites in the greater Atlanta area, Georgia (GA), and Centreville, Alabama (AL), for approximately 1 year as part of Southeastern Center for Air Pollution and Epidemiology study (SCAPE) and Southern Oxidant and Aerosol Study (SOAS). Organic aerosol (OA) accounts for more than half of NR-PM1 mass concentration regardless of sampling sites and seasons. Positive matrix factorization (PMF) analysis of HR-ToF-AMS measurements identified various OA sources, depending on location and season. Hydrocarbon-like OA (HOA) and cooking OA (COA) have important, but not dominant, contributions to total OA in urban sites (i.e., 21-38 % of total OA depending on site and season). Biomass burning OA (BBOA) concentration shows a distinct seasonal variation with a larger enhancement in winter than summer. We find a good correlation between BBOA and brown carbon, indicating biomass burning is an important source for brown carbon, although an additional, unidentified brown carbon source is likely present at the rural Yorkville site. Isoprene-derived OA factor (isoprene-OA) is only deconvolved in warmer months and contributes 18-36 % of total OA. The presence of isoprene-OA factor in urban sites is more likely from local production in the presence of NOx than transport from rural sites. More-oxidized and less-oxidized oxygenated organic aerosol (MO-OOA and LO-OOA, respectively) are dominant fractions (47-79 %) of OA in all sites. MO-OOA correlates well with ozone in summer but not in winter, indicating MO-OOA sources may vary with seasons. LO-OOA, which reaches a daily maximum at night, correlates better with estimated nitrate functionality from organic nitrates than total nitrates. Based on the HR-ToF-AMS measurements, we estimate that the nitrate functionality from organic nitrates contributes 63-100 % to the total measured nitrates in summer. Furthermore, the contribution of organic nitrates to total OA is estimated to be 5-12 % in summer, suggesting that organic nitrates are important components in the ambient aerosol in the southeastern USA. The spatial distribution of OA is investigated by comparing simultaneous HR-ToF-AMS measurements with ACSM measurements at two different sampling sites. OA is found to be spatially homogeneous in summer due possibly to stagnant air mass and a dominant amount of regional secondary organic aerosol (SOA) in the southeastern USA. The homogeneity is less in winter, which is likely due to spatial variation of primary emissions. We observe that the seasonality of OA concentration shows a clear urban/rural contrast. While OA exhibits weak seasonal variation in the urban sites, its concentration is higher in summer than winter for rural sites. This observation from our year-long measurements is consistent with 14 years of organic carbon (OC) data from the SouthEastern Aerosol Research and Characterization (SEARCH) network. The comparison between short-term measurements with advanced instruments and long-term measurements of basic air quality indicators not only tests the robustness of the short-term measurements but also provides insights in interpreting long-term measurements. We find that OA factors resolved from PMF analysis on HR-ToF-AMS measurements have distinctly different diurnal variations. The compensation of OA factors with different diurnal trends is one possible reason for the repeatedly observed, relatively flat OA diurnal profile in the southeastern USA. In addition, analysis of long-term measurements shows that the correlation between OC and sulfate is substantially stronger in summer than winter. This seasonality could be partly due to the effects of sulfate on isoprene SOA formation as revealed by the short-term intensive measurements.
Bremsstrahlung Dose Yield for High-Intensity Short-Pulse Laser–Solid Experiments
Liang, Taiee; Bauer, Johannes M.; Liu, James C.; ...
2016-12-01
A bremsstrahlung source term has been developed by the Radiation Protection (RP) group at SLAC National Accelerator Laboratory for high-intensity short-pulse laser–solid experiments between 10 17 and 10 22 W cm –2. This source term couples the particle-in-cell plasma code EPOCH and the radiation transport code FLUKA to estimate the bremsstrahlung dose yield from laser–solid interactions. EPOCH characterizes the energy distribution, angular distribution, and laser-to-electron conversion efficiency of the hot electrons from laser–solid interactions, and FLUKA utilizes this hot electron source term to calculate a bremsstrahlung dose yield (mSv per J of laser energy on target). The goal of thismore » paper is to provide RP guidelines and hazard analysis for high-intensity laser facilities. In conclusion, a comparison of the calculated bremsstrahlung dose yields to radiation measurement data is also made.« less
Bremsstrahlung Dose Yield for High-Intensity Short-Pulse Laser–Solid Experiments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liang, Taiee; Bauer, Johannes M.; Liu, James C.
A bremsstrahlung source term has been developed by the Radiation Protection (RP) group at SLAC National Accelerator Laboratory for high-intensity short-pulse laser–solid experiments between 10 17 and 10 22 W cm –2. This source term couples the particle-in-cell plasma code EPOCH and the radiation transport code FLUKA to estimate the bremsstrahlung dose yield from laser–solid interactions. EPOCH characterizes the energy distribution, angular distribution, and laser-to-electron conversion efficiency of the hot electrons from laser–solid interactions, and FLUKA utilizes this hot electron source term to calculate a bremsstrahlung dose yield (mSv per J of laser energy on target). The goal of thismore » paper is to provide RP guidelines and hazard analysis for high-intensity laser facilities. In conclusion, a comparison of the calculated bremsstrahlung dose yields to radiation measurement data is also made.« less
Code of Federal Regulations, 2010 CFR
2010-10-01
... other sources, (iii) The measures that you are taking and have taken to alleviate the situation, (iv) Financial projections, (v) Proposed term of the repayment, (vi) Current and projected market conditions...
Spectra of cosmic X-ray sources
NASA Technical Reports Server (NTRS)
Holt, S. S.; Mccray, R.
1982-01-01
X-ray measurements provide the most direct probes of astrophysical environments with temperatures exceeding one million K. Progress in experimental research utilizing dispersive techniques (e.g., Bragg and grating spectroscopy) is considerably slower than that in areas utilizing photometric techniques, because of the relative inefficiency of the former for the weak X-ray signals from celestial sources. As a result, the term "spectroscopy" as applied to X-ray astronomy has traditionally satisfied a much less restrictive definition (in terms of resolving power) than it has in other wavebands. Until quite recently, resolving powers of order unity were perfectly respectable, and still provide (in most cases) the most useful spectroscopic data. In the broadest sense, X-ray photometric measurements are spectroscopic, insofar as they represent samples of the overall electromagnetic continua of celestial objects.
Tanik, A
2000-01-01
The six main drinking water reservoirs of Istanbul are under the threat of pollution due to rapid population increase, unplanned urbanisation and insufficient infrastructure. In contrast to the present land use profile, the environmental evaluation of the catchment areas reveals that point sources of pollutants, especially of domestic origin, dominate over those from diffuse sources. The water quality studies also support these findings, emphasising that if no substantial precautions are taken, there will be no possibility of obtaining drinking water from them. In this paper, under the light of the present status of the reservoirs, possible and probable short- and long-term protective measures are outlined for reducing the impact of point sources. Immediate precautions mostly depend on reducing the pollution arising from the existing settlements. Long-term measures mainly emphasise the preparation of new land use plans taking into consideration the protection of unoccupied lands. Recommendations on protection and control of the reservoirs are stated.
A suggested glossary of terms and standards for measuring wood and bark mill residues
Jeffrey L. Wartluft
1976-01-01
Current information about wood and bark residues lacks the consistency needed to enable complete understanding and comparison from source to source. To make information about wood and bark residues more useful for production and marketing decisions, the Forest Products Marketing Laboratory of the USDA Forest Service and the Tennessee Valley Authority prepared this...
Discriminating Simulated Vocal Tremor Source Using Amplitude Modulation Spectra
Carbonell, Kathy M.; Lester, Rosemary A.; Story, Brad H.; Lotto, Andrew J.
2014-01-01
Objectives/Hypothesis Sources of vocal tremor are difficult to categorize perceptually and acoustically. This paper describes a preliminary attempt to discriminate vocal tremor sources through the use of spectral measures of the amplitude envelope. The hypothesis is that different vocal tremor sources are associated with distinct patterns of acoustic amplitude modulations. Study Design Statistical categorization methods (discriminant function analysis) were used to discriminate signals from simulated vocal tremor with different sources using only acoustic measures derived from the amplitude envelopes. Methods Simulations of vocal tremor were created by modulating parameters of a vocal fold model corresponding to oscillations of respiratory driving pressure (respiratory tremor), degree of vocal fold adduction (adductory tremor) and fundamental frequency of vocal fold vibration (F0 tremor). The acoustic measures were based on spectral analyses of the amplitude envelope computed across the entire signal and within select frequency bands. Results The signals could be categorized (with accuracy well above chance) in terms of the simulated tremor source using only measures of the amplitude envelope spectrum even when multiple sources of tremor were included. Conclusions These results supply initial support for an amplitude-envelope based approach to identify the source of vocal tremor and provide further evidence for the rich information about talker characteristics present in the temporal structure of the amplitude envelope. PMID:25532813
Evaluation of Long-term Performance of Enhanced Anaerobic Source Zone Bioremediation using mass flux
NASA Astrophysics Data System (ADS)
Haluska, A.; Cho, J.; Hatzinger, P.; Annable, M. D.
2017-12-01
Chlorinated ethene DNAPL source zones in groundwater act as potential long term sources of contamination as they dissolve yielding concentrations well above MCLs, posing an on-going public health risk. Enhanced bioremediation has been applied to treat many source zones with significant promise, but long-term sustainability of this technology has not been thoroughly assessed. This study evaluated the long-term effectiveness of enhanced anaerobic source zone bioremediation at chloroethene contaminated sites to determine if the treatment prevented contaminant rebound and removed NAPL from the source zone. Long-term performance was evaluated based on achieving MCL-based contaminant mass fluxes in parent compound concentrations during different monitoring periods. Groundwater concertation versus time data was compiled for 6-sites and post-remedial contaminant mass flux data was then measured using passive flux meters at wells both within and down-gradient of the source zone. Post-remedial mass flux data was then combined with pre-remedial water quality data to estimate pre-remedial mass flux. This information was used to characterize a DNAPL dissolution source strength function, such as the Power Law Model and the Equilibrium Stream tube model. The six-sites characterized for this study were (1) Former Charleston Air Force Base, Charleston, SC; (2) Dover Air Force Base, Dover, DE; (3) Treasure Island Naval Station, San Francisco, CA; (4) Former Raritan Arsenal, Edison, NJ; (5) Naval Air Station, Jacksonville, FL; and, (6) Former Naval Air Station, Alameda, CA. Contaminant mass fluxes decreased for all the sites by the end of the post-treatment monitoring period and rebound was limited within the source zone. Post remedial source strength function estimates suggest that decreases in contaminant mass flux will continue to occur at these sites, but a mass flux based on MCL levels may never be exceeded. Thus, site clean-up goals should be evaluated as order-of-magnitude reductions. Additionally, sites may require monitoring for a minimum of 5-years in order to sufficiently evaluate remedial performance. The study shows that enhanced anaerobic source zone bioremediation contributed to a modest reduction of source zone contaminant mass discharge and appears to have mitigated rebound of chlorinated ethenes.
NASA Astrophysics Data System (ADS)
Turbelin, Grégory; Singh, Sarvesh Kumar; Issartel, Jean-Pierre
2014-12-01
In the event of an accidental or intentional contaminant release in the atmosphere, it is imperative, for managing emergency response, to diagnose the release parameters of the source from measured data. Reconstruction of the source information exploiting measured data is called an inverse problem. To solve such a problem, several techniques are currently being developed. The first part of this paper provides a detailed description of one of them, known as the renormalization method. This technique, proposed by Issartel (2005), has been derived using an approach different from that of standard inversion methods and gives a linear solution to the continuous Source Term Estimation (STE) problem. In the second part of this paper, the discrete counterpart of this method is presented. By using matrix notation, common in data assimilation and suitable for numerical computing, it is shown that the discrete renormalized solution belongs to a family of well-known inverse solutions (minimum weighted norm solutions), which can be computed by using the concept of generalized inverse operator. It is shown that, when the weight matrix satisfies the renormalization condition, this operator satisfies the criteria used in geophysics to define good inverses. Notably, by means of the Model Resolution Matrix (MRM) formalism, we demonstrate that the renormalized solution fulfils optimal properties for the localization of single point sources. Throughout the article, the main concepts are illustrated with data from a wind tunnel experiment conducted at the Environmental Flow Research Centre at the University of Surrey, UK.
NASA Astrophysics Data System (ADS)
Vilain, J.
Approaches to major hazard assessment and prediction are reviewed. Source term: (phenomenology/modeling of release, influence on early stages of dispersion); dispersion (atmospheric advection, diffusion and deposition, emphasis on dense/cold gases); combustion (flammable clouds and mists covering flash fires, deflagration, transition to detonation; mostly unconfined/partly confined situations); blast formation, propagation, interaction with structures; catastrophic fires (pool fires, torches and fireballs; highly reactive substances) runaway reactions; features of more general interest; toxic substances, excluding toxicology; and dust explosions (phenomenology and protective measures) are discussed.
NASA Astrophysics Data System (ADS)
Pellerin, B. A.; Bergamaschi, B. A.; Downing, B. D.; Saraceno, J.; Fleck, J.; Shanley, J. B.; Aiken, G.; Boss, E.; Fujii, R.
2009-12-01
A critical challenge for understanding the sources, character and cycling of dissolved organic matter (DOM) is making measurements at the time scales in which changes occur in aquatic systems. Traditional approaches for data collection (daily to monthly discrete sampling) are often limited by analytical and field costs, site access and logistical challenges, particularly for long-term sampling at a large number of sites. The ability to make optical measurements of DOM in situ has been known for more than 50 years, but much of the work on in situ DOM absorbance and fluorescence using commercially-available instruments has taken place in the last few years. Here we present several recent examples that highlight the application of in situ measurements for understanding DOM dynamics in riverine systems at intervals of minutes to hours. Examples illustrate the utility of in situ optical sensors for studies of DOM over short-duration events of days to weeks (diurnal cycles, tidal cycles, storm events and snowmelt periods) as well as longer-term continuous monitoring for months to years. We also highlight the application of in situ optical DOM measurements as proxies for constituents that are significantly more difficult and expensive to measure at high frequencies (e.g. methylmercury, trihalomethanes). Relatively simple DOM absorbance and fluorescence measurements made in situ could be incorporated into short and long-term ecological research and monitoring programs, resulting in advanced understanding of organic matter sources, character and cycling in riverine systems.
Multiwavelength pyrometer for gray and non-gray surfaces in the presence of interfering radiation
NASA Technical Reports Server (NTRS)
Ng, Daniel L. P. (Inventor)
1994-01-01
A method and apparatus for detecting the temperature of gray and non-gray bodies in the presence of interfering radiation are presented. A gray body has a constant emissivity less than 1 and a non-gray body has an emissivity which varies with wavelength. The emissivity and reflectivity of the surface is determined over a range of wavelengths. Spectra are also measured of the extraneous interference radiation source and the surface of the object to be measured in the presence of the extraneous interference radiation source. An auxiliary radiation source is used to determine the reflectivity of the surface and also the emissivity. The measured spectrum of the surfaces in the presence of the extraneous interference radiation source is set equal to the emissivity of the surface multiplied by a Planck function containing a temperature term T plus the surface reflectivity multiplied by the spectrum of the extraneous interference radiation source. The equation is then solved for T to determine the temperature of the surface.
NASA Astrophysics Data System (ADS)
Zhu, Yiting; Narendran, Nadarajah; Tan, Jianchuan; Mou, Xi
2014-09-01
The organic light-emitting diode (OLED) has demonstrated its novelty in displays and certain lighting applications. Similar to white light-emitting diode (LED) technology, it also holds the promise of saving energy. Even though the luminous efficacy values of OLED products have been steadily growing, their longevity is still not well understood. Furthermore, currently there is no industry standard for photometric and colorimetric testing, short and long term, of OLEDs. Each OLED manufacturer tests its OLED panels under different electrical and thermal conditions using different measurement methods. In this study, an imaging-based photometric and colorimetric measurement method for OLED panels was investigated. Unlike an LED that can be considered as a point source, the OLED is a large form area source. Therefore, for an area source to satisfy lighting application needs, it is important that it maintains uniform light level and color properties across the emitting surface of the panel over a long period. This study intended to develop a measurement procedure that can be used to test long-term photometric and colorimetric properties of OLED panels. The objective was to better understand how test parameters such as drive current or luminance and temperature affect the degradation rate. In addition, this study investigated whether data interpolation could allow for determination of degradation and lifetime, L70, at application conditions based on the degradation rates measured at different operating conditions.
A model for jet-noise analysis using pressure-gradient correlations on an imaginary cone
NASA Technical Reports Server (NTRS)
Norum, T. D.
1974-01-01
The technique for determining the near and far acoustic field of a jet through measurements of pressure-gradient correlations on an imaginary conical surface surrounding the jet is discussed. The necessary analytical developments are presented, and their feasibility is checked by using a point source as the sound generator. The distribution of the apparent sources on the cone, equivalent to the point source, is determined in terms of the pressure-gradient correlations.
Source and long-term behavior of transuranic aerosols in the WIPP environment.
Thakur, P; Lemons, B G
2016-10-01
Source and long-term behavior transuranic aerosols ((239+240)Pu, (238)Pu, and (241)Am) in the ambient air samples collected at and near the Waste Isolation Pilot Plant (WIPP) deep geologic repository site were investigated using historical data from an independent monitoring program conducted by the Carlsbad Environmental Monitoring and Research Center and an oversight monitoring program conducted by the management and operating contractor for WIPP at and near the facility. An analysis of historical data indicates frequent detections of (239+240)Pu and (241)Am, whereas (238)Pu is detected infrequently. Peaks in (239+240)Pu and (241)Am concentrations in ambient air generally occur from March to June timeframe, which is when strong and gusty winds in the area frequently give rise to blowing dust. Long-term measurements of plutonium isotopes (1985-2015) in the WIPP environment suggest that the resuspension of previously contaminated soils is likely the primary source of plutonium in the ambient air samples from WIPP and its vicinity. There is no evidence that WIPP is a source of environmental contamination that can be considered significant by any health-based standard.
NASA Astrophysics Data System (ADS)
Dunlap, L.; Li, C.; Dickerson, R. R.; Krotkov, N. A.
2015-12-01
Weather systems, particularly mid-latitude wave cyclones, have been known to play an important role in the short-term variation of near-surface air pollution. Ground measurements and model simulations have demonstrated that stagnant air and minimal precipitation associated with high pressure systems are conducive to pollutant accumulation. With the passage of a cold front, built up pollution is transported downwind of the emission sources or washed out by precipitation. This concept is important to note when studying long-term changes in spatio-temporal pollution distribution, but has not been studied in detail from space. In this study, we focus on East Asia (especially the industrialized eastern China), where numerous large power plants and other point sources as well as area sources emit large amounts of SO2, an important gaseous pollutant and a precursor of aerosols. Using data from the Aura Ozone Monitoring Instrument (OMI) we show that such weather driven distribution can indeed be discerned from satellite data by utilizing probability distribution functions (PDFs) of SO2 column content. These PDFs are multimodal and give insight into the background pollution level at a given location and contribution from local and upwind emission sources. From these PDFs it is possible to determine the frequency for a given region to have SO2 loading that exceeds the background amount. By comparing OMI-observed long-term change in the frequency with meteorological data, we can gain insights into the effects of climate change (e.g., the weakening of Asian monsoon) on regional air quality. Such insight allows for better interpretation of satellite measurements as well as better prediction of future pollution distribution as a changing climate gives way to changing weather patterns.
NASA Astrophysics Data System (ADS)
Winiarek, Victor; Bocquet, Marc; Duhanyan, Nora; Roustan, Yelva; Saunier, Olivier; Mathieu, Anne
2013-04-01
A major difficulty when inverting the source term of an atmospheric tracer dispersion problem is the estimation of the prior errors: those of the atmospheric transport model, those ascribed to the representativeness of the measurements, the instrumental errors, and those attached to the prior knowledge on the variables one seeks to retrieve. In the case of an accidental release of pollutant, and specially in a situation of sparse observability, the reconstructed source is sensitive to these assumptions. This sensitivity makes the quality of the retrieval dependent on the methods used to model and estimate the prior errors of the inverse modeling scheme. In Winiarek et al. (2012), we proposed to use an estimation method for the errors' amplitude based on the maximum likelihood principle. Under semi-Gaussian assumptions, it takes into account, without approximation, the positivity assumption on the source. We applied the method to the estimation of the Fukushima Daiichi cesium-137 and iodine-131 source terms using activity concentrations in the air. The results were compared to an L-curve estimation technique, and to Desroziers's scheme. Additionally to the estimations of released activities, we provided related uncertainties (12 PBq with a std. of 15 - 20 % for cesium-137 and 190 - 380 PBq with a std. of 5 - 10 % for iodine-131). We also enlightened that, because of the low number of available observations (few hundreds) and even if orders of magnitude were consistent, the reconstructed activities significantly depended on the method used to estimate the prior errors. In order to use more data, we propose to extend the methods to the use of several data types, such as activity concentrations in the air and fallout measurements. The idea is to simultaneously estimate the prior errors related to each dataset, in order to fully exploit the information content of each one. Using the activity concentration measurements, but also daily fallout data from prefectures and cumulated deposition data over a region lying approximately 150 km around the nuclear power plant, we can use a few thousands of data in our inverse modeling algorithm to reconstruct the Cesium-137 source term. To improve the parameterization of removal processes, rainfall fields have also been corrected using outputs from the mesoscale meteorological model WRF and ground station rainfall data. As expected, the different methods yield closer results as the number of data increases. Reference : Winiarek, V., M. Bocquet, O. Saunier, A. Mathieu (2012), Estimation of errors in the inverse modeling of accidental release of atmospheric pollutant : Application to the reconstruction of the cesium-137 and iodine-131 source terms from the Fukushima Daiichi power plant, J. Geophys. Res., 117, D05122, doi:10.1029/2011JD016932.
Code of Federal Regulations, 2012 CFR
2012-07-01
... State elect to impose control measures on fossil fuel-fired NOX sources serving electric generators with... (g) of this section. (3) For purposes of paragraph (f)(2) of this section, the term “fossil fuel-fired” means, with regard to a NOX source: (i) The combustion of fossil fuel, alone or in combination...
Code of Federal Regulations, 2013 CFR
2013-07-01
... State elect to impose control measures on fossil fuel-fired NOX sources serving electric generators with... (g) of this section. (3) For purposes of paragraph (f)(2) of this section, the term “fossil fuel-fired” means, with regard to a NOX source: (i) The combustion of fossil fuel, alone or in combination...
Code of Federal Regulations, 2014 CFR
2014-07-01
... State elect to impose control measures on fossil fuel-fired NOX sources serving electric generators with... (g) of this section. (3) For purposes of paragraph (f)(2) of this section, the term “fossil fuel-fired” means, with regard to a NOX source: (i) The combustion of fossil fuel, alone or in combination...
Code of Federal Regulations, 2011 CFR
2011-07-01
... State elect to impose control measures on fossil fuel-fired NOX sources serving electric generators with... (g) of this section. (3) For purposes of paragraph (f)(2) of this section, the term “fossil fuel-fired” means, with regard to a NOX source: (i) The combustion of fossil fuel, alone or in combination...
Ravichandran, Ramamoorthy; Binukumar, Jp
2011-01-01
International Basic Safety Standards (International Atomic Energy Agency, IAEA) provide guidance levels for diagnostic procedures in nuclear medicine indicating the maximum usual activity for various diagnostic tests in terms of activities of injected radioactive formulations. An accuracy of ± 10% in the activities of administered radio-pharmaceuticals is being recommended, for expected outcome in diagnostic and therapeutic nuclear medicine procedures. It is recommended that the long-term stability of isotope calibrators used in nuclear medicine is to be checked periodically for their performance using a long-lived check source, such as Cs-137, of suitable activity. In view of the un-availability of such a radioactive source, we tried to develop methods to maintain traceability of these instruments, for certifying measured activities for human use. Two re-entrant chambers [(HDR 1000 and Selectron Source Dosimetry System (SSDS)] with I-125 and Ir-192 calibration factors in the Department of Radiotherapy were used to measure Iodine-131 (I-131) therapy capsules to establish traceability to Mark V isotope calibrator of the Department of Nuclear Medicine. Special nylon jigs were fabricated to keep I-131 capsule holder in position. Measured activities in all the chambers showed good agreement. The accuracy of SSDS chamber in measuring Ir-192 activities in the last 5 years was within 0.5%, validating its role as departmental standard for measuring activity. The above method is adopted because mean energies of I-131 and Ir-192 are comparable.
Analysis and Synthesis of Tonal Aircraft Noise Sources
NASA Technical Reports Server (NTRS)
Allen, Matthew P.; Rizzi, Stephen A.; Burdisso, Ricardo; Okcu, Selen
2012-01-01
Fixed and rotary wing aircraft operations can have a significant impact on communities in proximity to airports. Simulation of predicted aircraft flyover noise, paired with listening tests, is useful to noise reduction efforts since it allows direct annoyance evaluation of aircraft or operations currently in the design phase. This paper describes efforts to improve the realism of synthesized source noise by including short term fluctuations, specifically for inlet-radiated tones resulting from the fan stage of turbomachinery. It details analysis performed on an existing set of recorded turbofan data to isolate inlet-radiated tonal fan noise, then extract and model short term tonal fluctuations using the analytic signal. Methodologies for synthesizing time-variant tonal and broadband turbofan noise sources using measured fluctuations are also described. Finally, subjective listening test results are discussed which indicate that time-variant synthesized source noise is perceived to be very similar to recordings.
The low-frequency sound power measuring technique for an underwater source in a non-anechoic tank
NASA Astrophysics Data System (ADS)
Zhang, Yi-Ming; Tang, Rui; Li, Qi; Shang, Da-Jing
2018-03-01
In order to determine the radiated sound power of an underwater source below the Schroeder cut-off frequency in a non-anechoic tank, a low-frequency extension measuring technique is proposed. This technique is based on a unique relationship between the transmission characteristics of the enclosed field and those of the free field, which can be obtained as a correction term based on previous measurements of a known simple source. The radiated sound power of an unknown underwater source in the free field can thereby be obtained accurately from measurements in a non-anechoic tank. To verify the validity of the proposed technique, a mathematical model of the enclosed field is established using normal-mode theory, and the relationship between the transmission characteristics of the enclosed and free fields is obtained. The radiated sound power of an underwater transducer source is tested in a glass tank using the proposed low-frequency extension measuring technique. Compared with the free field, the radiated sound power level of the narrowband spectrum deviation is found to be less than 3 dB, and the 1/3 octave spectrum deviation is found to be less than 1 dB. The proposed testing technique can be used not only to extend the low-frequency applications of non-anechoic tanks, but also for measurement of radiated sound power from complicated sources in non-anechoic tanks.
Variability Search in GALFACTS
NASA Astrophysics Data System (ADS)
Kania, Joseph; Wenger, Trey; Ghosh, Tapasi; Salter, Christopher J.
2015-01-01
The Galactic ALFA Continuum Transit Survey (GALFACTS) is an all-Arecibo-sky survey using the seven-beam Arecibo L-band Feed Array (ALFA). The Survey is centered at 1.375 GHz with 300-MHz bandwidth, and measures all four Stokes parameters. We are looking for compact sources that vary in intensity or polarization on timescales of about a month via intra-survey comparisons and long term variations through comparisons with the NRAO VLA Sky Survey. Data processing includes locating and rejecting radio frequency interference, recognizing sources, two-dimensional Gaussian fitting to multiple cuts through the same source, and gain corrections. Our Python code is being used on the calibrations sources observed in conjunction with the survey measurements to determine the calibration parameters that will then be applied to data for the main field.
Associations of short-term exposure to fine particulate matter (PM2.5) with daily mortality may be due to specific PM2.5 chemical components. Objectives: Daily concentrations of PM2.5 chemical species were measured over five consecutive years in Denver, CO to investigate whethe...
NASA Technical Reports Server (NTRS)
Palosz, W.
2003-01-01
The amounts and composition of residual gases formed in sealed ampoules loaded with different sources (elements and II-VI and IV-VI compounds) after consecutive annealings were investigated. A given source was subjected to a series of heat treatments, with intermediate measurements and removal of the gas accumulated in the system. The results of these experiments are discussed in terms of the underlying thermochemical and kinetic phenomena and practical limitations of reducing the amount of residual gases in sealed ampoules.
Multisource Estimation of Long-term Global Terrestrial Surface Radiation
NASA Astrophysics Data System (ADS)
Peng, L.; Sheffield, J.
2017-12-01
Land surface net radiation is the essential energy source at the earth's surface. It determines the surface energy budget and its partitioning, drives the hydrological cycle by providing available energy, and offers heat, light, and energy for biological processes. Individual components in net radiation have changed historically due to natural and anthropogenic climate change and land use change. Decadal variations in radiation such as global dimming or brightening have important implications for hydrological and carbon cycles. In order to assess the trends and variability of net radiation and evapotranspiration, there is a need for accurate estimates of long-term terrestrial surface radiation. While large progress in measuring top of atmosphere energy budget has been made, huge discrepancies exist among ground observations, satellite retrievals, and reanalysis fields of surface radiation, due to the lack of observational networks, the difficulty in measuring from space, and the uncertainty in algorithm parameters. To overcome the weakness of single source datasets, we propose a multi-source merging approach to fully utilize and combine multiple datasets of radiation components separately, as they are complementary in space and time. First, we conduct diagnostic analysis of multiple satellite and reanalysis datasets based on in-situ measurements such as Global Energy Balance Archive (GEBA), existing validation studies, and other information such as network density and consistency with other meteorological variables. Then, we calculate the optimal weighted average of multiple datasets by minimizing the variance of error between in-situ measurements and other observations. Finally, we quantify the uncertainties in the estimates of surface net radiation and employ physical constraints based on the surface energy balance to reduce these uncertainties. The final dataset is evaluated in terms of the long-term variability and its attribution to changes in individual components. The goal of this study is to provide a merged observational benchmark for large-scale diagnostic analyses, remote sensing and land surface modeling.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Genn Saji
2006-07-01
The term 'ultimate risk' is used here to describe the probabilities and radiological consequences that should be incorporated in siting, containment design and accident management of nuclear power plants for hypothetical accidents. It is closely related with the source terms specified in siting criteria which assures an adequate separation of radioactive inventories of the plants from the public, in the event of a hypothetical and severe accident situation. The author would like to point out that current source terms which are based on the information from the Windscale accident (1957) through TID-14844 are very outdated and do not incorporate lessonsmore » learned from either the Three Miles Island (TMI, 1979) nor Chernobyl accident (1986), two of the most severe accidents ever experienced. As a result of the observations of benign radionuclides released at TMI, the technical community in the US felt that a more realistic evaluation of severe reactor accident source terms was necessary. In this background, the 'source term research project' was organized in 1984 to respond to these challenges. Unfortunately, soon after the time of the final report from this project was released, the Chernobyl accident occurred. Due to the enormous consequences induced by then accident, the one time optimistic perspectives in establishing a more realistic source term were completely shattered. The Chernobyl accident, with its human death toll and dispersion of a large part of the fission fragments inventories into the environment, created a significant degradation in the public's acceptance of nuclear energy throughout the world. In spite of this, nuclear communities have been prudent in responding to the public's anxiety towards the ultimate safety of nuclear plants, since there still remained many unknown points revolving around the mechanism of the Chernobyl accident. In order to resolve some of these mysteries, the author has performed a scoping study of the dispersion and deposition mechanisms of fuel particles and fission fragments during the initial phase of the Chernobyl accident. Through this study, it is now possible to generally reconstruct the radiological consequences by using a dispersion calculation technique, combined with the meteorological data at the time of the accident and land contamination densities of {sup 137}Cs measured and reported around the Chernobyl area. Although it is challenging to incorporate lessons learned from the Chernobyl accident into the source term issues, the author has already developed an example of safety goals by incorporating the radiological consequences of the accident. The example provides safety goals by specifying source term releases in a graded approach in combination with probabilities, i.e. risks. The author believes that the future source term specification should be directly linked with safety goals. (author)« less
The time variability of Jupiter's synchrotron radiation
NASA Astrophysics Data System (ADS)
Bolton, Scott Jay
1991-02-01
The time variability of the Jovian synchrotron emission is investigated by analyzing radio observations of Jupiter at decimetric wavelengths. The observations are composed from two distinct sets of measurements addressing both short term (days to weeks) and long term (months to years) variability. The study of long term variations utilizes a set of measurements made several times each month with the NASA Deep Space Network (DNS) antennas operating at 2295 MHz (13.1 cm). The DSN data set, covering 1971 through 1985, is compared with a set of measurements of the solar wind from a number of Earth orbiting spacecraft. The analysis indicates a maximum correlation between the synchrotron emission and the solar wind ram pressure with a two year time lag. Physical mechanisms affecting the synchrotron emission are discussed with an emphasis on radial diffusion. Calculations are performed that suggest the correlation is consistent with inward adiabatic diffusion of solar wind particles driven by Brice's model of ionospheric neutral wind convection (Brice 1972). The implication is that the solar wind could be a source of particles of Jupiter's radiation belts. The investigation of short term variability focuses on a three year Jupiter observing program using the University of California's Hat Creek radio telescope operating at 1400 MHz (21 cm). Measurements are made every two days during the months surrounding opposition. Results from the three year program suggest short term variability near the 10-20 percent level but should be considered inconclusive due to scheduling and observational limitations. A discussion of magneto-spheric processes on short term timescales identifies wave-particle interactions as a candidate source. Further analysis finds that the short term variations could be related to whistler mode wave-particles interactions in the radiation belts associated with atmospheric lightning on Jupiter. However, theoretical calculations on wave particle interactions imply thought if whistler mode waves are to interact with the synchrotron emitting electrons.
NASA Technical Reports Server (NTRS)
Horsham, Gray A. P.
1998-01-01
Market research sources were used to initially gather primary technological problems and needs data from non-aerospace companies in targeted industry sectors. The company-supplied information served as input data to activate or start-up an internal, phased match-making process. This process was based on technical-level relationship exploration followed by business-level agreement negotiations, and culminated with project management and execution. Space Act Agreements represented near-term outputs. Company product or process commercialization derived from Lewis support and measurable economic effects represented far-term outputs.
NASA Astrophysics Data System (ADS)
Fu, Shihang; Zhang, Li; Hu, Yao; Ding, Xiang
2018-01-01
Confocal Raman Microscopy (CRM) has matured to become one of the most powerful instruments in analytical science because of its molecular sensitivity and high spatial resolution. Compared with conventional Raman Microscopy, CRM can perform three dimensions mapping of tiny samples and has the advantage of high spatial resolution thanking to the unique pinhole. With the wide application of the instrument, there is a growing requirement for the evaluation of the imaging performance of the system. Point-spread function (PSF) is an important approach to the evaluation of imaging capability of an optical instrument. Among a variety of measurement methods of PSF, the point source method has been widely used because it is easy to operate and the measurement results are approximate to the true PSF. In the point source method, the point source size has a significant impact on the final measurement accuracy. In this paper, the influence of the point source sizes on the measurement accuracy of PSF is analyzed and verified experimentally. A theoretical model of the lateral PSF for CRM is established and the effect of point source size on full-width at half maximum of lateral PSF is simulated. For long-term preservation and measurement convenience, PSF measurement phantom using polydimethylsiloxane resin, doped with different sizes of polystyrene microspheres is designed. The PSF of CRM with different sizes of microspheres are measured and the results are compared with the simulation results. The results provide a guide for measuring the PSF of the CRM.
Rey-Martinez, Jorge; Pérez-Fernández, Nicolás
2016-12-01
The proposed validation goal of 0.9 in intra-class correlation coefficient was reached with the results of this study. With the obtained results we consider that the developed software (RombergLab) is a validated balance assessment software. The reliability of this software is dependent of the used force platform technical specifications. Develop and validate a posturography software and share its source code in open source terms. Prospective non-randomized validation study: 20 consecutive adults underwent two balance assessment tests, six condition posturography was performed using a clinical approved software and force platform and the same conditions were measured using the new developed open source software using a low cost force platform. Intra-class correlation index of the sway area obtained from the center of pressure variations in both devices for the six conditions was the main variable used for validation. Excellent concordance between RombergLab and clinical approved force platform was obtained (intra-class correlation coefficient =0.94). A Bland and Altman graphic concordance plot was also obtained. The source code used to develop RombergLab was published in open source terms.
Rotational and X-ray luminosity evolution of high-B radio pulsars
NASA Astrophysics Data System (ADS)
Benli, Onur; Ertan, Ünal
2018-05-01
In continuation of our earlier work on the long-term evolution of the so-called high-B radio pulsars (HBRPs) with measured braking indices, we have investigated the long-term evolution of the remaining five HBRPs for which braking indices have not been measured yet. This completes our source-by-source analyses of HBRPs in the fallback disc model that was also applied earlier to anomalous X-ray pulsars (AXPs), soft gamma repeaters (SGRs), and dim isolated neutron stars (XDINs). Our results show that the X-ray luminosities and the rotational properties of these rather different neutron star populations can be acquired by neutron stars with fallback discs as a result of differences in their initial conditions, namely the initial disc mass, initial period and the dipole field strength. For the five HBRPs, unlike for AXPs, SGRs and XDINs, our results do not constrain the dipole field strengths of the sources. We obtain evolutionary paths leading to the properties of HBRPs in the propeller phase with dipole fields sufficiently strong to produce pulsed radio emission.
Snow, Mathew S.; Snyder, Darin C.; Delmore, James E.
2016-01-18
Source term attribution of environmental contamination following the Fukushima Daiichi Nuclear Power Plant (FDNPP) disaster is complicated by a large number of possible similar emission source terms (e.g. FDNPP reactor cores 1–3 and spent fuel ponds 1–4). Cesium isotopic analyses can be utilized to discriminate between environmental contamination from different FDNPP source terms and, if samples are sufficiently temporally resolved, potentially provide insights into the extent of reactor core damage at a given time. Rice, soil, mushroom, and soybean samples taken 100–250 km from the FDNPP site were dissolved using microwave digestion. Radiocesium was extracted and purified using two sequentialmore » ammonium molybdophosphate-polyacrylonitrile columns, following which 135Cs/ 137Cs isotope ratios were measured using thermal ionization mass spectrometry (TIMS). Results were compared with data reported previously from locations to the northwest of FDNPP and 30 km to the south of FDNPP. 135Cs/ 137Cs isotope ratios from samples 100–250 km to the southwest of the FDNPP site show a consistent value of 0.376 ± 0.008. 135Cs/ 137Cs versus 134Cs/ 137Cs correlation plots suggest that radiocesium to the southwest is derived from a mixture of FDNPP reactor cores 1, 2, and 3. Conclusions from the cesium isotopic data are in agreement with those derived independently based upon the event chronology combined with meteorological conditions at the time of the disaster. In conclusion, cesium isotopic analyses provide a powerful tool for source term discrimination of environmental radiocesium contamination at the FDNPP site. For higher precision source term attribution and forensic determination of the FDNPP core conditions based upon cesium, analyses of a larger number of samples from locations to the north and south of the FDNPP site (particularly time-resolved air filter samples) are needed. Published in 2016. This article is a U.S. Government work and is in the public domain in the USA.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Snow, Mathew S.; Snyder, Darin C.; Delmore, James E.
Source term attribution of environmental contamination following the Fukushima Daiichi Nuclear Power Plant (FDNPP) disaster is complicated by a large number of possible similar emission source terms (e.g. FDNPP reactor cores 1–3 and spent fuel ponds 1–4). Cesium isotopic analyses can be utilized to discriminate between environmental contamination from different FDNPP source terms and, if samples are sufficiently temporally resolved, potentially provide insights into the extent of reactor core damage at a given time. Rice, soil, mushroom, and soybean samples taken 100–250 km from the FDNPP site were dissolved using microwave digestion. Radiocesium was extracted and purified using two sequentialmore » ammonium molybdophosphate-polyacrylonitrile columns, following which 135Cs/ 137Cs isotope ratios were measured using thermal ionization mass spectrometry (TIMS). Results were compared with data reported previously from locations to the northwest of FDNPP and 30 km to the south of FDNPP. 135Cs/ 137Cs isotope ratios from samples 100–250 km to the southwest of the FDNPP site show a consistent value of 0.376 ± 0.008. 135Cs/ 137Cs versus 134Cs/ 137Cs correlation plots suggest that radiocesium to the southwest is derived from a mixture of FDNPP reactor cores 1, 2, and 3. Conclusions from the cesium isotopic data are in agreement with those derived independently based upon the event chronology combined with meteorological conditions at the time of the disaster. In conclusion, cesium isotopic analyses provide a powerful tool for source term discrimination of environmental radiocesium contamination at the FDNPP site. For higher precision source term attribution and forensic determination of the FDNPP core conditions based upon cesium, analyses of a larger number of samples from locations to the north and south of the FDNPP site (particularly time-resolved air filter samples) are needed. Published in 2016. This article is a U.S. Government work and is in the public domain in the USA.« less
Snow, Mathew S; Snyder, Darin C; Delmore, James E
2016-02-28
Source term attribution of environmental contamination following the Fukushima Daiichi Nuclear Power Plant (FDNPP) disaster is complicated by a large number of possible similar emission source terms (e.g. FDNPP reactor cores 1-3 and spent fuel ponds 1-4). Cesium isotopic analyses can be utilized to discriminate between environmental contamination from different FDNPP source terms and, if samples are sufficiently temporally resolved, potentially provide insights into the extent of reactor core damage at a given time. Rice, soil, mushroom, and soybean samples taken 100-250 km from the FDNPP site were dissolved using microwave digestion. Radiocesium was extracted and purified using two sequential ammonium molybdophosphate-polyacrylonitrile columns, following which (135)Cs/(137) Cs isotope ratios were measured using thermal ionization mass spectrometry (TIMS). Results were compared with data reported previously from locations to the northwest of FDNPP and 30 km to the south of FDNPP. (135)Cs/(137)Cs isotope ratios from samples 100-250 km to the southwest of the FDNPP site show a consistent value of 0.376 ± 0.008. (135)Cs/(137)Cs versus (134)Cs/(137)Cs correlation plots suggest that radiocesium to the southwest is derived from a mixture of FDNPP reactor cores 1, 2, and 3. Conclusions from the cesium isotopic data are in agreement with those derived independently based upon the event chronology combined with meteorological conditions at the time of the disaster. Cesium isotopic analyses provide a powerful tool for source term discrimination of environmental radiocesium contamination at the FDNPP site. For higher precision source term attribution and forensic determination of the FDNPP core conditions based upon cesium, analyses of a larger number of samples from locations to the north and south of the FDNPP site (particularly time-resolved air filter samples) are needed. Published in 2016. This article is a U.S. Government work and is in the public domain in the USA.
NASA Astrophysics Data System (ADS)
Gelado-Caballero, MaríA. D.; López-GarcíA, Patricia; Prieto, Sandra; Patey, Matthew D.; Collado, Cayetano; HéRnáNdez-Brito, José J.
2012-02-01
There are very few sets of long-term measurements of aerosol concentrations over the North Atlantic Ocean, yet such data is invaluable in quantifying atmospheric dust inputs to this ocean region. We present an 8-year record of total suspended particles (TSP) collected at three stations on Gran Canaria Island, Spain (Taliarte at sea level, Tafira 269 m above sea level (a.s.l.) and Pico de la Gorra 1930 m a.s.l.). Using wet and dry deposition measurements, the mean dust flux was calculated at 42.3 mg m-2 d-1. Air mass back trajectories (HYSPLIT, NOAA) suggested that the Sahara desert is the major source of African dust (dominant during 32-50% of days), while the Sahel desert was the major source only 2-10% of the time (maximum in summer). Elemental composition ratios of African samples indicate that, despite the homogeneity of the dust in collected samples, some signatures of the bedrocks can still be detected. Differences were found for the Sahel, Central Sahara and North of Sahara regions in Ti/Al, Mg/Al and Ca/Al ratios, respectively. Elements often associated with pollution (Pb, Cd, Ni, Zn) appeared to share a common origin, while Cu may have a predominantly local source, as suggested by a decrease in the enrichment factor (EF) of Cu during dust events. The inter-annual variability of dust concentrations is investigated in this work. During winter, African dust concentration measurements at the Pico de la Gorra station were found to correlate with the North Atlantic Oscillation (NAO) index.
Gallastegi, Mara; Guxens, Mònica; Jiménez-Zabala, Ana; Calvente, Irene; Fernández, Marta; Birks, Laura; Struchen, Benjamin; Vrijheid, Martine; Estarlich, Marisa; Fernández, Mariana F; Torrent, Maties; Ballester, Ferrán; Aurrekoetxea, Juan J; Ibarluzea, Jesús; Guerra, David; González, Julián; Röösli, Martin; Santa-Marina, Loreto
2016-02-18
Analysis of the association between exposure to electromagnetic fields of non-ionising radiation (EMF-NIR) and health in children and adolescents is hindered by the limited availability of data, mainly due to the difficulties on the exposure assessment. This study protocol describes the methodologies used for characterising exposure of children to EMF-NIR in the INMA (INfancia y Medio Ambiente- Environment and Childhood) Project, a prospective cohort study. Indirect (proximity to emission sources, questionnaires on sources use and geospatial propagation models) and direct methods (spot and fixed longer-term measurements and personal measurements) were conducted in order to assess exposure levels of study participants aged between 7 and 18 years old. The methodology used varies depending on the frequency of the EMF-NIR and the environment (homes, schools and parks). Questionnaires assessed the use of sources contributing both to Extremely Low Frequency (ELF) and Radiofrequency (RF) exposure levels. Geospatial propagation models (NISMap) are implemented and validated for environmental outdoor sources of RFs using spot measurements. Spot and fixed longer-term ELF and RF measurements were done in the environments where children spend most of the time. Moreover, personal measurements were taken in order to assess individual exposure to RF. The exposure data are used to explore their relationships with proximity and/or use of EMF-NIR sources. Characterisation of the EMF-NIR exposure by this combination of methods is intended to overcome problems encountered in other research. The assessment of exposure of INMA cohort children and adolescents living in different regions of Spain to the full frequency range of EMF-NIR extends the characterisation of environmental exposures in this cohort. Together with other data obtained in the project, on socioeconomic and family characteristics and development of the children and adolescents, this will enable to evaluate the complex interaction between health outcomes in children and adolescents and the various environmental factors that surround them.
A time reversal algorithm in acoustic media with Dirac measure approximations
NASA Astrophysics Data System (ADS)
Bretin, Élie; Lucas, Carine; Privat, Yannick
2018-04-01
This article is devoted to the study of a photoacoustic tomography model, where one is led to consider the solution of the acoustic wave equation with a source term writing as a separated variables function in time and space, whose temporal component is in some sense close to the derivative of the Dirac distribution at t = 0. This models a continuous wave laser illumination performed during a short interval of time. We introduce an algorithm for reconstructing the space component of the source term from the measure of the solution recorded by sensors during a time T all along the boundary of a connected bounded domain. It is based at the same time on the introduction of an auxiliary equivalent Cauchy problem allowing to derive explicit reconstruction formula and then to use of a deconvolution procedure. Numerical simulations illustrate our approach. Finally, this algorithm is also extended to elasticity wave systems.
On the role of mean flows in Doppler shifted frequencies
NASA Astrophysics Data System (ADS)
Gerkema, Theo; Maas, Leo R. M.; van Haren, Hans
2013-04-01
In the oceanographic literature, the term 'Doppler shift' often features in the context of mean flows and (internal) waves. Closer inspection reveals that the term is in fact used for two different things, which should be carefully distinguished, for their conflation results in incorrect interpretations. One refers to the difference in frequencies measured by two observers, one at a fixed position and one moving with the mean flow. The other definition is the one used in physics, where the frequency measured by an observer is compared to that of the source. In the latter sense, Doppler shifts occur only if the source and observer move with respect to each other; a steady mean flow cannot create a Doppler shift. We rehash the classical theory to straighten out some misconceptions and discuss how wave dispersion affects the classical relations and their application, for example on near-inertial internal waves.
Grady, Caitlin A; Kipkorir, Emmanuel C; Nguyen, Kien; Blatchley, E R
2015-06-01
In recent decades, more than 2 billion people have gained access to improved drinking water sources thanks to extensive effort from governments, and public and private sector entities. Despite this progress, many water sector development interventions do not provide access to safe water or fail to be sustained for long-term use. The authors examined drinking water quality of previously implemented water improvement projects in three communities in western Kenya and three communities in southern Vietnam. The cross-sectional study of 219 households included measurements of viable Escherichia coli. High rates of E. coli prevalence in these improved water sources were found in many of the samples. These findings suggest that measures above and beyond the traditional 'improved source' definition may be necessary to ensure truly safe water throughout these regions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gambino, Nadia, E-mail: gambinon@ethz.ch; Brandstätter, Markus; Rollinger, Bob
2014-09-15
In this work, a new diagnostic tool for laser-produced plasmas (LPPs) is presented. The detector is based on a multiple array of six motorized Langmuir probes. It allows to measure the dynamics of a LPP in terms of charged particles detection with particular attention to droplet-based LPP sources for EUV lithography. The system design permits to temporally resolve the angular and radial plasma charge distribution and to obtain a hemispherical mapping of the ions and electrons around the droplet plasma. The understanding of these dynamics is fundamental to improve the debris mitigation techniques for droplet-based LPP sources. The device hasmore » been developed, built, and employed at the Laboratory for Energy Conversion, ETH Zürich. The experimental results have been obtained on the droplet-based LPP source ALPS II. For the first time, 2D mappings of the ion kinetic energy distribution around the droplet plasma have been obtained with an array of multiple Langmuir probes. These measurements show an anisotropic expansion of the ions in terms of kinetic energy and amount of ion charge around the droplet target. First estimations of the plasma density and electron temperature were also obtained from the analysis of the probe current signals.« less
Mass discharge assessment at a brominated DNAPL site: Effects of known DNAPL source mass removal
NASA Astrophysics Data System (ADS)
Johnston, C. D.; Davis, G. B.; Bastow, T. P.; Woodbury, R. J.; Rao, P. S. C.; Annable, M. D.; Rhodes, S.
2014-08-01
Management and closure of contaminated sites is increasingly being proposed on the basis of mass flux of dissolved contaminants in groundwater. Better understanding of the links between source mass removal and contaminant mass fluxes in groundwater would allow greater acceptance of this metric in dealing with contaminated sites. Our objectives here were to show how measurements of the distribution of contaminant mass flux and the overall mass discharge emanating from the source under undisturbed groundwater conditions could be related to the processes and extent of source mass depletion. In addition, these estimates of mass discharge were sought in the application of agreed remediation targets set in terms of pumped groundwater quality from offsite wells. Results are reported from field studies conducted over a 5-year period at a brominated DNAPL (tetrabromoethane, TBA; and tribromoethene, TriBE) site located in suburban Perth, Western Australia. Groundwater fluxes (qw; L3/L2/T) and mass fluxes (Jc; M/L2/T) of dissolved brominated compounds were simultaneously estimated by deploying Passive Flux Meters (PFMs) in wells in a heterogeneous layered aquifer. PFMs were deployed in control plane (CP) wells immediately down-gradient of the source zone, before (2006) and after (2011) 69-85% of the source mass was removed, mainly by groundwater pumping from the source zone. The high-resolution (26-cm depth interval) measures of qw and Jc along the source CP allowed investigation of the DNAPL source-zone architecture and impacts of source mass removal. Comparable estimates of total mass discharge (MD; M/T) across the source zone CP reduced from 104 g day- 1 to 24-31 g day- 1 (70-77% reductions). Importantly, this mass discharge reduction was consistent with the estimated proportion of source mass remaining at the site (15-31%). That is, a linear relationship between mass discharge and source mass is suggested. The spatial detail of groundwater and mass flux distributions also provided further evidence of the source zone architecture and DNAPL mass depletion processes. This was especially apparent in different mass-depletion rates from distinct parts of the CP. High mass fluxes and groundwater fluxes located near the base of the aquifer dominated in terms of the dissolved mass flux in the profile, although not in terms of concentrations. Reductions observed in Jc and MD were used to better target future remedial efforts. Integration of the observations from the PFM deployments and the source mass depletion provided a basis for establishing flux-based management criteria for the site.
Mass discharge assessment at a brominated DNAPL site: Effects of known DNAPL source mass removal.
Johnston, C D; Davis, G B; Bastow, T P; Woodbury, R J; Rao, P S C; Annable, M D; Rhodes, S
2014-08-01
Management and closure of contaminated sites is increasingly being proposed on the basis of mass flux of dissolved contaminants in groundwater. Better understanding of the links between source mass removal and contaminant mass fluxes in groundwater would allow greater acceptance of this metric in dealing with contaminated sites. Our objectives here were to show how measurements of the distribution of contaminant mass flux and the overall mass discharge emanating from the source under undisturbed groundwater conditions could be related to the processes and extent of source mass depletion. In addition, these estimates of mass discharge were sought in the application of agreed remediation targets set in terms of pumped groundwater quality from offsite wells. Results are reported from field studies conducted over a 5-year period at a brominated DNAPL (tetrabromoethane, TBA; and tribromoethene, TriBE) site located in suburban Perth, Western Australia. Groundwater fluxes (qw; L(3)/L(2)/T) and mass fluxes (Jc; M/L(2)/T) of dissolved brominated compounds were simultaneously estimated by deploying Passive Flux Meters (PFMs) in wells in a heterogeneous layered aquifer. PFMs were deployed in control plane (CP) wells immediately down-gradient of the source zone, before (2006) and after (2011) 69-85% of the source mass was removed, mainly by groundwater pumping from the source zone. The high-resolution (26-cm depth interval) measures of qw and Jc along the source CP allowed investigation of the DNAPL source-zone architecture and impacts of source mass removal. Comparable estimates of total mass discharge (MD; M/T) across the source zone CP reduced from 104gday(-1) to 24-31gday(-1) (70-77% reductions). Importantly, this mass discharge reduction was consistent with the estimated proportion of source mass remaining at the site (15-31%). That is, a linear relationship between mass discharge and source mass is suggested. The spatial detail of groundwater and mass flux distributions also provided further evidence of the source zone architecture and DNAPL mass depletion processes. This was especially apparent in different mass-depletion rates from distinct parts of the CP. High mass fluxes and groundwater fluxes located near the base of the aquifer dominated in terms of the dissolved mass flux in the profile, although not in terms of concentrations. Reductions observed in Jc and MD were used to better target future remedial efforts. Integration of the observations from the PFM deployments and the source mass depletion provided a basis for establishing flux-based management criteria for the site. Copyright © 2013 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Tichý, Ondřej; Šmídl, Václav; Hofman, Radek; Šindelářová, Kateřina; Hýža, Miroslav; Stohl, Andreas
2017-10-01
In the fall of 2011, iodine-131 (131I) was detected at several radionuclide monitoring stations in central Europe. After investigation, the International Atomic Energy Agency (IAEA) was informed by Hungarian authorities that 131I was released from the Institute of Isotopes Ltd. in Budapest, Hungary. It was reported that a total activity of 342 GBq of 131I was emitted between 8 September and 16 November 2011. In this study, we use the ambient concentration measurements of 131I to determine the location of the release as well as its magnitude and temporal variation. As the location of the release and an estimate of the source strength became eventually known, this accident represents a realistic test case for inversion models. For our source reconstruction, we use no prior knowledge. Instead, we estimate the source location and emission variation using only the available 131I measurements. Subsequently, we use the partial information about the source term available from the Hungarian authorities for validation of our results. For the source determination, we first perform backward runs of atmospheric transport models and obtain source-receptor sensitivity (SRS) matrices for each grid cell of our study domain. We use two dispersion models, FLEXPART and Hysplit, driven with meteorological analysis data from the global forecast system (GFS) and from European Centre for Medium-range Weather Forecasts (ECMWF) weather forecast models. Second, we use a recently developed inverse method, least-squares with adaptive prior covariance (LS-APC), to determine the 131I emissions and their temporal variation from the measurements and computed SRS matrices. For each grid cell of our simulation domain, we evaluate the probability that the release was generated in that cell using Bayesian model selection. The model selection procedure also provides information about the most suitable dispersion model for the source term reconstruction. Third, we select the most probable location of the release with its associated source term and perform a forward model simulation to study the consequences of the iodine release. Results of these procedures are compared with the known release location and reported information about its time variation. We find that our algorithm could successfully locate the actual release site. The estimated release period is also in agreement with the values reported by IAEA and the reported total released activity of 342 GBq is within the 99 % confidence interval of the posterior distribution of our most likely model.
NASA Astrophysics Data System (ADS)
Brocklehurst, Aidan; Boon, Alex; Barlow, Janet; Hayden, Paul; Robins, Alan
2014-05-01
The source area of an instrument is an estimate of the area of ground over which the measurement is generated. Quantification of the source area of a measurement site provides crucial context for analysis and interpretation of the data. A range of computational models exists to calculate the source area of an instrument, but these are usually based on assumptions which do not hold for instruments positioned very close to the surface, particularly those surrounded by heterogeneous terrain i.e. urban areas. Although positioning instrumentation at higher elevation (i.e. on masts) is ideal in urban areas, this can be costly in terms of installation and maintenance costs and logistically difficult to position instruments in the ideal geographical location. Therefore, in many studies, experimentalists turn to rooftops to position instrumentation. Experimental validations of source area models for these situations are very limited. In this study, a controlled tracer gas experiment was conducted in a wind tunnel based on a 1:200 scale model of a measurement site used in previous experimental work in central London. The detector was set at the location of the rooftop site as the tracer was released at a range of locations within the surrounding streets and rooftops. Concentration measurements are presented for a range of wind angles, with the spread of concentration measurements indicative of the source area distribution. Clear evidence of wind channeling by streets is seen with the shape of the source area strongly influenced by buildings upwind of the measurement point. The results of the wind tunnel study are compared to scalar concentration source areas generated by modelling approaches based on meteorological data from the central London experimental site and used in the interpretation of continuous carbon dioxide (CO2) concentration data. Initial conclusions will be drawn as to how to apply scalar concentration source area models to rooftop measurement sites and suggestions for their improvement to incorporate effects such as channeling.
Yi, Qitao; Chen, Qiuwen; Hu, Liuming; Shi, Wenqing
2017-05-16
This research developed an innovative approach to reveal nitrogen sources, transformation, and transport in large and complex river networks in the Taihu Lake basin using measurement of dual stable isotopes of nitrate. The spatial patterns of δ 15 N corresponded to the urbanization level, and the nitrogen cycle was associated with the hydrological regime at the basin level. During the high flow season of summer, nonpoint sources from fertilizer/soils and atmospheric deposition constituted the highest proportion of the total nitrogen load. The point sources from sewage/manure, with high ammonium concentrations and high δ 15 N and δ 18 O contents in the form of nitrate, accounted for the largest inputs among all sources during the low flow season of winter. Hot spot areas with heavy point source pollution were identified, and the pollutant transport routes were revealed. Nitrification occurred widely during the warm seasons, with decreased δ 18 O values; whereas great potential for denitrification existed during the low flow seasons of autumn and spring. The study showed that point source reduction could have effects over the short-term; however, long-term efforts to substantially control agriculture nonpoint sources are essential to eutrophication alleviation for the receiving lake, which clarifies the relationship between point and nonpoint source control.
Calibration of Photon Sources for Brachytherapy
NASA Astrophysics Data System (ADS)
Rijnders, Alex
Source calibration has to be considered an essential part of the quality assurance program in a brachytherapy department. Not only it will ensure that the source strength value used for dose calculation agrees within some predetermined limits to the value stated on the source certificate, but also it will ensure traceability to international standards. At present calibration is most often still given in terms of reference air kerma rate, although calibration in terms of absorbed dose to water would be closer to the users interest. It can be expected that in a near future several standard laboratories will be able to offer this latter service, and dosimetry protocols will have to be adapted in this way. In-air measurement using ionization chambers (e.g. a Baldwin—Farmer ionization chamber for 192Ir high dose rate HDR or pulsed dose rate PDR sources) is still considered the method of choice for high energy source calibration, but because of their ease of use and reliability well type chambers are becoming more popular and are nowadays often recommended as the standard equipment. For low energy sources well type chambers are in practice the only equipment available for calibration. Care should be taken that the chamber is calibrated at the standard laboratory for the same source type and model as used in the clinic, and using the same measurement conditions and setup. Several standard laboratories have difficulties to provide these calibration facilities, especially for the low energy seed sources (125I and 103Pd). Should a user not be able to obtain properly calibrated equipment to verify the brachytherapy sources used in his department, then at least for sources that are replaced on a regular basis, a consistency check program should be set up to ensure a minimal level of quality control before these sources are used for patient treatment.
Long-Term Stability of the NIST Standard Ultrasonic Source.
Fick, Steven E
2008-01-01
The National Institute of Standards and Technology (NIST) Standard Ultrasonic Source (SUS) is a system comprising a transducer capable of output power levels up to 1 W at multiple frequencies between 1 MHz and 30 MHz, and an electrical impedance-matching network that allows the system to be driven by a conventional 50 Ω rf (radio-frequency) source. It is designed to allow interlaboratory replication of ultrasonic power levels with high accuracy using inexpensive readily available ancillary equipment. The SUS was offered for sale for 14 years (1985 to 1999). Each system was furnished with data for the set of calibration points (combinations of power level and frequency) specified by the customer. Of the systems that had been ordered with some calibration points in common, three were returned more than once to NIST for recalibration. Another system retained at NIST has been recalibrated periodically since 1984. The collective data for these systems comprise 9 calibration points and 102 measurements spanning a 17 year interval ending in 2001, the last year NIST ultrasonic power measurement services were available to the public. These data have been analyzed to compare variations in output power with frequency, power level, and time elapsed since the first calibration. The results verify the claim, made in the instruction sheet furnished with every SUS, that "long-term drift, if any, in the calibration of NIST Standard Sources is insignificant compared to the uncertainties associated with a single measurement of ultrasonic power by any method available at NIST."
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lee, Yuna; Park, Yeong-Shin; Jo, Jong-Gab
2012-02-15
Microwave plasma ion source with rectangular cavity resonator has been examined to improve ion beam current by changing wave launcher type from single-port to double-port. The cavity resonators with double-port and single-port wave launchers are designed to get resonance effect at TE-103 mode and TE-102 mode, respectively. In order to confirm that the cavities are acting as resonator, the microwave power for breakdown is measured and compared with the E-field strength estimated from the HFSS (High Frequency Structure Simulator) simulation. Langmuir probe measurements show that double-port cavity enhances central density of plasma ion source by modifying non-uniform plasma density profilemore » of the single-port cavity. Correspondingly, beam current from the plasma ion source utilizing the double-port resonator is measured to be higher than that utilizing single-port resonator. Moreover, the enhancement in plasma density and ion beam current utilizing the double-port resonator is more pronounced as higher microwave power applied to the plasma ion source. Therefore, the rectangular cavity resonator utilizing the double-port is expected to enhance the performance of plasma ion source in terms of ion beam extraction.« less
Lee, Yuna; Park, Yeong-Shin; Jo, Jong-Gab; Yang, J J; Hwang, Y S
2012-02-01
Microwave plasma ion source with rectangular cavity resonator has been examined to improve ion beam current by changing wave launcher type from single-port to double-port. The cavity resonators with double-port and single-port wave launchers are designed to get resonance effect at TE-103 mode and TE-102 mode, respectively. In order to confirm that the cavities are acting as resonator, the microwave power for breakdown is measured and compared with the E-field strength estimated from the HFSS (High Frequency Structure Simulator) simulation. Langmuir probe measurements show that double-port cavity enhances central density of plasma ion source by modifying non-uniform plasma density profile of the single-port cavity. Correspondingly, beam current from the plasma ion source utilizing the double-port resonator is measured to be higher than that utilizing single-port resonator. Moreover, the enhancement in plasma density and ion beam current utilizing the double-port resonator is more pronounced as higher microwave power applied to the plasma ion source. Therefore, the rectangular cavity resonator utilizing the double-port is expected to enhance the performance of plasma ion source in terms of ion beam extraction.
NASA Technical Reports Server (NTRS)
Wang, H. T.
1979-01-01
Three kinds of frequency measuring systems are described: frequency comparison, phase comparison, and time comparison. With the help of the portable cesium clock in determining the time delay between two stations, a time synchronization, experiment was conducted using the Symphonie satellite. A result with an accuracy of 30 ns and an uncertainty of about 10 ns was obtained. Another experiment, applying the television pulse technique for time synchronization, yielded a result with an error of about 0.5 mu s in 24 hours. In order to measure the short term frequency stability of crystal oscillators or other frequency sources, a rubidium maser atomic frequency standard was developed as well as a short term stability measuring system.
NASA Technical Reports Server (NTRS)
Osaka, K.; Murata, T.; Okamoto, S.; Ohta, T.; Ozaki, T.; Maeda, T.; Mori, K.; Handa, H.; Matsumoto, S.; Sakaguchi, I.
1982-01-01
A completely implantable intracranial pressure sensor designed for long-term measurement of intraventricular pressure in hydrocephalic patients is described. The measurement principal of the device is discussed along with the electronic and component structure and sources of instrument error. Clinical tests of this implanted pressure device involving both humans and animals showed it to be comparable to other methods of intracranial pressure measurement.
Two-micron Laser Atmospheric Wind Sounder (LAWS) pointing/tracking study
NASA Technical Reports Server (NTRS)
Manlief, Scott
1995-01-01
The objective of the study was to identify and model major sources of short-term pointing jitter for a free-flying, full performance 2 micron LAWS system and evaluate the impact of the short-term jitter on wind-measurement performance. A fast steering mirror controls system was designed for the short-term jitter compensation. The performance analysis showed that the short-term jitter performance of the controls system over the 5.2 msec round-trip time for a realistic spacecraft environment was = 0.3 micro rad, rms, within the specified value of less than 0.5 micro rad, rms, derived in a 2 micron LAWS System Study. Disturbance modes were defined for: (1) the Bearing and Power Transfer Assembly (BAPTA) scan bearing, (2) the spacecraft reaction wheel torques, and (3) the solar array drive torques. The scan bearing disturbance was found to be the greatest contributing noise source to the jitter performance. Disturbances from the fast steering mirror reaction torques and a boom-mounted cross-link antenna clocking were also considered but were judged to be small compared to the three principal disturbance sources above and were not included in the final controls analysis.
Land surface temperature measurements from EOS MODIS data
NASA Technical Reports Server (NTRS)
Wan, Zhengming
1993-01-01
The task objectives of this reporting phase included: (1) completing the draft of the LST Algorithms Theoretical Basic Document by July 30, 1993; (2) making a detailed characterization of the thermal infrared measurement system including spectrometer, blackbody, and radiation sources; (3) making TIR spectral measurements of water and snow-cover surfaces with the MIDAC M2401 spectrometer; and (4) making conceptual and engineering design of an accessory system for spectrometric measurements at variable angles. These objectives are based on the requirements by the MODIS Science Team and the unique challenge in the development of MODIS LST algorithms: to acquire accurate spectral emissivity data of land covers in the near-term and to make ground validations of the LST product in the long-term with a TIR measurement system.
A Novel Airborne Carbon Isotope Analyzer for Methane and Carbon Dioxide Source Fingerprinting
NASA Astrophysics Data System (ADS)
Berman, E. S.; Huang, Y. W.; Owano, T. G.; Leifer, I.
2014-12-01
Recent field studies on major sources of the important greenhouse gas methane (CH4) indicate significant underestimation of methane release from fossil fuel industrial (FFI) and animal husbandry sources, among others. In addition, uncertainties still exist with respect to carbon dioxide (CO2) measurements, especially source fingerprinting. CO2 isotopic analysis provides a valuable in situ measurement approach to fingerprint CH4 and CO2as associated with combustion sources, leakage from geologic reservoirs, or biogenic sources. As a result, these measurements can characterize strong combustion source plumes, such as power plant emissions, and discriminate these emissions from other sources. As part of the COMEX (CO2 and MEthane eXperiment) campaign, a novel CO2 isotopic analyzer was installed and collected data aboard the CIRPAS Twin Otter aircraft. Developing methods to derive CH4 and CO2 budgets from remote sensing data is the goal of the summer 2014 COMEX campaign, which combines hyperspectral imaging (HSI) and non-imaging spectroscopy (NIS) with in situ airborne and surface data. COMEX leverages the synergy between high spatial resolution HSI and moderate spatial resolution NIS. The carbon dioxide isotope analyzer developed by Los Gatos Research (LGR) uses LGR's patented Off-Axis ICOS (Integrated Cavity Output Spectroscopy) technology and incorporates proprietary internal thermal control for high sensitivity and optimal instrument stability. This analyzer measures CO2 concentration as well as δ13C, δ18O, and δ17O from CO2 at natural abundance (100-3000 ppm). The laboratory accuracy is ±1.2 ppm (1σ) in CO2 from 370-1000 ppm, with a long-term (1000 s) precision of ±0.012 ppm. The long-term precision for both δ13C and δ18O is 0.04 ‰, and for δ17O is 0.06 ‰. The analyzer was field-tested as part of the COWGAS campaign, a pre-cursor campaign to COMEX in March 2014, where it successfully discriminated plumes related to combustion processes associated with dairy activities (tractor exhaust) from plumes and sources in air enriched in methane and ammonia from bovine activities including waste maintenance. Methodology, laboratory data, field data from COWGAS, and field data from the COMEX campaign acquired by LGR's carbon isotope analyzer as well as other COMEX analyzers are presented.
Multi-Scale Analysis of Trends in Northeastern Temperate Forest Springtime Phenology
NASA Astrophysics Data System (ADS)
Moon, M.; Melaas, E. K.; Sulla-menashe, D. J.; Friedl, M. A.
2017-12-01
The timing of spring leaf emergence is highly variable in many ecosystems, exerts first-order control growing season length, and significantly modulates seasonally-integrated photosynthesis. Numerous studies have reported trends toward earlier spring phenology in temperate forests, with some papers indicating that this trend is also leading to increased carbon uptake. At broad spatial scales, however, most of these studies have used data from coarse spatial resolution instruments such as MODIS, which does not resolve ecologically important landscape-scale patterns in phenology. In this work, we examine how long-term trends in spring phenology differ across three data sources acquired at different scales of measurements at the Harvard Forest in central Massachusetts. Specifically, we compared trends in the timing of phenology based on long-term in-situ measurements of phenology, estimates based on eddy-covariance measurements of net carbon uptake transition dates, and from two sources of satellite-based remote sensing (MODIS and Landsat) land surface phenology (LSP) data. Our analysis focused on the flux footprint surrounding the Harvard Forest Environmental Measurements (EMS) tower. Our results reveal clearly defined trends toward earlier springtime phenology in Landsat LSP and in the timing of tower-based net carbon uptake. However, we find no statistically significant trend in springtime phenology measured from MODIS LSP data products, possibly because the time series of MODIS observations is relatively short (13 years). The trend in tower-based transition data exhibited a larger negative value than the trend derived from Landsat LSP data (-0.42 and -0.28 days per year for 21 and 28 years, respectively). More importantly, these results have two key implications regarding how changes in spring phenology are impacting carbon uptake at landscape-scale. First, long-term trends in spring phenology can be quite different, depending on what data source is used to estimate the trend, and 2) the response of carbon uptake to climate change may be more sensitive than the response of land surface phenology itself.
Natural convection in symmetrically heated vertical parallel plates with discrete heat sources
DOE Office of Scientific and Technical Information (OSTI.GOV)
Manca, O.; Nardini, S.; Naso, V.
Laminar air natural convection in a symmetrically heated vertical channel with uniform flush-mounted discrete heat sources has been experimentally investigated. The effects of heated strips location and of their number are pointed out in terms of the maximum wall temperatures. A flow visualization in the entrance region of the channel was carried out and air temperatures and velocities in two cross sections have been measured. Dimensionless local heat transfer coefficients have been evaluated and monomial correlations among relevant parameters have bee derived in the local Rayleigh number range 10--10{sup 6}. Channel Nusselt number has been correlated in a polynomial formmore » in terms of channel Rayleigh number.« less
Long-Term Temporal Trends of Polychlorinated Biphenyls and Their Controlling Sources in China.
Zhao, Shizhen; Breivik, Knut; Liu, Guorui; Zheng, Minghui; Jones, Kevin C; Sweetman, Andrew J
2017-03-07
Polychlorinated biphenyls (PCBs) are industrial organic contaminants identified as persistent, bioaccumulative, toxic (PBT), and subject to long-range transport (LRT) with global scale significance. This study focuses on a reconstruction and prediction for China of long-term emission trends of intentionally and unintentionally produced (UP) ∑ 7 PCBs (UP-PCBs, from the manufacture of steel, cement and sinter iron) and their re-emissions from secondary sources (e.g., soils and vegetation) using a dynamic fate model (BETR-Global). Contemporary emission estimates combined with predictions from the multimedia fate model suggest that primary sources still dominate, although unintentional sources are predicted to become a main contributor from 2035 for PCB-28. Imported e-waste is predicted to play an increasing role until 2020-2030 on a national scale due to the decline of intentionally produced (IP) emissions. Hypothetical emission scenarios suggest that China could become a potential source to neighboring regions with a net output of ∼0.4 t year -1 by around 2050. However, future emission scenarios and hence model results will be dictated by the efficiency of control measures.
Exploiting New Data Sources to Quantify Arterial Congestion and Performance Measures.
DOT National Transportation Integrated Search
2017-01-01
Transit travel time, operating speed and reliability all influence service attractiveness, operating cost and system efficiency. These metrics have a long-term impact on system effectiveness through a change in ridership. As part of its bus dispatch ...
Current Situation for Management of Disused Sealed Radioactive Sources in Japan - 13025
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kusama, Keiji; Miyamoto, Yoichi
2013-07-01
As for the Sealed Radioactive Source currently used in Japan, many of them are imported from overseas. The U.S., Canada, Germany, the Netherlands, Belgium and Czech Republic are the main exporting States. Many of disused sealed radioactive sources are being returned to exporting States. The sealed radioactive sources which cannot be returned to exporting States are appropriately kept in the domestic storage facility. So, there are not main problem on the long term management of disused sealed radioactive sources in Japan. However, there are some difficulties on repatriate. One is reservation of a means of transport. The sea mail whichmore » conveys radioactive sources owing to reduction of movement of international cargo is decreasing in number. And there is a denial of shipment. Other one is that the manufacturer has already resigned from the work and cannot return disused sealed radioactive sources, or a manufacturer cannot specify and disused sources cannot be returned. The disused sealed radioactive source which cannot be repatriated is a little in term of radioactivity. As for the establishment of national measure of final disposal facility for disused sealed radioactive sources, in Japan, it is not yet installed with difficulty. Since there are many countries for which installation of a final disposal facility for disused sealed radioactive sources is difficult, the source manufacture country should respond positively to return the source which was manufactured and sold in the past. (authors)« less
NASA Astrophysics Data System (ADS)
Flassak, Thomas; de Witt, Helmut; Hahnfeld, Peter; Knaup, Andreas; Kramer, Lothar
1995-09-01
COMPAS is a decision support system designed to assist in the assessment of the consequences of accidental releases of toxic and flammable substances. One of the key elements of COMPAS is a feedback algorithm which allows us to calculate the source term with the aid of concentration measurements. Up to now the feedback technique is applied to concentration measurements done with test tubes or conventional point sensors. In this paper the extension of the actual method is presented which is the combination of COMPAS and an optical remote sensing system like the KAYSER-THREDE K300 FTIR system. Active remote sensing methods based on FTIR are, among other applications, ideal for the so-called fence line monitoring of the diffuse emissions and accidental releases from industrial facilities, since from the FTIR spectra averaged concentration levels along the measurement path can be achieved. The line-averaged concentrations are ideally suited as on-line input for COMPAS' feedback technique. Uncertainties in the assessment of the source term related with both shortcomings of the dispersion model itself and also problems of a feedback strategy based on point measurements are reduced.
An overview of applied visibility fundamentals: Survey and synthesis of a visibility literature
NASA Astrophysics Data System (ADS)
1982-06-01
The concept of visibility has been defined in numerous ways. One source that is widely used in the United States describes visibility as the greatest distance at which it is just possible to see and identify selected objects or light sources with the unaided eye, or, it may be considered the distance an observer would have to back away from a target before it disappears. The process by which we see distant objects is based on the characteristics of the target, its immediate surroundings, air quality, illumination of the sight path, and the eyes and brain of the observer. Thus, four components of visibility have been noted: (1) target, (2) background, (3) atmosphere, and (4) observer. Most definitions of visibility include the observer as a key element, but it is also possible to measure visibility without the immediate presence of humans by means of instruments such as telephotometers, transmissometers, or nephelometers. Such measurements can be kept in non-dimensional terms and can be compared without human intervention. When used in the field, however, visibility sensors must be calibrated in human terms of visibility measurement.
VLF Source Localization with a Freely Drifting Sensor Array
1992-09-01
Simultaneous Measurement of Infra - sonic Acoustic Particle Velocity and Acoustic Pressure in the Ocean by F-ely Drifting Swallow Floats," IEEEJ. Ocean. Eng., vol...Pacific. Marine Physical Laboratory’s set of nine freely drifting, infrasonic sensors, capable of recording ocean ambient noise in the 1- to 25-Hz range...Terms. 15. Number of Pages, Swallow float, matched-field processing, infrasonic sensor, vlf source localization 153 16. Price Code. 17. Seorlity
Novel techniques for characterization of hydrocarbon emission sources in the Barnett Shale
NASA Astrophysics Data System (ADS)
Nathan, Brian Joseph
Changes in ambient atmospheric hydrocarbon concentrations can have both short-term and long-term effects on the atmosphere and on human health. Thus, accurate characterization of emissions sources is critically important. The recent boom in shale gas production has led to an increase in hydrocarbon emissions from associated processes, though the exact extent is uncertain. As an original quantification technique, a model airplane equipped with a specially-designed, open-path methane sensor was flown multiple times over a natural gas compressor station in the Barnett Shale in October 2013. A linear optimization was introduced to a standard Gaussian plume model in an effort to determine the most probable emission rate coming from the station. This is shown to be a suitable approach given an ideal source with a single, central plume. Separately, an analysis was performed to characterize the nonmethane hydrocarbons in the Barnett during the same period. Starting with ambient hourly concentration measurements of forty-six hydrocarbon species, Lagrangian air parcel trajectories were implemented in a meteorological model to extend the resolution of these measurements and achieve domain-fillings of the region for the period of interest. A self-organizing map (a type of unsupervised classification) was then utilized to reduce the dimensionality of the total multivariate set of grids into characteristic one-dimensional signatures. By also introducing a self-organizing map classification of the contemporary wind measurements, the spatial hydrocarbon characterizations are analyzed for periods with similar wind conditions. The accuracy of the classification is verified through assessment of observed spatial mixing ratio enhancements of key species, through site-comparisons with a related long-term study, and through a random forest analysis (an ensemble learning method of supervised classification) to determine the most important species for defining key classes. The hydrocarbon classification is shown to have performed very well in identifying expected signatures near and downwind-of oil and gas facilities with active permits, which showcases this method's usefulness for future regional hydrocarbon source-apportionment analyses.
Mapping water availability, projected use and cost in the western United States
NASA Astrophysics Data System (ADS)
Tidwell, Vincent C.; Moreland, Barbara D.; Zemlick, Katie M.; Roberts, Barry L.; Passell, Howard D.; Jensen, Daniel; Forsgren, Christopher; Sehlke, Gerald; Cook, Margaret A.; King, Carey W.; Larsen, Sara
2014-05-01
New demands for water can be satisfied through a variety of source options. In some basins surface and/or groundwater may be available through permitting with the state water management agency (termed unappropriated water), alternatively water might be purchased and transferred out of its current use to another (termed appropriated water), or non-traditional water sources can be captured and treated (e.g., wastewater). The relative availability and cost of each source are key factors in the development decision. Unfortunately, these measures are location dependent with no consistent or comparable set of data available for evaluating competing water sources. With the help of western water managers, water availability was mapped for over 1200 watersheds throughout the western US. Five water sources were individually examined, including unappropriated surface water, unappropriated groundwater, appropriated water, municipal wastewater and brackish groundwater. Also mapped was projected change in consumptive water use from 2010 to 2030. Associated costs to acquire, convey and treat the water, as necessary, for each of the five sources were estimated. These metrics were developed to support regional water planning and policy analysis with initial application to electric transmission planning in the western US.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lawrie, Scott R., E-mail: scott.lawrie@stfc.ac.uk; John Adams Institute for Accelerator Science, Department of Physics, University of Oxford; Faircloth, Daniel C.
2015-04-08
In order to facilitate the testing of advanced H{sup −} ion sources for the ISIS and Front End Test Stand (FETS) facilities at the Rutherford Appleton Laboratory (RAL), a Vessel for Extraction and Source Plasma Analyses (VESPA) has been constructed. This will perform the first detailed plasma measurements on the ISIS Penning-type H{sup −} ion source using emission spectroscopic techniques. In addition, the 30-year-old extraction optics are re-designed from the ground up in order to fully transport the beam. Using multiple beam and plasma diagnostics devices, the ultimate aim is improve H{sup −} production efficiency and subsequent transport for eithermore » long-term ISIS user operations or high power FETS requirements. The VESPA will also accommodate and test a new scaled-up Penning H{sup −} source design. This paper details the VESPA design, construction and commissioning, as well as initial beam and spectroscopy results.« less
Jiao, Shuliang; Todorović, Milos; Stoica, George; Wang, Lihong V
2005-09-10
We report on a new configuration of fiber-based polarization-sensitive Mueller matrix optical coherence tomography that permits the acquisition of the round-trip Jones matrix of a biological sample using only one light source and a single depth scan. In this new configuration, a polarization modulator is used in the source arm to continuously modulate the incident polarization state for both the reference and the sample arms. The Jones matrix of the sample can be calculated from the two frequency terms in the two detection channels. The first term is modulated by the carrier frequency, which is determined by the longitudinal scanning mechanism, whereas the other term is modulated by the beat frequency between the carrier frequency and the second harmonic of the modulation frequency of the polarization modulator. One important feature of this system is that, for the first time to our knowledge, the Jones matrix of the sample can be calculated with a single detection channel and a single measurement when diattenuation is negligible. The system was successfully tested by imaging both standard polarization elements and biological samples.
NASA Technical Reports Server (NTRS)
Karchmer, A. M.
1977-01-01
Fluctuating pressure measurements within the combustor and tailpipe of a turbofan engine are made simultaneously with far field acoustic measurements. The pressure measurements within the engine are accomplished with cooled semi-infinite waveguide probes utilizing conventional condenser microphones as the transducers. The measurements are taken over a broad range of engine operating conditions and for 16 far field microphone positions between 10 deg and 160 deg relative to the engine inlet axis. Correlation and coherence techniques are used to determine the relative phase and amplitude relationships between the internal pressures and far field acoustic pressures. The results indicate that the combustor is a low frequency source region for acoustic propagation through the tailpipe and out to the far field. Specifically, it is found that the relation between source pressure and the resulting sound pressure involves a 180 deg phase shift. The latter result is obtained by Fourier transforming the cross correlation function between the source pressure and acoustic pressure after removing the propagation delay time. Further, it is found that the transfer function between the source pressure and acoustic pressure has a magnitude approximately proportional to frequency squared. These results are shown to be consistent with a model using a modified source term in Lighthill's turbulence stress tensor, wherein the fluctuating Reynolds stresses are replaced with the pressure fluctuations due to fluctuating entropy.
New Experiments with Antiprotons
NASA Astrophysics Data System (ADS)
Kaplan, D. M.
2011-12-01
Fermilab operates the world's most intense antiproton source. Recently proposed experiments can use those antiprotons either parasitically during Teva-tron Collider running or after the Tevatron Collider finishes in about 2011. For example, the annihilation of 8 GeV antiprotons might make the world's most intense source of tagged D0 mesons, and thus the best near-term opportunity to study charm mixing and search for new physics via its CP-violation signature. Other possible precision measurements include properties of the X(3872) and the charmonium system. An experiment using a Penning trap and an atom interferometer could make the world's first measurement of the gravitational force on antimatter. These and other potential measurements using antiprotons could yield a broad physics program at Fermilab in the post-Tevatron era.
Acoustic device and method for measuring gas densities
NASA Technical Reports Server (NTRS)
Shakkottai, Parthasarathy (Inventor); Kwack, Eug Y. (Inventor); Back, Lloyd (Inventor)
1992-01-01
Density measurements can be made in a gas contained in a flow through enclosure by measuring the sound pressure level at a receiver or microphone located near a dipole sound source which is driven at constant velocity amplitude at low frequencies. Analytical results, which are provided in terms of geometrical parameters, wave numbers, and sound source type for systems of this invention, agree well with published data. The relatively simple designs feature a transmitter transducer at the closed end of a small tube and a receiver transducer on the circumference of the small tube located a small distance away from the transmitter. The transmitter should be a dipole operated at low frequency with the kL value preferable less that about 0.3.
Mainhagu, Jon; Morrison, C.; Truex, Michael J.; ...
2014-08-05
A method termed vapor-phase tomography has recently been proposed to characterize the distribution of volatile organic contaminant mass in vadose-zone source areas, and to measure associated three-dimensional distributions of local contaminant mass discharge. The method is based on measuring the spatial variability of vapor flux, and thus inherent to its effectiveness is the premise that the magnitudes and temporal variability of vapor concentrations measured at different monitoring points within the interrogated area will be a function of the geospatial positions of the points relative to the source location. A series of flow-cell experiments was conducted to evaluate this premise. Amore » well-defined source zone was created by injection and extraction of a non-reactive gas (SF6). Spatial and temporal concentration distributions obtained from the tests were compared to simulations produced with a mathematical model describing advective and diffusive transport. Tests were conducted to characterize both areal and vertical components of the application. Decreases in concentration over time were observed for monitoring points located on the opposite side of the source zone from the local–extraction point, whereas increases were observed for monitoring points located between the local–extraction point and the source zone. We found that the results illustrate that comparison of temporal concentration profiles obtained at various monitoring points gives a general indication of the source location with respect to the extraction and monitoring points.« less
Intercomparison of Open-Path Trace Gas Measurements with Two Dual Frequency Comb Spectrometers
Waxman, Eleanor M.; Cossel, Kevin C.; Truong, Gar-Wing; Giorgetta, Fabrizio R.; Swann, William C.; Coburn, Sean; Wright, Robert J.; Rieker, Gregory B.; Coddington, Ian; Newbury, Nathan R.
2017-01-01
We present the first quantitative intercomparison between two open-path dual comb spectroscopy (DCS) instruments which were operated across adjacent 2-km open-air paths over a two-week period. We used DCS to measure the atmospheric absorption spectrum in the near infrared from 6021 to 6388 cm−1 (1565 to 1661 nm), corresponding to a 367 cm−1 bandwidth, at 0.0067 cm−1 sample spacing. The measured absorption spectra agree with each other to within 5×10−4 without any external calibration of either instrument. The absorption spectra are fit to retrieve concentrations for carbon dioxide (CO2), methane (CH4), water (H2O), and deuterated water (HDO). The retrieved dry mole fractions agree to 0.14% (0.57 ppm) for CO2, 0.35% (7 ppb) for CH4, and 0.40% (36 ppm) for H2O over the two-week measurement campaign, which included 23 °C outdoor temperature variations and periods of strong atmospheric turbulence. This agreement is at least an order of magnitude better than conventional active-source open-path instrument intercomparisons and is particularly relevant to future regional flux measurements as it allows accurate comparisons of open-path DCS data across locations and time. We additionally compare the open-path DCS retrievals to a WMO-calibrated cavity ringdown point sensor located along the path with good agreement. Short-term and long-term differences between the two systems are attributed, respectively, to spatial sampling discrepancies and to inaccuracies in the current spectral database used to fit the DCS data. Finally, the two-week measurement campaign yields diurnal cycles of CO2 and CH4 that are consistent with the presence of local sources of CO2 and absence of local sources of CH4. PMID:29276547
Solute source depletion control of forward and back diffusion through low-permeability zones
NASA Astrophysics Data System (ADS)
Yang, Minjune; Annable, Michael D.; Jawitz, James W.
2016-10-01
Solute diffusive exchange between low-permeability aquitards and high-permeability aquifers acts as a significant mediator of long-term contaminant fate. Aquifer contaminants diffuse into aquitards, but as contaminant sources are depleted, aquifer concentrations decline, triggering back diffusion from aquitards. The dynamics of the contaminant source depletion, or the source strength function, controls the timing of the transition of aquitards from sinks to sources. Here, we experimentally evaluate three archetypical transient source depletion models (step-change, linear, and exponential), and we use novel analytical solutions to accurately account for dynamic aquitard-aquifer diffusive transfer. Laboratory diffusion experiments were conducted using a well-controlled flow chamber to assess solute exchange between sand aquifer and kaolinite aquitard layers. Solute concentration profiles in the aquitard were measured in situ using electrical conductivity. Back diffusion was shown to begin earlier and produce larger mass flux for rapidly depleting sources. The analytical models showed very good correspondence with measured aquifer breakthrough curves and aquitard concentration profiles. The modeling approach links source dissolution and back diffusion, enabling assessment of human exposure risk and calculation of the back diffusion initiation time, as well as the resulting plume persistence.
Solute source depletion control of forward and back diffusion through low-permeability zones.
Yang, Minjune; Annable, Michael D; Jawitz, James W
2016-10-01
Solute diffusive exchange between low-permeability aquitards and high-permeability aquifers acts as a significant mediator of long-term contaminant fate. Aquifer contaminants diffuse into aquitards, but as contaminant sources are depleted, aquifer concentrations decline, triggering back diffusion from aquitards. The dynamics of the contaminant source depletion, or the source strength function, controls the timing of the transition of aquitards from sinks to sources. Here, we experimentally evaluate three archetypical transient source depletion models (step-change, linear, and exponential), and we use novel analytical solutions to accurately account for dynamic aquitard-aquifer diffusive transfer. Laboratory diffusion experiments were conducted using a well-controlled flow chamber to assess solute exchange between sand aquifer and kaolinite aquitard layers. Solute concentration profiles in the aquitard were measured in situ using electrical conductivity. Back diffusion was shown to begin earlier and produce larger mass flux for rapidly depleting sources. The analytical models showed very good correspondence with measured aquifer breakthrough curves and aquitard concentration profiles. The modeling approach links source dissolution and back diffusion, enabling assessment of human exposure risk and calculation of the back diffusion initiation time, as well as the resulting plume persistence. Copyright © 2016 Elsevier B.V. All rights reserved.
Fermi Large Area Telescope Second Source Catalog
NASA Technical Reports Server (NTRS)
Nolan, P. L.; Abdo, A. A.; Ackermann, M.; Ajello, M; Allafort, A.; Antolini, E; Bonnell, J.; Cannon, A.; Celik O.; Corbet, R.;
2012-01-01
We present the second catalog of high-energy gamma-ray sources detected by the Large Area Telescope (LAT), the primary science instrument on the Fermi Gamma-ray Space Telescope (Fermi), derived from data taken during the first 24 months of the science phase of the mission, which began on 2008 August 4. Source detection is based on the average flux over the 24-month period. The Second Fermi-LAT catalog (2FGL) includes source location regions, defined in terms of elliptical fits to the 95% confidence regions and spectral fits in terms of power-law, exponentially cutoff power-law, or log-normal forms. Also included are flux measurements in 5 energy bands and light curves on monthly intervals for each source. Twelve sources in the catalog are modeled as spatially extended. We provide a detailed comparison of the results from this catalog with those from the first Fermi-LAT catalog (1FGL). Although the diffuse Galactic and isotropic models used in the 2FGL analysis are improved compared to the 1FGL catalog, we attach caution flags to 162 of the sources to indicate possible confusion with residual imperfections in the diffuse model. The 2FGL catalog contains 1873 sources detected and characterized in the 100 11eV to 100 GeV range of which we consider 127 as being firmly identified and 1171 as being reliably associated with counterparts of known or likely gamma-ray-producing source classes.
NASA Astrophysics Data System (ADS)
Hu, Minpeng; Liu, Yanmei; Wang, Jiahui; Dahlgren, Randy A.; Chen, Dingjiang
2018-06-01
Source apportionment is critical for guiding development of efficient watershed nitrogen (N) pollution control measures. The ReNuMa (Regional Nutrient Management) model, a semi-empirical, semi-process-oriented model with modest data requirements, has been widely used for riverine N source apportionment. However, the ReNuMa model contains limitations for addressing long-term N dynamics by ignoring temporal changes in atmospheric N deposition rates and N-leaching lag effects. This work modified the ReNuMa model by revising the source code to allow yearly changes in atmospheric N deposition and incorporation of N-leaching lag effects into N transport processes. The appropriate N-leaching lag time was determined from cross-correlation analysis between annual watershed individual N source inputs and riverine N export. Accuracy of the modified ReNuMa model was demonstrated through analysis of a 31-year water quality record (1980-2010) from the Yongan watershed in eastern China. The revisions considerably improved the accuracy (Nash-Sutcliff coefficient increased by ∼0.2) of the modified ReNuMa model for predicting riverine N loads. The modified model explicitly identified annual and seasonal changes in contributions of various N sources (i.e., point vs. nonpoint source, surface runoff vs. groundwater) to riverine N loads as well as the fate of watershed anthropogenic N inputs. Model results were consistent with previously modeled or observed lag time length as well as changes in riverine chloride and nitrate concentrations during the low-flow regime and available N levels in agricultural soils of this watershed. The modified ReNuMa model is applicable for addressing long-term changes in riverine N sources, providing decision-makers with critical information for guiding watershed N pollution control strategies.
NASA Astrophysics Data System (ADS)
Di Sieno, L.; Contini, D.; Dalla Mora, A.; Torricelli, A.; Spinelli, L.; Cubeddu, R.; Tosi, A.; Boso, G.; Pifferi, A.
2013-06-01
In this article, we show experimental results of time-resolved optical spectroscopy performed with small distance between launching and detecting fibers. It was already demonstrated that depth discrimination is independent of source-detector separation and that measurements at small source detector distance provide better contrast and spatial resolution. The main disadvantage is represent by the huge increase in early photons (scarcely diffused by tissue) peak that can saturate the dynamic range of most detectors, hiding information carried by late photons. Thanks to a fast-gated Single- Photon Avalanche Diode (SPAD) module, we are able to reject the peak of early photons and to obtain high-dynamic range acquisitions. We exploit fast-gated SPAD module to perform for the first time functional near-infrared spectroscopy (fNIRS) at small source-detector distance for in vivo measurements and we demonstrate the possibility to detect non-invasively the dynamics of oxygenated and deoxygenated haemoglobin occurring in the motor cortex during a motor task. We also show the improvement in terms of signal amplitude and Signal-to-Noise Ratio (SNR) obtained exploiting fast-gated SPAD performances with respect to "non-gated" measurements.
Bayesian source term estimation of atmospheric releases in urban areas using LES approach.
Xue, Fei; Kikumoto, Hideki; Li, Xiaofeng; Ooka, Ryozo
2018-05-05
The estimation of source information from limited measurements of a sensor network is a challenging inverse problem, which can be viewed as an assimilation process of the observed concentration data and the predicted concentration data. When dealing with releases in built-up areas, the predicted data are generally obtained by the Reynolds-averaged Navier-Stokes (RANS) equations, which yields building-resolving results; however, RANS-based models are outperformed by large-eddy simulation (LES) in the predictions of both airflow and dispersion. Therefore, it is important to explore the possibility of improving the estimation of the source parameters by using the LES approach. In this paper, a novel source term estimation method is proposed based on LES approach using Bayesian inference. The source-receptor relationship is obtained by solving the adjoint equations constructed using the time-averaged flow field simulated by the LES approach based on the gradient diffusion hypothesis. A wind tunnel experiment with a constant point source downwind of a single building model is used to evaluate the performance of the proposed method, which is compared with that of the existing method using a RANS model. The results show that the proposed method reduces the errors of source location and releasing strength by 77% and 28%, respectively. Copyright © 2018 Elsevier B.V. All rights reserved.
Pollution monitoring using networks of honey bees
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bromenshenk, J.J.; Dewart, M.L.; Thomas, J.M.
1983-08-01
Each year thousands of chemicals in large quantities are introduced into the global environment and the need for effective methods of monitoring these substances has steadily increased. Most monitoring programs rely upon instrumentation to measure specific contaminants in air, water, or soil. However, it has become apparent that humans and their environment are exposed to complex mixtures of chemicals rather than single entities. As our ability to detect ever smaller quantities of pollutants has increased, the biological significance of these findings has become more uncertain. Also, it is clear that monitoring efforts should shift from short-term studies of easily identifiablemore » sources in localized areas to long-term studies of multiple sources over widespread regions. Our investigations aim at providing better tools to meet these exigencies. Honey bees are discussed as an effective, long-term, self-sustaining system for monitoring environmental impacts. Our results indicate that the use of regional, and possibly national or international, capability can be realized with the aid of beekeepers in obtaining samples and conducting measurements. This approach has the added advantage of public involvement in environmental problem solving and protection of human health and environmental quality.« less
NASA Astrophysics Data System (ADS)
Sridhara, Basavapatna Sitaramaiah
In an internal combustion engine, the engine is the noise source and the exhaust pipe is the main transmitter of noise. Mufflers are often used to reduce engine noise level in the exhaust pipe. To optimize a muffler design, a series of experiments could be conducted using various mufflers installed in the exhaust pipe. For each configuration, the radiated sound pressure could be measured. However, this is not a very efficient method. A second approach would be to develop a scheme involving only a few measurements which can predict the radiated sound pressure at a specified distance from the open end of the exhaust pipe. In this work, the engine exhaust system was modelled as a lumped source-muffler-termination system. An expression for the predicted sound pressure level was derived in terms of the source and termination impedances, and the muffler geometry. The pressure source and monopole radiation models were used for the source and the open end of the exhaust pipe. The four pole parameters were used to relate the acoustic properties at two different cross sections of the muffler and the pipe. The developed formulation was verified through a series of experiments. Two loudspeakers and a reciprocating type vacuum pump were used as sound sources during the tests. The source impedance was measured using the direct, two-load and four-load methods. A simple expansion chamber and a side-branch resonator were used as mufflers. Sound pressure level measurements for the prediction scheme were made for several source-muffler and source-straight pipe combinations. The predicted and measured sound pressure levels were compared for all cases considered. In all cases, correlation of the experimental results and those predicted by the developed expressions was good. Predicted and measured values of the insertion loss of the mufflers were compared. The agreement between the two was good. Also, an error analysis of the four-load method was done.
Audio visual speech source separation via improved context dependent association model
NASA Astrophysics Data System (ADS)
Kazemi, Alireza; Boostani, Reza; Sobhanmanesh, Fariborz
2014-12-01
In this paper, we exploit the non-linear relation between a speech source and its associated lip video as a source of extra information to propose an improved audio-visual speech source separation (AVSS) algorithm. The audio-visual association is modeled using a neural associator which estimates the visual lip parameters from a temporal context of acoustic observation frames. We define an objective function based on mean square error (MSE) measure between estimated and target visual parameters. This function is minimized for estimation of the de-mixing vector/filters to separate the relevant source from linear instantaneous or time-domain convolutive mixtures. We have also proposed a hybrid criterion which uses AV coherency together with kurtosis as a non-Gaussianity measure. Experimental results are presented and compared in terms of visually relevant speech detection accuracy and output signal-to-interference ratio (SIR) of source separation. The suggested audio-visual model significantly improves relevant speech classification accuracy compared to existing GMM-based model and the proposed AVSS algorithm improves the speech separation quality compared to reference ICA- and AVSS-based methods.
Uras, Yusuf; Uysal, Yagmur; Arikan, Tugba Atilan; Kop, Alican; Caliskan, Mustafa
2015-06-01
The aim of this study was to investigate the sources of drinking water for Derebogazi Village, Kahramanmaras Province, Turkey, in terms of hydrogeochemistry, isotope geochemistry, and medical geology. Water samples were obtained from seven different water sources in the area, all of which are located within quartzite units of Paleozoic age, and isotopic analyses of (18)O and (2)H (deuterium) were conducted on the samples. Samples were collected from the region for 1 year. Water quality of the samples was assessed in terms of various water quality parameters, such as temperature, pH, conductivity, alkalinity, trace element concentrations, anion-cation measurements, and metal concentrations, using ion chromatography, inductively coupled plasma (ICP) mass spectrometry, ICP-optical emission spectrometry techniques. Regional health surveys had revealed that the heights of local people are significantly below the average for the country. In terms of medical geology, the sampled drinking water from the seven sources was deficient in calcium and magnesium ions, which promote bone development. Bone mineral density screening tests were conducted on ten females using dual energy X-ray absorptiometry to investigate possible developmental disorder(s) and potential for mineral loss in the region. Of these ten women, three had T-scores close to the osteoporosis range (T-score < -2.5).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schiefelbein, C.; Ho, T.
Changes in the physical properties (measured in terms of vitrinite reflectance, elemental analysis, and C-13 nuclear magnetic resonance) of an immature coal (0.46% R{sub o}) from Craig County, Colorado, that was thermally altered using hydrous pyrolysis were used to establish a correspondence between hydrous pyrolysis time/temperature reaction conditions and relative maturity (expressed in terms of vitrinite reflectance). This correspondence was used to determine the oil generation maturity limits for an immature hydrogen-rich (Type I fluorescing amorphous oil-prone kerogen) source rock from an offshore Congo well that was thermally altered using the same reaction conditions as applied to the immature coal.more » The resulting changes in the physical properties of the altered source rock, measured in terms of decreasing reactive carbon content (from Rock-Eval pyrolysis), were used to construct a hydrocarbon yield curve from which the relative maturity associated with the onset, main phase, and peak of oil generation was determined. Results, substantiated by anhydrous pyrolysis techniques, indicate that the source rock from Congo has a late onset of appreciable ({gt}10% transformation) oil generation (0.9% R{sub o} {plus minus} 0.1%), generates maximum quantities of oil from about 1.1 to 1.3% R{sub o}, and reaches the end (or peak) of the primary oil generating window at approximately 1.4% R{sub o} ({plus minus}0.1%) when secondary cracking reactions become important. However, the bottom of the oil window can be extended to about 1.6% R{sub o} because the heavy molecular weight degradation by-products (asphaltenes) that are not efficiently expelled from source rocks continue to degrade into progressively lower molecular weight hydrocarbons.« less
Suspended-sediment sources in an urban watershed, Northeast Branch Anacostia River, Maryland
Devereux, Olivia H.; Prestegaard, Karen L.; Needelman, Brian A.; Gellis, Allen C.
2010-01-01
Fine sediment sources were characterized by chemical composition in an urban watershed, the Northeast Branch Anacostia River, which drains to the Chesapeake Bay. Concentrations of 63 elements and two radionuclides were measured in possible land-based sediment sources and suspended sediment collected from the water column at the watershed outlet during storm events. These tracer concentrations were used to determine the relative quantity of suspended sediment contributed by each source. Although this is an urbanized watershed, there was not a distinct urban signature that can be evaluated except for the contributions from road surfaces. We identified the sources of fine sediment by both physiographic province (Piedmont and Coastal Plain) and source locale (streambanks, upland and street residue) by using different sets of elemental tracers. The Piedmont contributed the majority of the fine sediment for seven of the eight measured storms. The streambanks contributed the greatest quantity of fine sediment when evaluated by source locale. Street residue contributed 13% of the total suspended sediment on average and was the source most concentrated in anthropogenically enriched elements. Combining results from the source locale and physiographic province analyses, most fine sediment in the Northeast Branch watershed is derived from streambanks that contain sediment eroded from the Piedmont physiographic province of the watershed. Sediment fingerprinting analyses are most useful when longer term evaluations of sediment erosion and storage are also available from streambank-erosion measurements, sediment budget and other methods.
2014-10-01
de l’exactitude et de la précision), comparativement au modèle de mesure plus simple qui n’utilise pas de multiplicateurs. Importance pour la défense...3) Bayesian experimental design for receptor placement in order to maximize the expected information in the measured concen- tration data for...applications of the Bayesian inferential methodology for source recon- struction have used high-quality concentration data from well- designed atmospheric
Assessing the Gap Between Top-down and Bottom-up Measured Methane Emissions in Indianapolis, IN.
NASA Astrophysics Data System (ADS)
Prasad, K.; Lamb, B. K.; Cambaliza, M. O. L.; Shepson, P. B.; Stirm, B. H.; Salmon, O. E.; Lavoie, T. N.; Lauvaux, T.; Ferrara, T.; Howard, T.; Edburg, S. L.; Whetstone, J. R.
2014-12-01
Releases of methane (CH4) from the natural gas supply chain in the United States account for approximately 30% of the total US CH4 emissions. However, there continues to be large questions regarding the accuracy of current emission inventories for methane emissions from natural gas usage. In this paper, we describe results from top-down and bottom-up measurements of methane emissions from the large isolated city of Indianapolis. The top-down results are based on aircraft mass balance and tower based inverse modeling methods, while the bottom-up results are based on direct component sampling at metering and regulating stations, surface enclosure measurements of surveyed pipeline leaks, and tracer/modeling methods for other urban sources. Mobile mapping of methane urban concentrations was also used to identify significant sources and to show an urban-wide low level enhancement of methane levels. The residual difference between top-down and bottom-up measured emissions is large and cannot be fully explained in terms of the uncertainties in top-down and bottom-up emission measurements and estimates. Thus, the residual appears to be, at least partly, attributed to a significant wide-spread diffusive source. Analyses are included to estimate the size and nature of this diffusive source.
Frequency standard stability for Doppler measurements on-board the shuttle
NASA Technical Reports Server (NTRS)
Harton, P. L.
1974-01-01
The short and long term stability characteristics of crystal and atomic standards are described. Emphasis is placed on crystal oscillators because of the selection which was made for the shuttle baseline and the complexities which are introduced by the shuttle environment. Attention is given, first, to the definitions of stability and the application of these definitions to the shuttle system and its mission. Data from time domain measurements are used to illustrate the definitions. Results of a literature survey to determine environmental effects on frequency reference sources are then presented. Finally, methods of standard frequency dissemination over radio frequency carriers are noted as a possible means of measuring absolute accuracy and long term stability characteristics during on one way Doppler equipment.
Modeling the contribution of point sources and non-point sources to Thachin River water pollution.
Schaffner, Monika; Bader, Hans-Peter; Scheidegger, Ruth
2009-08-15
Major rivers in developing and emerging countries suffer increasingly of severe degradation of water quality. The current study uses a mathematical Material Flow Analysis (MMFA) as a complementary approach to address the degradation of river water quality due to nutrient pollution in the Thachin River Basin in Central Thailand. This paper gives an overview of the origins and flow paths of the various point- and non-point pollution sources in the Thachin River Basin (in terms of nitrogen and phosphorus) and quantifies their relative importance within the system. The key parameters influencing the main nutrient flows are determined and possible mitigation measures discussed. The results show that aquaculture (as a point source) and rice farming (as a non-point source) are the key nutrient sources in the Thachin River Basin. Other point sources such as pig farms, households and industries, which were previously cited as the most relevant pollution sources in terms of organic pollution, play less significant roles in comparison. This order of importance shifts when considering the model results for the provincial level. Crosschecks with secondary data and field studies confirm the plausibility of our simulations. Specific nutrient loads for the pollution sources are derived; these can be used for a first broad quantification of nutrient pollution in comparable river basins. Based on an identification of the sensitive model parameters, possible mitigation scenarios are determined and their potential to reduce the nutrient load evaluated. A comparison of simulated nutrient loads with measured nutrient concentrations shows that nutrient retention in the river system may be significant. Sedimentation in the slow flowing surface water network as well as nitrogen emission to the air from the warm oxygen deficient waters are certainly partly responsible, but also wetlands along the river banks could play an important role as nutrient sinks.
Brusseau, M. L.; Hatton, J.; DiGuiseppi, W.
2011-01-01
The long-term impact of source-zone remediation efforts was assessed for a large site contaminated by trichloroethene. The impact of the remediation efforts (soil vapor extraction and in-situ chemical oxidation) was assessed through analysis of plume-scale contaminant mass discharge, which was measured using a high-resolution data set obtained from 23 years of operation of a large pump-and-treat system. The initial contaminant mass discharge peaked at approximately 7 kg/d, and then declined to approximately 2 kg/d. This latter value was sustained for several years prior to the initiation of source-zone remediation efforts. The contaminant mass discharge in 2010, measured several years after completion of the two source-zone remediation actions, was approximately 0.2 kg/d, which is ten times lower than the value prior to source-zone remediation. The time-continuous contaminant mass discharge data can be used to evaluate the impact of the source-zone remediation efforts on reducing the time required to operate the pump-and-treat system, and to estimate the cost savings associated with the decreased operational period. While significant reductions have been achieved, it is evident that the remediation efforts have not completely eliminated contaminant mass discharge and associated risk. Remaining contaminant mass contributing to the current mass discharge is hypothesized to comprise poorly-accessible mass in the source zones, as well as aqueous (and sorbed) mass present in the extensive lower-permeability units located within and adjacent to the contaminant plume. The fate of these sources is an issue of critical import to the remediation of chlorinated-solvent contaminated sites, and development of methods to address these sources will be required to achieve successful long-term management of such sites and to ultimately transition them to closure. PMID:22115080
A routinely applied atmospheric dispersion model was modified to evaluate alternative modeling techniques which allowed for more detailed source data, onsite meteorological data, and several dispersion methodologies. These were evaluated with hourly SO2 concentrations measured at...
Clean the Air and Breathe Easier.
ERIC Educational Resources Information Center
Guevin, John
1997-01-01
Failure to prevent indoor air quality problems or act promptly can result in increased chances for long- or short-term health problems for staff and students, reduced productivity, faster plant deterioration, and strained school-community relations. Basic pollution control measures include source management, local exhausts, ventilation, exposure…
IMPROVEMENTS IN THE THERMAL NEUTRON CALIBRATION UNIT, TNF2, AT LNMRI/IRD.
Astuto, A; Fernandes, S S; Patrão, K C S; Fonseca, E S; Pereira, W W; Lopes, R T
2018-02-21
The standard thermal neutron flux unit, TNF2, in the Brazilian National Ionizing Radiation Metrology Laboratory was rebuilt. Fluence is still achieved by moderating of four 241Am-Be sources with 0.6 TBq each. The facility was again simulated and redesigned with graphite core and paraffin added graphite blocks surrounding it. Simulations using the MCNPX code on different geometric arrangements of moderator materials and neutron sources were performed. The resulting neutron fluence quality in terms of intensity, spectrum and cadmium ratio was evaluated. After this step, the system was assembled based on the results obtained from the simulations and measurements were performed with equipment existing in LNMRI/IRD and by simulated equipment. This work focuses on the characterization of a central chamber point and external points around the TNF2 in terms of neutron spectrum, fluence and ambient dose equivalent, H*(10). This system was validated with spectra measurements, fluence and H*(10) to ensure traceability.
Eslinger, P W; Biegalski, S R; Bowyer, T W; Cooper, M W; Haas, D A; Hayes, J C; Hoffman, I; Korpach, E; Yi, J; Miley, H S; Rishel, J P; Ungar, K; White, B; Woods, V T
2014-01-01
Systems designed to monitor airborne radionuclides released from underground nuclear explosions detected radioactive fallout across the northern hemisphere resulting from the Fukushima Dai-ichi Nuclear Power Plant accident in March 2011. Sampling data from multiple International Modeling System locations are combined with atmospheric transport modeling to estimate the magnitude and time sequence of releases of (133)Xe. Modeled dilution factors at five different detection locations were combined with 57 atmospheric concentration measurements of (133)Xe taken from March 18 to March 23 to estimate the source term. This analysis suggests that 92% of the 1.24 × 10(19) Bq of (133)Xe present in the three operating reactors at the time of the earthquake was released to the atmosphere over a 3 d period. An uncertainty analysis bounds the release estimates to 54-129% of available (133)Xe inventory. Copyright © 2013 Elsevier Ltd. All rights reserved.
Tang, Xiao-Bin; Meng, Jia; Wang, Peng; Cao, Ye; Huang, Xi; Wen, Liang-Sheng; Chen, Da
2016-04-01
A small-sized UAV (NH-UAV) airborne system with two gamma spectrometers (LaBr3 detector and HPGe detector) was developed to monitor activity concentration in serious nuclear accidents, such as the Fukushima nuclear accident. The efficiency calibration and determination of minimum detectable activity concentration (MDAC) of the specific system were studied by MC simulations at different flight altitudes, different horizontal distances from the detection position to the source term center and different source term sizes. Both air and ground radiation were considered in the models. The results obtained may provide instructive suggestions for in-situ radioactivity measurements of NH-UAV. Copyright © 2016 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friese, Judah I.; Kephart, Rosara F.; Lucas, Dawn D.
2013-05-01
The Comprehensive Nuclear Test Ban Treaty (CTBT) has remote radionuclide monitoring followed by an On Site Inspection (OSI) to clarify the nature of a suspect event. An important aspect of radionuclide measurements on site is the discrimination of other potential sources of similar radionuclides such as reactor accidents or medical isotope production. The Chernobyl and Fukushima nuclear reactor disasters offer two different reactor source term environmental inputs that can be compared against historical measurements of nuclear explosions. The comparison of whole-sample gamma spectrometry measurements from these three events and the analysis of similarities and differences are presented. This analysis ismore » a step toward confirming what is needed for measurements during an OSI under the auspices of the Comprehensive Test Ban Treaty.« less
Accuracy-preserving source term quadrature for third-order edge-based discretization
NASA Astrophysics Data System (ADS)
Nishikawa, Hiroaki; Liu, Yi
2017-09-01
In this paper, we derive a family of source term quadrature formulas for preserving third-order accuracy of the node-centered edge-based discretization for conservation laws with source terms on arbitrary simplex grids. A three-parameter family of source term quadrature formulas is derived, and as a subset, a one-parameter family of economical formulas is identified that does not require second derivatives of the source term. Among the economical formulas, a unique formula is then derived that does not require gradients of the source term at neighbor nodes, thus leading to a significantly smaller discretization stencil for source terms. All the formulas derived in this paper do not require a boundary closure, and therefore can be directly applied at boundary nodes. Numerical results are presented to demonstrate third-order accuracy at interior and boundary nodes for one-dimensional grids and linear triangular/tetrahedral grids over straight and curved geometries.
Hamilton, Clayon B; Wong, Ming-Kin; Gignac, Monique A M; Davis, Aileen M; Chesworth, Bert M
2017-01-01
To identify validated measures that capture illness perception and behavior and have been used to assess people who have knee pain/osteoarthritis. A scoping review was performed. Nine electronic databases were searched for records from inception through April 19, 2015. Search terms included illness perception, illness behavior, knee, pain, osteoarthritis, and their related terms. This review included English language publications of primary data on people with knee pain/osteoarthritis who were assessed with validated measures capturing any of 4 components of illness perception and behavior: monitor body, define and interpret symptoms, take remedial action, and utilize sources of help. Seventy-one publications included relevant measures. Two reviewers independently coded and analyzed each relevant measure within the 4 components. Sixteen measures were identified that capture components of illness perception and behavior in the target population. These measures were originally developed to capture constructs that include coping strategies/skills/styles, illness belief, illness perception, self-efficacy, and pain behavior. Coding results indicated that 5, 11, 12, and 5 of these measures included the monitor body, define and interpret symptoms, take remedial action, and utilize sources of help components, respectively. Several validated measures were interpreted as capturing some components, and only 1 measure was interpreted as capturing all of the components of illness perception and behavior in the target population. A measure that comprehensively captures illness perception and behavior could be valuable for informing and evaluating therapy for patients along a continuum of symptomatic knee osteoarthritis. © 2016 World Institute of Pain.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Paul L. Wichlacz
2003-09-01
This source-term summary document is intended to describe the current understanding of contaminant source terms and the conceptual model for potential source-term release to the environment at the Idaho National Engineering and Environmental Laboratory (INEEL), as presented in published INEEL reports. The document presents a generalized conceptual model of the sources of contamination and describes the general categories of source terms, primary waste forms, and factors that affect the release of contaminants from the waste form into the vadose zone and Snake River Plain Aquifer. Where the information has previously been published and is readily available, summaries of the inventorymore » of contaminants are also included. Uncertainties that affect the estimation of the source term release are also discussed where they have been identified by the Source Term Technical Advisory Group. Areas in which additional information are needed (i.e., research needs) are also identified.« less
Angular dependence of source-target-detector in active mode standoff infrared detection
NASA Astrophysics Data System (ADS)
Pacheco-Londoño, Leonardo C.; Castro-Suarez, John R.; Aparicio-Bolaños, Joaquín. A.; Hernández-Rivera, Samuel P.
2013-06-01
Active mode standoff measurement using infrared spectroscopy were carried out in which the angle between target and the source was varied from 0-70° with respect to the surface normal of substrates containing traces of highly energetic materials (explosives). The experiments were made using three infrared sources: a modulated source (Mod-FTIR), an unmodulated source (UnMod-FTIR) and a scanning quantum cascade laser (QCL), part of a dispersive mid infrared (MIR) spectrometer. The targets consisted of PENT 200 μg/cm2 deposited on aluminum plates placed at 1 m from the sources. The evaluation of the three modalities was aimed at verifying the influence of the highly collimated laser beam in the detection in comparison with the other sources. The Mod-FTIR performed better than QCL source in terms of the MIR signal intensity decrease with increasing angle.
NASA Astrophysics Data System (ADS)
Davoine, X.; Bocquet, M.
2007-03-01
The reconstruction of the Chernobyl accident source term has been previously carried out using core inventories, but also back and forth confrontations between model simulations and activity concentration or deposited activity measurements. The approach presented in this paper is based on inverse modelling techniques. It relies both on the activity concentration measurements and on the adjoint of a chemistry-transport model. The location of the release is assumed to be known, and one is looking for a source term available for long-range transport that depends both on time and altitude. The method relies on the maximum entropy on the mean principle and exploits source positivity. The inversion results are mainly sensitive to two tuning parameters, a mass scale and the scale of the prior errors in the inversion. To overcome this hardship, we resort to the statistical L-curve method to estimate balanced values for these two parameters. Once this is done, many of the retrieved features of the source are robust within a reasonable range of parameter values. Our results favour the acknowledged three-step scenario, with a strong initial release (26 to 27 April), followed by a weak emission period of four days (28 April-1 May) and again a release, longer but less intense than the initial one (2 May-6 May). The retrieved quantities of iodine-131, caesium-134 and caesium-137 that have been released are in good agreement with the latest reported estimations. Yet, a stronger apportionment of the total released activity is ascribed to the first period and less to the third one. Finer chronological details are obtained, such as a sequence of eruptive episodes in the first two days, likely related to the modulation of the boundary layer diurnal cycle. In addition, the first two-day release surges are found to have effectively reached an altitude up to the top of the domain (5000 m).
Systematic review of studies of staffing and quality in nursing homes.
Bostick, Jane E; Rantz, Marilyn J; Flesner, Marcia K; Riggs, C Jo
2006-07-01
To evaluate a range of staffing measures and data sources for long-term use in public reporting of staffing as a quality measure in nursing homes. Eighty-seven research articles and government documents published from 1975 to 2003 were reviewed and summarized. Relevant content was extracted and organized around 3 themes: staffing measures, quality measures, and risk adjustment variables. Data sources for staffing information were also identified. There is a proven association between higher total staffing levels (especially licensed staff) and improved quality of care. Studies also indicate a significant relationship between high turnover and poor resident outcomes. Functional ability, pressure ulcers, and weight loss are the most sensitive quality indicators linked to staffing. The best national data sources for staffing and quality include the Minimum Data Set (MDS) and On-line Survey and Certification Automated Records (OSCAR). However, the accuracy of this self-reported information requires further reliability and validity testing. A nationwide instrument needs to be developed to accurately measure staff turnover. Large-scale studies using payroll data to measure staff retention and its impact on resident outcomes are recommended. Future research should use the most nurse-sensitive quality indicators such as pressure ulcers, functional status, and weight loss.
Effort-reward imbalance and its association with health among permanent and fixed-term workers
2010-01-01
Background In the past decade, the changing labor market seems to have rejected the traditional standards employment and has begun to support a variety of non-standard forms of work in their place. The purpose of our study was to compare the degree of job stress, sources of job stress, and association of high job stress with health among permanent and fixed-term workers. Methods Our study subjects were 709 male workers aged 30 to 49 years in a suburb of Tokyo, Japan. In 2008, we conducted a cross-sectional study to compare job stress using an effort-reward imbalance (ERI) model questionnaire. Lifestyles, subjective symptoms, and body mass index were also observed from the 2008 health check-up data. Results The rate of job stress of the high-risk group measured by ERI questionnaire was not different between permanent and fixed-term workers. However, the content of the ERI components differed. Permanent workers were distressed more by effort, overwork, or job demand, while fixed-term workers were distressed more by their job insecurity. Moreover, higher ERI was associated with existence of subjective symptoms (OR = 2.07, 95% CI: 1.42-3.03) and obesity (OR = 2.84, 95% CI:1.78-4.53) in fixed-term workers while this tendency was not found in permanent workers. Conclusions Our study showed that workers with different employment types, permanent and fixed-term, have dissimilar sources of job stress even though their degree of job stress seems to be the same. High ERI was associated with existing subjective symptoms and obesity in fixed-term workers. Therefore, understanding different sources of job stress and their association with health among permanent and fixed-term workers should be considered to prevent further health problems. PMID:21054838
Integrated Global Observation Strategy - Ozone and Atmospheric Chemistry Project
NASA Technical Reports Server (NTRS)
Hilsenrath, Ernest; Readings, C. J.; Kaye, J.; Mohnen, V.; Einaudi, Franco (Technical Monitor)
2000-01-01
The "Long Term Continuity of Stratospheric Ozone Measurements and Atmospheric Chemistry" project was one of six established by the Committee on Earth Observing Satellites (CEOS) in response to the Integrated Global Observing Strategy (IGOS) initiative. IGOS links satellite and ground based systems for global environmental observations. The strategy of this project is to develop a consensus of user requirements including the scientific (SPARC, IGAC, WCRP) and the applications community (WMO, UNEP) and to develop a long-term international plan for ozone and atmospheric chemistry measurements. The major components of the observing system include operational and research (meeting certain criteria) satellite platforms planned by the space faring nations which are integrated with a well supported and sustained ground, aircraft, and balloon measurements program for directed observations as well satellite validation. Highly integrated and continuous measurements of ozone, validation, and reanalysis efforts are essential to meet the international scientific and applications goals. In order to understand ozone trends, climate change, and air quality, it is essential to conduct long term measurements of certain other atmospheric species. These species include key source, radical, and reservoir constituents.
A study of comprehension and use of weather information by various agricultural groups in Wisconsin
NASA Technical Reports Server (NTRS)
Smith, J. L.
1972-01-01
An attempt was made to determine whether current techniques are adequate for communicating improved weather forecasts to users. Primary concern was for agricultural users. Efforts were made to learn the preferred source of weather forecasts and the frequency of use. Attempts were also made to measure knowledge of specific terms having to do with weather and comprehension of terms less often used but critical to varying intensities of weather.
40 CFR 408.11 - Specialized definitions.
Code of Federal Regulations, 2011 CFR
2011-07-01
... STANDARDS CANNED AND PRESERVED SEAFOOD PROCESSING POINT SOURCE CATEGORY Farm-Raised Catfish Processing... apply to this subpart. (b) The term oil and grease shall mean those components of a waste water amenable to measurement by the method described in Methods for Chemical Analysis of Water and Wastes, 1971...
Anaerobic fitness tests: what are we measuring?
Van Praagh, Emmanuel
2007-01-01
Anaerobic fitness, during growth and development, has not received the same attention from researchers as aerobic fitness. This is surprising given the level of anaerobic energy used daily during childhood and adolescence. During physical activity and sport, the child is spontaneously more attracted to short-burst movements than to long-term activities. It is, however, well known that in anaerobic activities such as sprint cycling, sprint running or sprint swimming, the child's performance is distinctly poorer than that of the adult. This partly reflects the child's lesser ability to generate mechanical energy from chemical energy sources during short-term high-intensity work or exercise. Direct measurements of the rate or capacity of anaerobic pathways for energy turnover presents several ethical and methodological difficulties. Therefore, rather than measure energy supply, pediatric exercise scientists have concentrated on measuring short-term power output by means of standardized protocol tests such as short-term cycling power tests, running tests or vertical jump tests. There is, however, no perfect test and, therefore, it is important to acknowledge the benefits and limitations of each testing method. Mass-related short-term power output was shown to increase dramatically during growth and development, whereas the corresponding increase in peak blood lactate was considerably lower. This suggests that the observed difference between children and adolescents during short-term power output testing may be related to neuromuscular factors, hormonal factors and improved motor coordination.
Inversion of Atmospheric Tracer Measurements, Localization of Sources
NASA Astrophysics Data System (ADS)
Issartel, J.-P.; Cabrit, B.; Hourdin, F.; Idelkadi, A.
When abnormal concentrations of a pollutant are observed in the atmosphere, the question of its origin arises immediately. The radioactivity from Tchernobyl was de- tected in Sweden before the accident was announced. This situation emphasizes the psychological, political and medical stakes of a rapid identification of sources. In tech- nical terms, most industrial sources can be modeled as a fixed point at ground level with undetermined duration. The classical method of identification involves the cal- culation of a backtrajectory departing from the detector with an upstream integration of the wind field. We were first involved in such questions as we evaluated the ef- ficiency of the international monitoring network planned in the frame of the Com- prehensive Test Ban Treaty. We propose a new approach of backtracking based upon the use of retroplumes associated to available measurements. Firstly the retroplume is related to inverse transport processes, describing quantitatively how the air in a sam- ple originates from regions that are all the more extended and diffuse as we go back far in the past. Secondly it clarifies the sensibility of the measurement with respect to all potential sources. It is therefore calculated by adjoint equations including of course diffusive processes. Thirdly, the statistical interpretation, valid as far as sin- gle particles are concerned, should not be used to investigate the position and date of a macroscopic source. In that case, the retroplume rather induces a straightforward constraint between the intensity of the source and its position. When more than one measurements are available, including zero valued measurements, the source satisfies the same number of linear relations tightly related to the retroplumes. This system of linear relations can be handled through the simplex algorithm in order to make the above intensity-position correlation more restrictive. This method enables to manage in a quantitative manner the unavoidable ambiguity of atmospheric phenomena. When several measurements are available the ambiguity about the identification of a source is reduced significantly.
Predicting Near-Term Water Quality from Satellite Observations of Watershed Conditions
NASA Astrophysics Data System (ADS)
Weiss, W. J.; Wang, L.; Hoffman, K.; West, D.; Mehta, A. V.; Lee, C.
2017-12-01
Despite the strong influence of watershed conditions on source water quality, most water utilities and water resource agencies do not currently have the capability to monitor watershed sources of contamination with great temporal or spatial detail. Typically, knowledge of source water quality is limited to periodic grab sampling; automated monitoring of a limited number of parameters at a few select locations; and/or monitoring relevant constituents at a treatment plant intake. While important, such observations are not sufficient to inform proactive watershed or source water management at a monthly or seasonal scale. Satellite remote sensing data on the other hand can provide a snapshot of an entire watershed at regular, sub-monthly intervals, helping analysts characterize watershed conditions and identify trends that could signal changes in source water quality. Accordingly, the authors are investigating correlations between satellite remote sensing observations of watersheds and source water quality, at a variety of spatial and temporal scales and lags. While correlations between remote sensing observations and direct in situ measurements of water quality have been well described in the literature, there are few studies that link remote sensing observations across a watershed with near-term predictions of water quality. In this presentation, the authors will describe results of statistical analyses and discuss how these results are being used to inform development of a desktop decision support tool to support predictive application of remote sensing data. Predictor variables under evaluation include parameters that describe vegetative conditions; parameters that describe climate/weather conditions; and non-remote sensing, in situ measurements. Water quality parameters under investigation include nitrogen, phosphorus, organic carbon, chlorophyll-a, and turbidity.
Screening and validation of EXTraS data products
NASA Astrophysics Data System (ADS)
Carpano, Stefania; Haberl, F.; De Luca, A.; Tiengo, A.: Israel, G.; Rodriguez, G.; Belfiore, A.; Rosen, S.; Read, A.; Wilms, J.; Kreikenbohm, A.; Law-Green, D.
2015-09-01
The EXTraS project (Exploring the X-ray Transient and variable Sky) is aimed at fullyexploring the serendipitous content of the XMM-Newton EPIC database in the timedomain. The project is funded within the EU/FP7-Cooperation Space framework and is carried out by a collaboration including INAF (Italy), IUSS (Italy), CNR/IMATI (Italy), University of Leicester (UK), MPE (Germany) and ECAP (Germany). The several tasks consist in characterise aperiodicvariability for all 3XMM sources, search for short-term periodic variability on hundreds of thousands sources, detect new transient sources that are missed by standard source detection and hence not belonging to the 3XMM catalogue, search for long term variability by measuring fluxes or upper limits for both pointed and slew observations, and finally perform multiwavelength characterisation andclassification. Screening and validation of the different products is essentially in order to reject flawed results, generated by the automatic pipelines. We present here the screening tool we developed in the form of a Graphical User Interface and our plans for a systematic screening of the different catalogues.
NASA Astrophysics Data System (ADS)
Ferro, Andrea R.; Klepeis, Neil E.; Ott, Wayne R.; Nazaroff, William W.; Hildemann, Lynn M.; Switzer, Paul
Residential interior door positions influence the pollutant concentrations that result from short-term indoor sources, such as cigarettes, candles, and incense. To elucidate this influence, we reviewed past studies and conducted new experiments in three residences: a single-story 714 m 3 ranch-style house, a 510 m 3 two-story split-level house, and a 200 m 3 two-story house. During the experiments, we released sulfur hexafluoride or carbon monoxide tracer gas over short periods (≤30 min) and measured concentrations in the source room and at least one other (receptor) room for various interior door opening positions. We found that closing a door between rooms effectively prevented transport of air pollutants, reducing the average concentration in the receptor room relative to the source room by 57-100% over exposure periods of 1-8 h. When intervening doors were partially or fully open, the reduction in average concentrations ranged from 3% to 99%, varying as a function of door opening width and the distance between source and receptor rooms.
Multi-Decadal Variation of Aerosols: Sources, Transport, and Climate Effects
NASA Technical Reports Server (NTRS)
Chin, Mian; Diehl, Thomas; Bian, Huisheng; Streets, David
2008-01-01
We present a global model study of multi-decadal changes of atmospheric aerosols and their climate effects using a global chemistry transport model along with the near-term to longterm data records. We focus on a 27-year time period of satellite era from 1980 to 2006, during which a suite of aerosol data from satellite observations, ground-based measurements, and intensive field experiments have become available. We will use the Goddard Chemistry Aerosol Radiation and Transport (GOCART) model, which involves a time-varying, comprehensive global emission dataset that we put together in our previous investigations and will be improved/extended in this project. This global emission dataset includes emissions of aerosols and their precursors from fuel combustion, biomass burning, volcanic eruptions, and other sources from 1980 to the present. Using the model and satellite data, we will analyze (1) the long-term global and regional aerosol trends and their relationship to the changes of aerosol and precursor emissions from anthropogenic and natural sources, (2) the intercontinental source-receptor relationships controlled by emission, transport pathway, and climate variability.
Electron Energy Deposition in Atomic Nitrogen
1987-10-06
knovn theoretical results, and their relative accuracy in comparison to existing measurements and calculations is given elsevhere. 20 2.1 The Source Term...with the proper choice of parameters, reduces to vell-known theoretical results. 20 Table 2 gives the parameters for collisional excitation of the...calculations of McGuire 36 and experimental measurements of Brook et al.3 7 Additional theoretical and experimental results are discussed in detail elsevhere
1983-11-01
successfully. I- Accession For NTIS -GO iiiONa DTIC TAB t Unannounced - Justificatio Distribution/ I Availability Codes vail and/or DIst Special IA-11...terms of initial signal power. An active sensor must be excited externally. Such a sensor receives its power from an external source and merely modulates...electrons in the material to gain L enough energy to be emitted. The voltage source causes a positive potential to be felt on the collector, thus causing the
Alexander, Angel M.; Flynn, Kathryn E.; Hahn, Elizabeth A.; Jeffery, Diana D.; Keefe, Francis J.; Reeve, Bryce B.; Schultz, Wesley; Reese, Jennifer Barsky; Shelby, Rebecca A.; Weinfurt, Kevin P.
2014-01-01
Introduction There is a significant gap in research regarding the readability and comprehension of existing sexual function measures. Patient-reported outcome measures may use terms not well understood by respondents with low literacy. Aim To test comprehension of words and phrases typically used in sexual function measures to improve validity for all individuals, including those with low literacy. Methods We recruited 20 men and 28 women for cognitive interviews on version 2.0 of the PROMIS Sexual Function and Satisfaction measures. We assessed participants’ reading level using the word reading subtest of the Wide Range Achievement Test (WRAT). Sixteen participants were classified as having low literacy. Main Outcome Measures In the first round of cognitive interviews, each survey item was reviewed by 5 or more people, at least 2 of whom had lower than a ninth-grade reading level (low literacy). Patient feedback was incorporated into a revised version of the items. In the second round of interviews, an additional 3 or more people (at least 1 with low literacy) reviewed each revised item. Results Participants with low literacy had difficulty comprehending terms such as aroused, orgasm, erection, ejaculation, incontinence, and vaginal penetration. Women across a range of literacy levels had difficulty with clinical terms like labia and clitoris. We modified unclear terms to include parenthetical descriptors or slang equivalents, which generally improved comprehension. Conclusions Common words and phrases used across measures of self-reported sexual function are not universally understood. Researchers should appreciate these misunderstandings as a potential source of error in studies using self-reported measures of sexual function. PMID:24902984
On epicardial potential reconstruction using regularization schemes with the L1-norm data term.
Shou, Guofa; Xia, Ling; Liu, Feng; Jiang, Mingfeng; Crozier, Stuart
2011-01-07
The electrocardiographic (ECG) inverse problem is ill-posed and usually solved by regularization schemes. These regularization methods, such as the Tikhonov method, are often based on the L2-norm data and constraint terms. However, L2-norm-based methods inherently provide smoothed inverse solutions that are sensitive to measurement errors, and also lack the capability of localizing and distinguishing multiple proximal cardiac electrical sources. This paper presents alternative regularization schemes employing the L1-norm data term for the reconstruction of epicardial potentials (EPs) from measured body surface potentials (BSPs). During numerical implementation, the iteratively reweighted norm algorithm was applied to solve the L1-norm-related schemes, and measurement noises were considered in the BSP data. The proposed L1-norm data term-based regularization schemes (with L1 and L2 penalty terms of the normal derivative constraint (labelled as L1TV and L1L2)) were compared with the L2-norm data terms (Tikhonov with zero-order and normal derivative constraints, labelled as ZOT and FOT, and the total variation method labelled as L2TV). The studies demonstrated that, with averaged measurement noise, the inverse solutions provided by the L1L2 and FOT algorithms have less relative error values. However, when larger noise occurred in some electrodes (for example, signal lost during measurement), the L1TV and L1L2 methods can obtain more accurate EPs in a robust manner. Therefore the L1-norm data term-based solutions are generally less perturbed by measurement noises, suggesting that the new regularization scheme is promising for providing practical ECG inverse solutions.
Harris, Peter; Philip, Rachel; Robinson, Stephen; Wang, Lian
2016-03-22
Monitoring ocean acoustic noise has been the subject of considerable recent study, motivated by the desire to assess the impact of anthropogenic noise on marine life. A combination of measuring ocean sound using an acoustic sensor network and modelling sources of sound and sound propagation has been proposed as an approach to estimating the acoustic noise map within a region of interest. However, strategies for developing a monitoring network are not well established. In this paper, considerations for designing a network are investigated using a simulated scenario based on the measurement of sound from ships in a shipping lane. Using models for the sources of the sound and for sound propagation, a noise map is calculated and measurements of the noise map by a sensor network within the region of interest are simulated. A compressive sensing algorithm, which exploits the sparsity of the representation of the noise map in terms of the sources, is used to estimate the locations and levels of the sources and thence the entire noise map within the region of interest. It is shown that although the spatial resolution to which the sound sources can be identified is generally limited, estimates of aggregated measures of the noise map can be obtained that are more reliable compared with those provided by other approaches.
Harris, Peter; Philip, Rachel; Robinson, Stephen; Wang, Lian
2016-01-01
Monitoring ocean acoustic noise has been the subject of considerable recent study, motivated by the desire to assess the impact of anthropogenic noise on marine life. A combination of measuring ocean sound using an acoustic sensor network and modelling sources of sound and sound propagation has been proposed as an approach to estimating the acoustic noise map within a region of interest. However, strategies for developing a monitoring network are not well established. In this paper, considerations for designing a network are investigated using a simulated scenario based on the measurement of sound from ships in a shipping lane. Using models for the sources of the sound and for sound propagation, a noise map is calculated and measurements of the noise map by a sensor network within the region of interest are simulated. A compressive sensing algorithm, which exploits the sparsity of the representation of the noise map in terms of the sources, is used to estimate the locations and levels of the sources and thence the entire noise map within the region of interest. It is shown that although the spatial resolution to which the sound sources can be identified is generally limited, estimates of aggregated measures of the noise map can be obtained that are more reliable compared with those provided by other approaches. PMID:27011187
Hernández, Klaudia L; Yannicelli, Beatriz; Olsen, Lasse M; Dorador, Cristina; Menschel, Eduardo J; Molina, Verónica; Remonsellez, Francisco; Hengst, Martha B; Jeffrey, Wade H
2016-01-01
In high altitude environments, extreme levels of solar radiation and important differences of ionic concentrations over narrow spatial scales may modulate microbial activity. In Salar de Huasco, a high-altitude wetland in the Andean mountains, the high diversity of microbial communities has been characterized and associated with strong environmental variability. Communities that differed in light history and environmental conditions, such as nutrient concentrations and salinity from different spatial locations, were assessed for bacterial secondary production (BSP, 3 H-leucine incorporation) response from short-term exposures to solar radiation. We sampled during austral spring seven stations categorized as: (a) source stations, with recently emerged groundwater (no-previous solar exposure); (b) stream running water stations; (c) stations connected to source waters but far downstream from source points; and (d) isolated ponds disconnected from ground sources or streams with a longer isolation and solar exposure history. Very high values of 0.25 μE m -2 s -1 , 72 W m -2 and 12 W m -2 were measured for PAR, UVA, and UVB incident solar radiation, respectively. The environmental factors measured formed two groups of stations reflected by principal component analyses (near to groundwater sources and isolated systems) where isolated ponds had the highest BSP and microbial abundance (35 microalgae taxa, picoeukaryotes, nanoflagellates, and bacteria) plus higher salinities and PO 4 3- concentrations. BSP short-term response (4 h) to solar radiation was measured by 3 H-leucine incorporation under four different solar conditions: full sun, no UVB, PAR, and dark. Microbial communities established in waters with the longest surface exposure (e.g., isolated ponds) had the lowest BSP response to solar radiation treatments, and thus were likely best adapted to solar radiation exposure contrary to ground source waters. These results support our light history (solar exposure) hypothesis where the more isolated the community is from ground water sources, the better adapted it is to solar radiation. We suggest that factors other than solar radiation (e.g., salinity, PO 4 3- , NO 3 - ) are also important in determining microbial productivity in heterogeneous environments such as the Salar de Huasco.
Hernández, Klaudia L.; Yannicelli, Beatriz; Olsen, Lasse M.; Dorador, Cristina; Menschel, Eduardo J.; Molina, Verónica; Remonsellez, Francisco; Hengst, Martha B.; Jeffrey, Wade H.
2016-01-01
In high altitude environments, extreme levels of solar radiation and important differences of ionic concentrations over narrow spatial scales may modulate microbial activity. In Salar de Huasco, a high-altitude wetland in the Andean mountains, the high diversity of microbial communities has been characterized and associated with strong environmental variability. Communities that differed in light history and environmental conditions, such as nutrient concentrations and salinity from different spatial locations, were assessed for bacterial secondary production (BSP, 3H-leucine incorporation) response from short-term exposures to solar radiation. We sampled during austral spring seven stations categorized as: (a) source stations, with recently emerged groundwater (no-previous solar exposure); (b) stream running water stations; (c) stations connected to source waters but far downstream from source points; and (d) isolated ponds disconnected from ground sources or streams with a longer isolation and solar exposure history. Very high values of 0.25 μE m-2 s-1, 72 W m-2 and 12 W m-2 were measured for PAR, UVA, and UVB incident solar radiation, respectively. The environmental factors measured formed two groups of stations reflected by principal component analyses (near to groundwater sources and isolated systems) where isolated ponds had the highest BSP and microbial abundance (35 microalgae taxa, picoeukaryotes, nanoflagellates, and bacteria) plus higher salinities and PO43- concentrations. BSP short-term response (4 h) to solar radiation was measured by 3H-leucine incorporation under four different solar conditions: full sun, no UVB, PAR, and dark. Microbial communities established in waters with the longest surface exposure (e.g., isolated ponds) had the lowest BSP response to solar radiation treatments, and thus were likely best adapted to solar radiation exposure contrary to ground source waters. These results support our light history (solar exposure) hypothesis where the more isolated the community is from ground water sources, the better adapted it is to solar radiation. We suggest that factors other than solar radiation (e.g., salinity, PO43-, NO3-) are also important in determining microbial productivity in heterogeneous environments such as the Salar de Huasco. PMID:27920763
Above and beyond short-term mating, long-term mating is uniquely tied to human personality.
Holtzman, Nicholas S; Strube, Michael J
2013-12-16
To what extent are personality traits and sexual strategies linked? The literature does not provide a clear answer, as it is based on the Sociosexuality model, a one-dimensional model that fails to measure long-term mating (LTM). An improved two-dimensional model separately assesses long-term and short-term mating (STM; Jackson and Kirkpatrick, 2007). In this paper, we link this two-dimensional model to an array of personality traits (Big 5, Dark Triad, and Schizoid Personality). We collected data from different sources (targets and peers; Study 1), and from different nations (United States, Study 1; India, Study 2). We demonstrate for the first time that, above and beyond STM, LTM captures variation in personality.
Improving the limits of detection of low background alpha emission measurements
NASA Astrophysics Data System (ADS)
McNally, Brendan D.; Coleman, Stuart; Harris, Jack T.; Warburton, William K.
2018-01-01
Alpha particle emission - even at extremely low levels - is a significant issue in the search for rare events (e.g., double beta decay, dark matter detection). Traditional measurement techniques require long counting times to measure low sample rates in the presence of much larger instrumental backgrounds. To address this, a commercially available instrument developed by XIA uses pulse shape analysis to discriminate alpha emissions produced by the sample from those produced by other surfaces of the instrument itself. Experience with this system has uncovered two residual sources of background: cosmogenics and radon emanation from internal components. An R&D program is underway to enhance the system and extend the pulse shape analysis technique further, so that these residual sources can be identified and rejected as well. In this paper, we review the theory of operation and pulse shape analysis techniques used in XIA's alpha counter, and briefly explore data suggesting the origin of the residual background terms. We will then present our approach to enhance the system's ability to identify and reject these terms. Finally, we will describe a prototype system that incorporates our concepts and demonstrates their feasibility.
Pesaranghader, Ahmad; Matwin, Stan; Sokolova, Marina; Beiko, Robert G
2016-05-01
Measures of protein functional similarity are essential tools for function prediction, evaluation of protein-protein interactions (PPIs) and other applications. Several existing methods perform comparisons between proteins based on the semantic similarity of their GO terms; however, these measures are highly sensitive to modifications in the topological structure of GO, tend to be focused on specific analytical tasks and concentrate on the GO terms themselves rather than considering their textual definitions. We introduce simDEF, an efficient method for measuring semantic similarity of GO terms using their GO definitions, which is based on the Gloss Vector measure commonly used in natural language processing. The simDEF approach builds optimized definition vectors for all relevant GO terms, and expresses the similarity of a pair of proteins as the cosine of the angle between their definition vectors. Relative to existing similarity measures, when validated on a yeast reference database, simDEF improves correlation with sequence homology by up to 50%, shows a correlation improvement >4% with gene expression in the biological process hierarchy of GO and increases PPI predictability by > 2.5% in F1 score for molecular function hierarchy. Datasets, results and source code are available at http://kiwi.cs.dal.ca/Software/simDEF CONTACT: ahmad.pgh@dal.ca or beiko@cs.dal.ca Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Fermi large area telescope second source catalog
Nolan, P. L.; Abdo, A. A.; Ackermann, M.; ...
2012-03-28
Here, we present the second catalog of high-energy γ-ray sources detected by the Large Area Telescope (LAT), the primary science instrument on the Fermi Gamma-ray Space Telescope (Fermi), derived from data taken during the first 24 months of the science phase of the mission, which began on 2008 August 4. Source detection is based on the average flux over the 24 month period. The second Fermi-LAT catalog (2FGL) includes source location regions, defined in terms of elliptical fits to the 95% confidence regions and spectral fits in terms of power-law, exponentially cutoff power-law, or log-normal forms. Also included are fluxmore » measurements in five energy bands and light curves on monthly intervals for each source. Twelve sources in the catalog are modeled as spatially extended. Furthermore, we provide a detailed comparison of the results from this catalog with those from the first Fermi-LAT catalog (1FGL). Although the diffuse Galactic and isotropic models used in the 2FGL analysis are improved compared to the 1FGL catalog, we attach caution flags to 162 of the sources to indicate possible confusion with residual imperfections in the diffuse model. Finally, the 2FGL catalog contains 1873 sources detected and characterized in the 100 MeV to 100 GeV range of which we consider 127 as being firmly identified and 1171 as being reliably associated with counterparts of known or likely γ-ray-producing source classes.« less
Laser-Based and Ultra-Portable Gas Sensor for Indoor and Outdoor Formaldehyde (HCHO) Monitoring
NASA Astrophysics Data System (ADS)
Shutter, J. D.; Allen, N.; Paul, J.; Thiebaud, J.; So, S.; Scherer, J. J.; Keutsch, F. N.
2017-12-01
While used as a key tracer of oxidative chemistry in the atmosphere, formaldehyde (HCHO) is also a known human carcinogen and is listed and regulated by the United States EPA as a hazardous air pollutant. Combustion processes and photochemical oxidation of volatile organic compounds (VOCs) are the major outdoor sources of HCHO, and building materials and household products are ubiquitous sources of indoor HCHO. Due to the ease with which humans can be exposed to HCHO, it is imperative to monitor levels of both indoor and outdoor HCHO exposure in both short and long-term studies.High-quality direct and indirect methods of quantifying HCHO mixing ratios exist, but instrument size and user-friendliness can make them cumbersome or impractical for certain types of indoor and long-term outdoor measurements. In this study, we present urban HCHO measurements by using a new, commercially-available, ppbv-level accurate HCHO gas sensor (Aeris Technologies' MIRA Pico VOC Laser-Based Gas Analyzer) that is highly portable (29 cm x 20 cm x 10 cm), lightweight (3 kg), easy-to-use, and has low power (15 W) consumption. Using an ultra-compact multipass cell, an absorption path length of 13 m is achieved, resulting in a sensor capable of achieving ppbv/s sensitivity levels with no significant spectral interferences.To demonstrate the utility of the gas sensor for emissions measurements, a GPS was attached to the sensor's housing in order to map mobile HCHO measurements in real-time around the Boston, Massachusetts, metro area. Furthermore, the sensor was placed in residential and industrial environments to show its usefulness for indoor and outdoor pollution measurements. Lastly, we show the feasibility of using the HCHO sensor (or a network of them) in long-term monitoring stations for hazardous air pollutants.
Johnson, W B; Lall, R; Bongar, B; Nordlund, M D
1999-01-01
Objective personality assessment instruments offer a comparatively underutilized source of clinical data in attempts to evaluate and predict risk for suicide. In contrast to focal suicide risk measures, global personality inventories may be useful in identification of long-standing styles that predispose persons to eventual suicidal behavior. This article reviews the empirical literature regarding the efficacy of established personality inventories in predicting suicidality. The authors offer several recommendations for future research with these measures and conclude that such objective personality instruments offer only marginal utility as sources of clinical information in comprehensive suicide risk evaluations. Personality inventories may offer greatest utility in long-term assessment of suicide risk.
NASA Technical Reports Server (NTRS)
Hayden, R. E.; Kadman, Y.; Chanaud, R. C.
1972-01-01
The feasibility of quieting the externally-blown-flap (EBF) noise sources which are due to interaction of jet exhaust flow with deployed flaps was demonstrated on a 1/15-scale 3-flap EBF model. Sound field characteristics were measured and noise reduction fundamentals were reviewed in terms of source models. Test of the 1/15-scale model showed broadband noise reductions of up to 20 dB resulting from combination of variable impedance flap treatment and mesh grids placed in the jet flow upstream of the flaps. Steady-state lift, drag, and pitching moment were measured with and without noise reduction treatment.
NASA Technical Reports Server (NTRS)
Diamante, J. M.; Englar, T. S., Jr.; Jazwinski, A. H.
1977-01-01
Estimation theory, which originated in guidance and control research, is applied to the analysis of air quality measurements and atmospheric dispersion models to provide reliable area-wide air quality estimates. A method for low dimensional modeling (in terms of the estimation state vector) of the instantaneous and time-average pollutant distributions is discussed. In particular, the fluctuating plume model of Gifford (1959) is extended to provide an expression for the instantaneous concentration due to an elevated point source. Individual models are also developed for all parameters in the instantaneous and the time-average plume equations, including the stochastic properties of the instantaneous fluctuating plume.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tabbagh, A.; Lardy, M.
Temperature measurements on shallow vertical profiles undertaken on Matthews and Hunter volcanoes of the New Hebrides arc (SW Pacific) demonstrate the absence of both unsteady and steady conductive abnormal flux at the location of the studied profiles. The reasons for this absence are explained in terms of limits in depth or magnitude for possible sources of heat inside the volcanoes. It implies that the magma chamber is of rather limited extent. This type of flux measurement has a low cost and it will be possible to implant a series of such temperature profiles on an edifice in order to obtainmore » a map of the flux that could be widely used for the location of heat sources.« less
Poggi, L A; Malizia, A; Ciparisse, J F; Gaudio, P
2016-10-01
An open issue still under investigation by several international entities working on the safety and security field for the foreseen nuclear fusion reactors is the estimation of source terms that are a hazard for the operators and public, and for the machine itself in terms of efficiency and integrity in case of severe accident scenarios. Source term estimation is a crucial key safety issue to be addressed in the future reactors safety assessments, and the estimates available at the time are not sufficiently satisfactory. The lack of neutronic data along with the insufficiently accurate methodologies used until now, calls for an integrated methodology for source term estimation that can provide predictions with an adequate accuracy. This work proposes a complete methodology to estimate dust source terms starting from a broad information gathering. The wide number of parameters that can influence dust source term production is reduced with statistical tools using a combination of screening, sensitivity analysis, and uncertainty analysis. Finally, a preliminary and simplified methodology for dust source term production prediction for future devices is presented.
A novel method for detecting light source for digital images forensic
NASA Astrophysics Data System (ADS)
Roy, A. K.; Mitra, S. K.; Agrawal, R.
2011-06-01
Manipulation in image has been in practice since centuries. These manipulated images are intended to alter facts — facts of ethics, morality, politics, sex, celebrity or chaos. Image forensic science is used to detect these manipulations in a digital image. There are several standard ways to analyze an image for manipulation. Each one has some limitation. Also very rarely any method tried to capitalize on the way image was taken by the camera. We propose a new method that is based on light and its shade as light and shade are the fundamental input resources that may carry all the information of the image. The proposed method measures the direction of light source and uses the light based technique for identification of any intentional partial manipulation in the said digital image. The method is tested for known manipulated images to correctly identify the light sources. The light source of an image is measured in terms of angle. The experimental results show the robustness of the methodology.
NASA Astrophysics Data System (ADS)
Galluzzi, V.; Massardi, M.; Bonaldi, A.; Casasola, V.; Gregorini, L.; Trombetti, T.; Burigana, C.; Bonato, M.; De Zotti, G.; Ricci, R.; Stevens, J.; Ekers, R. D.; Bonavera, L.; di Serego Alighieri, S.; Liuzzo, E.; López-Caniego, M.; Paladino, R.; Toffolatti, L.; Tucci, M.; Callingham, J. R.
2018-03-01
We present high sensitivity (σP ≃ 0.6 mJy) polarimetric observations in seven bands, from 2.1 to 38 GHz, of a complete sample of 104 compact extragalactic radio sources brighter than 200 mJy at 20 GHz. Polarization measurements in six bands, in the range 5.5-38 GHz, for 53 of these objects were reported by Galluzzi et al. We have added new measurements in the same six bands for another 51 sources and measurements at 2.1 GHz for the full sample of 104 sources. Also, the previous measurements at 18, 24, 33, and 38 GHz were re-calibrated using the updated model for the flux density absolute calibrator, PKS1934-638, not available for the earlier analysis. The observations, carried out with the Australia Telescope Compact Array, achieved a 90 per cent detection rate (at 5σ) in polarization. 89 of our sources have a counterpart in the 72-231 MHz GLEAM (GaLactic and Extragalactic All-sky Murchison Widefield Array) survey, providing an unparalleled spectral coverage of 2.7 decades of frequency for these sources. While the total intensity data from 5.5 to 38 GHz could be interpreted in terms of single component emission, a joint analysis of more extended total intensity spectra presented here, and of the polarization spectra, reveals that over 90 per cent of our sources show clear indications of at least two emission components. We interpret this as an evidence of recurrent activity. Our high sensitivity polarimetry has allowed a 5σ detection of the weak circular polarization for ˜ 38 per cent of the data set, and a deeper estimate of 20 GHz polarization source counts than has been possible so far.
NASA Astrophysics Data System (ADS)
Nagai, Haruyasu; Terada, Hiroaki; Tsuduki, Katsunori; Katata, Genki; Ota, Masakazu; Furuno, Akiko; Akari, Shusaku
2017-09-01
In order to assess the radiological dose to the public resulting from the Fukushima Daiichi Nuclear Power Station (FDNPS) accident in Japan, especially for the early phase of the accident when no measured data are available for that purpose, the spatial and temporal distribution of radioactive materials in the environment are reconstructed by computer simulations. In this study, by refining the source term of radioactive materials discharged into the atmosphere and modifying the atmospheric transport, dispersion and deposition model (ATDM), the atmospheric dispersion simulation of radioactive materials is improved. Then, a database of spatiotemporal distribution of radioactive materials in the air and on the ground surface is developed from the output of the simulation. This database is used in other studies for the dose assessment by coupling with the behavioral pattern of evacuees from the FDNPS accident. By the improvement of the ATDM simulation to use a new meteorological model and sophisticated deposition scheme, the ATDM simulations reproduced well the 137Cs and 131I deposition patterns. For the better reproducibility of dispersion processes, further refinement of the source term was carried out by optimizing it to the improved ATDM simulation by using new monitoring data.
Development of a wireless air pollution sensor package for aerial-sampling of emissions
A new sensor system for mobile and aerial emission sampling was developed for open area pollutant sources, such as prescribed forest burns. The sensor system, termed “Kolibri”, consists of multiple low-cost air quality sensors measuring CO2, CO, samplers for particulate matter wi...
Monitoring Knowledge Base (MKB)
The Monitoring Knowledge Base (MKB) is a compilation of emissions measurement and monitoring techniques associated with air pollution control devices, industrial process descriptions, and permitting techniques, including flexible permit development. Using MKB, one can gain a comprehensive understanding of emissions sources, control devices, and monitoring techniques, enabling one to determine appropriate permit terms and conditions.
Why Save Wilderness?--Fruits and Veggies!
ERIC Educational Resources Information Center
Kowalewski, David
2015-01-01
Why save wilderness? Environmental educators usually offer ecosystemic and aesthetic reasons, yet clearly this abstract approach has failed to resonate with the wider public. In this article I adopt a nutritional strategy based on a broad array of sources. Wild plant food, in terms of economics, ubiquity, and other measures, performs very well…
Operating envelopes of particle sizing instrumentation used for icing research
NASA Technical Reports Server (NTRS)
Hovenac, Edward A.
1987-01-01
The Forward Scattering Spectrometer Probe and the Optical Array Probe are analyzed in terms of their ability to make accurate determinations of water droplet size distributions. Sources of counting and sizing errors are explained. The paper describes ways of identifying these errors and how they can affect measurement.
Gestalt concept of closure: a construct without closure.
Wasserstein, Jeanette
2002-12-01
This comment reviews the original Gestalt literature which introduced the concept of 'closure'. It is argued that the meaning of 'closure' was confounded in the source literature and, thus, the term connotes more than it denotes. Research based on different measures of this ambiguous construct inevitably may not always converge.
Assessment of Methane Emissions from Oil and Gas Production Pads using Mobile Measurements
Journal Article Abstract --- "A mobile source inspection approach called OTM 33A was used to quantify short-term methane emission rates from 218 oil and gas production pads in Texas, Colorado, and Wyoming from 2010 to 2013. The emission rates were log-normally distributed with ...
Morgan, Kimbly; Osterling, Kristin; Gilbert, Robert; Dechman, Gail
2015-01-01
To determine the effects of short- and long-term use of autogenic drainage (AD) on pulmonary function and sputum recovery in people with cystic fibrosis (CF). The authors conducted a systematic review of randomized and quasi-randomized clinical trials in which participants were people with CF who use AD as their sole airway clearance technique. Searches in 4 databases and secondary sources using 5 key terms yielded 735 articles, of which 58 contained the terms autogenic drainage and cystic fibrosis. Ultimately, 4 studies, 2 of which were long term, were included. All measured forced expiratory volume in 1 second (FEV1) and found no change. The long-term studies were underpowered to detect change in FEV1; however, the short-term studies found a clinically significant sputum yield (≥4 g). AD has been shown to produce clinically significant sputum yields in a limited number of investigations. The effect of AD on the function of the pulmonary system remains uncertain, and questions have emerged regarding the appropriateness of FEV1 as a valid measure of airway clearance from peripheral lung regions. Further consideration should be given to the use of FEV1 as a primary measure of the effect of AD.
Osterling, Kristin; Gilbert, Robert; Dechman, Gail
2015-01-01
ABSTRACT Purpose: To determine the effects of short- and long-term use of autogenic drainage (AD) on pulmonary function and sputum recovery in people with cystic fibrosis (CF). Methods: The authors conducted a systematic review of randomized and quasi-randomized clinical trials in which participants were people with CF who use AD as their sole airway clearance technique. Results: Searches in 4 databases and secondary sources using 5 key terms yielded 735 articles, of which 58 contained the terms autogenic drainage and cystic fibrosis. Ultimately, 4 studies, 2 of which were long term, were included. All measured forced expiratory volume in 1 second (FEV1) and found no change. The long-term studies were underpowered to detect change in FEV1; however, the short-term studies found a clinically significant sputum yield (≥4 g). Conclusion: AD has been shown to produce clinically significant sputum yields in a limited number of investigations. The effect of AD on the function of the pulmonary system remains uncertain, and questions have emerged regarding the appropriateness of FEV1 as a valid measure of airway clearance from peripheral lung regions. Further consideration should be given to the use of FEV1 as a primary measure of the effect of AD. PMID:27504031
NASA Technical Reports Server (NTRS)
Granados Munoz, Maria Jose; Johnson, Matthew S.; Leblanc, Thierry
2016-01-01
In the past decades, significant efforts have been made to increase tropospheric ozone long-term monitoring. A large number of ground-based, airborne and space-borne instruments are currently providing valuable data to contribute to better understand tropospheric ozone budget and variability. Nonetheless, most of these instruments provide in-situ surface and column-integrated data, whereas vertically resolved measurements are still scarce. Besides ozonesondes and aircraft, lidar measurements have proven to be valuable tropospheric ozone profilers. Using the measurements from the tropospheric ozone differential absorption lidar (DIAL) located at the JPL Table Mountain Facility, California, and the GEOS-Chem and GEOS-5 model outputs, the impact of the North American monsoon on tropospheric ozone during summer 2014 is investigated. The influence of the Monsoon lightning-induced NOx will be evaluated against other sources (e.g. local anthropogenic emissions and the stratosphere) using also complementary data such as backward-trajectories analysis, coincident water vapor lidar measurements, and surface ozone in-situ measurements.
Systematic study of error sources in supersonic skin-friction balance measurements
NASA Technical Reports Server (NTRS)
Allen, J. M.
1976-01-01
An experimental study was performed to investigate potential error sources in data obtained with a self-nulling, moment-measuring, skin-friction balance. The balance was installed in the sidewall of a supersonic wind tunnel, and independent measurements of the three forces contributing to the balance output (skin friction, lip force, and off-center normal force) were made for a range of gap size and element protrusion. The relatively good agreement between the balance data and the sum of these three independently measured forces validated the three-term model used. No advantage to a small gap size was found; in fact, the larger gaps were preferable. Perfect element alignment with the surrounding test surface resulted in very small balance errors. However, if small protrusion errors are unavoidable, no advantage was found in having the element slightly below the surrounding test surface rather than above it.
van der Geest, Ivana M M; van den Heuvel-Eibrink, Marry M; Falkenburg, Nette; Michiels, Erna M C; van Vliet, Liesbeth; Pieters, Rob; Darlington, Anne-Sophie E
2015-05-01
The loss of a child is associated with an increased risk for developing psychological problems. However, studies investigating the impact of parents' faith and hope for a cure during the palliative phase on long-term parental psychological functioning are limited. The study's objective was to explore the role of faith and hope as a source of coping and indicator of long-term parental adjustment. Eighty-nine parents of 57 children who died of cancer completed questionnaires retrospectively, exploring faith, hope, and sources of coping, and measuring parents' current level of grief and depression. For 19 parents (21%) faith was very important during the palliative phase. The majority of parents remained hopeful for a meaningful time with their child (n=68, 76%); a pain-free death (n=58, 65%); and a cure (n=30, 34%). Their child (n=70, 79%) was parents' main source of coping. Twelve parents (14%) suffered from traumatic grief, and 22 parents (25%) showed symptoms of depression. Parents' faith was not associated with less long-term traumatic grief (OR=0.86, p=0.51) or symptoms of depression (OR=0.95, p=0.74), and parents' hope for a cure was not related to more long-term traumatic grief (OR=1.07, p=0.71) or symptoms of depression (OR=1.12, p=0.47). Faith was important for a minority of parents and was not associated with less long-term traumatic grief or symptoms of depression. The majority of parents remained hopeful. Hope for a cure was not associated with more long-term traumatic grief or symptoms of depression.
Numerical and experimental evaluations of the flow past nested chevrons
NASA Technical Reports Server (NTRS)
Foss, J. F.; Foss, J. K.; Spalart, P. R.
1989-01-01
An effort is made to contribute to the development of CFD by relating the successful use of vortex dynamics in the computation of the pressure drop past a planar array of chevron-shaped obstructions. An ensemble of results was used to compute the loss coefficient k, stimulating an experimental program for the assessment of the measured loss coefficient for the same geometry. The most provocative result of this study has been the representation of kinetic energy production in terms of vorticity source terms.
NASA Astrophysics Data System (ADS)
Čehovski, Marko; Döring, Sebastian; Rabe, Torsten; Caspary, Reinhard; Kowalsky, Wolfgang
2016-04-01
Organic laser sources offer the opportunity to integrate flexible and widely tunable lasers in polymer waveguide circuits, e.g. for Lab-on-Foil applications. Therefore, it is necessary to understand gain and degradation processes for long-term operation. In this paper we address the challenge of life-time (degradation) measurements of photoluminescence (PL) and optical gain in thin-film lasers. The well known guest-host system of aluminum-chelate Alq3 (Tris-(8-hydroxyquinoline)aluminum) as host material and the laser dye DCM2 (4-(Dicyanomethylene)-2- methyl-6-julolidyl-9-enyl-4H-pyran) as guest material is employed as laser active material. Sample layers have been built up by co-evaporation in an ultrahigh (UHV) vacuum chamber. 200nm thick films of Alq3:DCM2 with different doping concentrations have been processed onto glass and thermally oxidized silicon substrates. The gain measurements have been performed by the variable stripe length (VSL) method. This measurement technique allows to determine the thin-film waveguide gain and loss, respectively. For the measurements the samples were excited with UV irradiation (ƛ = 355nm) under nitrogen atmosphere by a passively Q-switched laser source. PL degradation measurements with regard to the optical gain have been done at laser threshold (approximately 3 μJ/cm2), five times above laser threshold and 10 times above laser threshold. A t50-PL lifetime of > 107 pulses could be measured at a maximum excitation energy density of 32 μJ/cm2. This allows for a detailed analysis of the gain degradation mechanism and therefore of the stimulated cross section. Depending on the DCM2 doping concentration C the stimulated cross section was reduced by 35 %. Nevertheless, the results emphasizes the necessity of the investigation of degradation processes in organic laser sources for long-term applications.
Metrological-grade tunable coherent source in the mid-infrared for molecular precision spectroscopy
NASA Astrophysics Data System (ADS)
Insero, G.; Clivati, C.; D'Ambrosio, D.; Cancio Pastor, P.; Verde, M.; Schunemann, P. G.; Zondy, J.-J.; Inguscio, M.; Calonico, D.; Levi, F.; De Natale, P.; Santambrogio, G.; Borri, S.
2018-02-01
We report on a metrological-grade mid-IR source with a 10-14 short-term instability for high-precision spectroscopy. Our source is based on the combination of a quantum cascade laser and a coherent radiation obtained by difference-frequency generation in an orientation-patterned gallium phosphide (OP-GaP) crystal. The pump and signal lasers are locked to an optical frequency comb referenced to the primary frequency standard via an optical fiber link. We demonstrate the robustness of the apparatus by measuring a vibrational transition around 6 μm on a metastable state of CO molecuels with 11 digits of precision.
Alexander, Angel M; Flynn, Kathryn E; Hahn, Elizabeth A; Jeffery, Diana D; Keefe, Francis J; Reeve, Bryce B; Schultz, Wesley; Reese, Jennifer Barsky; Shelby, Rebecca A; Weinfurt, Kevin P
2014-08-01
There is a significant gap in research regarding the readability and comprehension of existing sexual function measures. Patient-reported outcome measures may use terms not well understood by respondents with low literacy. This study aims to test comprehension of words and phrases typically used in sexual function measures to improve validity for all individuals, including those with low literacy. We recruited 20 men and 28 women for cognitive interviews on version 2.0 of the Patient-Reported Outcome Measurement Information System(®) (PROMIS(®) ) Sexual Function and Satisfaction measures. We assessed participants' reading level using the word reading subtest of the Wide Range Achievement Test. Sixteen participants were classified as having low literacy. In the first round of cognitive interviews, each survey item was reviewed by five or more people, at least two of whom had lower than a ninth-grade reading level (low literacy). Patient feedback was incorporated into a revised version of the items. In the second round of interviews, an additional three or more people (at least one with low literacy) reviewed each revised item. Participants with low literacy had difficulty comprehending terms such as aroused, orgasm, erection, ejaculation, incontinence, and vaginal penetration. Women across a range of literacy levels had difficulty with clinical terms like labia and clitoris. We modified unclear terms to include parenthetical descriptors or slang equivalents, which generally improved comprehension. Common words and phrases used across measures of self-reported sexual function are not universally understood. Researchers should appreciate these misunderstandings as a potential source of error in studies using self-reported measures of sexual function. This study also provides evidence for the importance of including individuals with low literacy in cognitive pretesting during the measure development. © 2014 International Society for Sexual Medicine.
Long-term changes after brief dynamic psychotherapy: symptomatic versus dynamic assessments.
Høglend, P; Sørlie, T; Sørbye, O; Heyerdahl, O; Amlo, S
1992-08-01
Dynamic change in psychotherapy, as measured by theory-related or mode-specific instruments, have been criticized for being too intercorrelated with symptomatic change measures. In this study, long-term changes after brief dynamic psychotherapy were studied in 45 moderately disturbed neurotic patients by a reliable outcome battery. The factor structure of all the change variables suggested that they tapped 2 distinct and stable sources of variance: dynamic and symptomatic change. The categories of overall dynamic change were different from categories of change on the Global Assessment Scale. A small systematic difference was found between the categories of overall dynamic change and the categories of target complaints change also, due to false solutions of dynamic conflicts.
The sound of moving bodies. Ph.D. Thesis - Cambridge Univ.
NASA Technical Reports Server (NTRS)
Brentner, Kenneth Steven
1990-01-01
The importance of the quadrupole source term in the Ffowcs, Williams, and Hawkings (FWH) equation was addressed. The quadrupole source contains fundamental components of the complete fluid mechanics problem, which are ignored only at the risk of error. The results made it clear that any application of the acoustic analogy should begin with all of the source terms in the FWH theory. The direct calculation of the acoustic field as part of the complete unsteady fluid mechanics problem using CFD is considered. It was shown that aeroelastic calculation can indeed be made with CFD codes. The results indicate that the acoustic field is the most susceptible component of the computation to numerical error. Therefore, the ability to measure the damping of acoustic waves is absolutely essential both to develop acoustic computations. Essential groundwork for a new approach to the problem of sound generation by moving bodies is presented. This new computational acoustic approach holds the promise of solving many problems hitherto pushed aside.
Dynamic power balance analysis in JET
NASA Astrophysics Data System (ADS)
Matthews, G. F.; Silburn, S. A.; Challis, C. D.; Eich, T.; Iglesias, D.; King, D.; Sieglin, B.; Contributors, JET
2017-12-01
The full scale realisation of nuclear fusion as an energy source requires a detailed understanding of power and energy balance in current experimental devices. In this we explore whether a global power balance model in which some of the calibration factors applied to the source or sink terms are fitted to the data can provide insight into possible causes of any discrepancies in power and energy balance seen in the JET tokamak. We show that the dynamics in the power balance can only be properly reproduced by including the changes in the thermal stored energy which therefore provides an additional opportunity to cross calibrate other terms in the power balance equation. Although the results are inconclusive with respect to the original goal of identifying the source of the discrepancies in the energy balance, we do find that with optimised parameters an extremely good prediction of the total power measured at the outer divertor target can be obtained over a wide range of pulses with time resolution up to ∼25 ms.
NASA Astrophysics Data System (ADS)
Zaccheo, T. S.; Pernini, T.; Dobler, J. T.; Blume, N.; Braun, M.
2017-12-01
This work highlights the use of the greenhouse-gas laser imaging tomography experiment (GreenLITETM) data in conjunction with a sparse tomography approach to identify and quantify both urban and industrial sources of CO2 and CH4. The GreenLITETM system provides a user-defined set of time-sequenced intersecting chords or integrated column measurements at a fixed height through a quasi-horizontal plane of interest. This plane, with unobstructed views along the lines of sight, may range from complex industrial facilities to a small city scale or urban sector. The continuous time phased absorption measurements are converted to column concentrations and combined with a plume based model to estimate the 2-D distribution of gas concentration over extended areas ranging from 0.04-25 km2. Finally, these 2-D maps of concentration are combined with ancillary meteorological and atmospheric data to identify potential emission sources and provide first order estimates of their associated fluxes. In this presentation, we will provide a brief overview of the systems and results from both controlled release experiments and a long-term system deployment in Paris, FR. These results provide a quantitative assessment of the system's ability to detect and estimate CO2 and CH4 sources, and demonstrate its ability to perform long-term autonomous monitoring and quantification of either persistent or sporadic emissions that may have both health and safety as well as environmental impacts.
Computations of steady-state and transient premixed turbulent flames using pdf methods
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hulek, T.; Lindstedt, R.P.
1996-03-01
Premixed propagating turbulent flames are modeled using a one-point, single time, joint velocity-composition probability density function (pdf) closure. The pdf evolution equation is solved using a Monte Carlo method. The unclosed terms in the pdf equation are modeled using a modified version of the binomial Langevin model for scalar mixing of Valino and Dopazo, and the Haworth and Pope (HP) and Lagrangian Speziale-Sarkar-Gatski (LSSG) models for the viscous dissipation of velocity and the fluctuating pressure gradient. The source terms for the presumed one-step chemical reaction are extracted from the rate of fuel consumption in laminar premixed hydrocarbon flames, computed usingmore » a detailed chemical kinetic mechanism. Steady-state and transient solutions are obtained for planar turbulent methane-air and propane-air flames. The transient solution method features a coupling with a Finite Volume (FV) code to obtain the mean pressure field. The results are compared with the burning velocity measurements of Abdel-Gayed et al. and with velocity measurements obtained in freely propagating propane-air flames by Videto and Santavicca. The effects of different upstream turbulence fields, chemical source terms (different fuels and strained/unstrained laminar flames) and the influence of the velocity statistics models (HP and LSSG) are assessed.« less
Parvez, Shahid; Frost, Kali; Sundararajan, Madhura
2017-01-01
In the absence of shorter term disinfectant byproducts (DBPs) data on regulated Trihalomethanes (THMs) and Haloacetic acids (HAAs), epidemiologists and risk assessors have used long-term annual compliance (LRAA) or quarterly (QA) data to evaluate the association between DBP exposure and adverse birth outcomes, which resulted in inconclusive findings. Therefore, we evaluated the reliability of using long-term LRAA and QA data as an indirect measure for short-term exposure. Short-term residential tap water samples were collected in peak DBP months (May–August) in a community water system with five separate treatment stations and were sourced from surface or groundwater. Samples were analyzed for THMs and HAAs per the EPA (U.S. Environmental Protection Agency) standard methods (524.2 and 552.2). The measured levels of total THMs and HAAs were compared temporally and spatially with LRAA and QA data, which showed significant differences (p < 0.05). Most samples from surface water stations showed higher levels than LRAA or QA. Significant numbers of samples in surface water stations exceeded regulatory permissible limits: 27% had excessive THMs and 35% had excessive HAAs. Trichloromethane, trichloroacetic acid, and dichloroacetic acid were the major drivers of variability. This study suggests that LRAA and QA data are not good proxies of short-term exposure. Further investigation is needed to determine if other drinking water systems show consistent findings for improved regulation. PMID:28531123
Parvez, Shahid; Frost, Kali; Sundararajan, Madhura
2017-05-20
In the absence of shorter term disinfectant byproducts (DBPs) data on regulated Trihalomethanes (THMs) and Haloacetic acids (HAAs), epidemiologists and risk assessors have used long-term annual compliance (LRAA) or quarterly (QA) data to evaluate the association between DBP exposure and adverse birth outcomes, which resulted in inconclusive findings. Therefore, we evaluated the reliability of using long-term LRAA and QA data as an indirect measure for short-term exposure. Short-term residential tap water samples were collected in peak DBP months (May-August) in a community water system with five separate treatment stations and were sourced from surface or groundwater. Samples were analyzed for THMs and HAAs per the EPA (U.S. Environmental Protection Agency) standard methods (524.2 and 552.2). The measured levels of total THMs and HAAs were compared temporally and spatially with LRAA and QA data, which showed significant differences ( p < 0.05). Most samples from surface water stations showed higher levels than LRAA or QA. Significant numbers of samples in surface water stations exceeded regulatory permissible limits: 27% had excessive THMs and 35% had excessive HAAs. Trichloromethane, trichloroacetic acid, and dichloroacetic acid were the major drivers of variability. This study suggests that LRAA and QA data are not good proxies of short-term exposure. Further investigation is needed to determine if other drinking water systems show consistent findings for improved regulation.
Kong, Shaofei; Yan, Qin; Zheng, Huang; Liu, Haibiao; Wang, Wei; Zheng, Shurui; Yang, Guowei; Zheng, Mingming; Wu, Jian; Qi, Shihua; Shen, Guofeng; Tang, Lili; Yin, Yan; Zhao, Tianliang; Yu, Huan; Liu, Dantong; Zhao, Delong; Zhang, Tao; Ruan, Jujun; Huang, Mingzhi
2018-05-01
Under great efforts in fighting against serious haze problem of China since 2013, decreasing of air pollutants especially for fine particles (PM 2.5 ) has been revealed for several key regions. This study tried to answer whether the reduction of PM 2.5 -bound polycyclic aromatic hydrocarbons (PAHs) was coincident with PM 2.5 because of long-term pollution control measures (PCM), and to assess source-oriented health risks associated with inhalation exposure to PAHs. Field measurements were carried out before and after the publishing of local air pollution protection plan for Nanjing, a mega-city in east China. Results indicated that the air quality was substantially improving, with a significant reduction in annual average PM 2.5 by 34%, and moreover, PM 2.5 -bound PAHs significantly reduced by 63% (p < 0.001). The remarkable reduction was mainly attributable to the change of emission sources, compared to the influence of atmospheric circulation patterns, surface meteorological conditions, and atmospheric chemical reaction. Four PAHs sources including coal combustion (CC), petroleum and oil burning (PO), wood burning (WB) and vehicle emission (VE) were identified. On an annual basis, contributions to ambient PM 2.5 -PAHs from WB, PO, CC and VE sources in the period before the action of control measures were 2.26, 2.20, 1.96 and 5.62 ng m -3 , respectively. They reduced to 1.09, 0.37, 1.31 and 1.77 ng m -3 for the four source types, with the reduction percentages as 51, 83, 33 and 68%, respectively. The estimated reduction in lifetime lung cancer risk was around 61%. The study that firstly assessed the health effects of PAHs reduction as a co-benefit raised by air PCM sustained for a long period is believed to be applicable and referential for other mega-cities around the world for assessing the benefits of PCM. Copyright © 2018 Elsevier Ltd. All rights reserved.
NASA Technical Reports Server (NTRS)
Green, Del L.; Walker, Eric L.; Everhart, Joel L.
2006-01-01
Minimization of uncertainty is essential to extend the usable range of the 15-psid Electronically Scanned Pressure [ESP) transducer measurements to the low free-stream static pressures found in hypersonic wind tunnels. Statistical characterization of environmental error sources inducing much of this uncertainty requires a well defined and controlled calibration method. Employing such a controlled calibration system, several studies were conducted that provide quantitative information detailing the required controls needed to minimize environmental and human induced error sources. Results of temperature, environmental pressure, over-pressurization, and set point randomization studies for the 15-psid transducers are presented along with a comparison of two regression methods using data acquired with both 0.36-psid and 15-psid transducers. Together these results provide insight into procedural and environmental controls required for long term high-accuracy pressure measurements near 0.01 psia in the hypersonic testing environment using 15-psid ESP transducers.
NASA Technical Reports Server (NTRS)
Green, Del L.; Walker, Eric L.; Everhart, Joel L.
2006-01-01
Minimization of uncertainty is essential to extend the usable range of the 15-psid Electronically Scanned Pressure (ESP) transducer measurements to the low free-stream static pressures found in hypersonic wind tunnels. Statistical characterization of environmental error sources inducing much of this uncertainty requires a well defined and controlled calibration method. Employing such a controlled calibration system, several studies were conducted that provide quantitative information detailing the required controls needed to minimize environmental and human induced error sources. Results of temperature, environmental pressure, over-pressurization, and set point randomization studies for the 15-psid transducers are presented along with a comparison of two regression methods using data acquired with both 0.36-psid and 15-psid transducers. Together these results provide insight into procedural and environmental controls required for long term high-accuracy pressure measurements near 0.01 psia in the hypersonic testing environment using 15-psid ESP transducers.
Parworth, Caroline; Tilp, Alison; Fast, Jerome; ...
2015-04-01
In this study the long-term trends of non-refractory submicrometer aerosol (NR-PM1) composition and mass concentration measured by an Aerosol Chemical Speciation Monitor (ACSM) at the Atmospheric Radiation Measurement (ARM) program's Southern Great Plains (SGP) site are discussed. NR-PM1 data was recorded at ~30 min intervals over a period of 19 months between November 2010 and June 2012. Positive Matrix Factorization (PMF) was performed on the measured organic mass spectral matrix using a rolling window technique to derive factors associated with distinct sources, evolution processes, and physiochemical properties. The rolling window approach also allows us to capture the dynamic variations ofmore » the chemical properties in the organic aerosol (OA) factors over time. Three OA factors were obtained including two oxygenated OA (OOA) factors, differing in degrees of oxidation, and a biomass burning OA (BBOA) factor. Back trajectory analyses were performed to investigate possible sources of major NR-PM1 species at the SGP site. Organics dominated NR-PM1 mass concentration for the majority of the study with the exception of winter, when ammonium nitrate increases due to transport of precursor species from surrounding urban and agricultural areas and also due to cooler temperatures. Sulfate mass concentrations have little seasonal variation with mixed regional and local sources. In the spring BBOA emissions increase and are mainly associated with local fires. Isoprene and carbon monoxide emission rates were obtained by the Model of Emissions of Gases and Aerosols from Nature (MEGAN) and the 2011 U.S. National Emissions Inventory to represent the spatial distribution of biogenic and anthropogenic sources, respectively. The combined spatial distribution of isoprene emissions and air mass trajectories suggest that biogenic emissions from the southeast contribute to SOA formation at the SGP site during the summer.« less
Status of a standard for neutron skyshine calculation and measurement
DOE Office of Scientific and Technical Information (OSTI.GOV)
Westfall, R.M.; Wright, R.Q.; Greenborg, J.
1990-01-01
An effort has been under way for several years to prepare a draft standard, ANS-6.6.2, Calculation and Measurement of Direct and Scattered Neutron Radiation from Contained Sources Due to Nuclear Power Operations. At the outset, the work group adopted a three-phase study involving one-dimensional analyses, a measurements program, and multi-dimensional analyses. Of particular interest are the neutron radiation levels associated with dry-fuel storage at reactor sites. The need for dry storage has been investigated for various scenarios of repository and monitored retrievable storage (MRS) facilities availability with the waste stream analysis model. The concern is with long-term integrated, low-level dosesmore » at long distances from a multiplicity of sources. To evaluate the conservatism associated with one-dimensional analyses, the work group has specified a series of simple problems. Sources as a function of fuel exposure were determined for a Westinghouse 17 x 17 pressurized water reactor assembly with the ORIGEN-S module of the SCALE system. The energy degradation of the 35 GWd/ton U sources was determined for two generic designs of dry-fuel storage casks.« less
NASA Astrophysics Data System (ADS)
Shifflett, J. A.
2008-08-01
We modify the Einstein-Schrödinger theory to include a cosmological constant Λ z which multiplies the symmetric metric, and we show how the theory can be easily coupled to additional fields. The cosmological constant Λ z is assumed to be nearly cancelled by Schrödinger’s cosmological constant Λ b which multiplies the nonsymmetric fundamental tensor, such that the total Λ = Λ z + Λ b matches measurement. The resulting theory becomes exactly Einstein-Maxwell theory in the limit as | Λ z | → ∞. For | Λ z | ~ 1/(Planck length)2 the field equations match the ordinary Einstein and Maxwell equations except for extra terms which are < 10-16 of the usual terms for worst-case field strengths and rates-of-change accessible to measurement. Additional fields can be included in the Lagrangian, and these fields may couple to the symmetric metric and the electromagnetic vector potential, just as in Einstein-Maxwell theory. The ordinary Lorentz force equation is obtained by taking the divergence of the Einstein equations when sources are included. The Einstein-Infeld-Hoffmann (EIH) equations of motion match the equations of motion for Einstein-Maxwell theory to Newtonian/Coulombian order, which proves the existence of a Lorentz force without requiring sources. This fixes a problem of the original Einstein-Schrödinger theory, which failed to predict a Lorentz force. An exact charged solution matches the Reissner-Nordström solution except for additional terms which are ~10-66 of the usual terms for worst-case radii accessible to measurement. An exact electromagnetic plane-wave solution is identical to its counterpart in Einstein-Maxwell theory.
Polynomial-interpolation algorithm for van der Pauw Hall measurement in a metal hydride film
NASA Astrophysics Data System (ADS)
Koon, D. W.; Ares, J. R.; Leardini, F.; Fernández, J. F.; Ferrer, I. J.
2008-10-01
We apply a four-term polynomial-interpolation extension of the van der Pauw Hall measurement technique to a 330 nm Mg-Pd bilayer during both absorption and desorption of hydrogen at room temperature. We show that standard versions of the van der Pauw DC Hall measurement technique produce an error of over 100% due to a drifting offset signal and can lead to unphysical interpretations of the physical processes occurring in this film. The four-term technique effectively removes this source of error, even when the offset signal is drifting by an amount larger than the Hall signal in the time interval between successive measurements. This technique can be used to increase the resolution of transport studies of any material in which the resistivity is rapidly changing, particularly when the material is changing from metallic to insulating behavior.
The Denver Aerosol Sources and Health (DASH) Study: Overview and Early Findings
Vedal, S.; Hannigan, M.P.; Dutton, S.J.; Miller, S. L.; Milford, J.B.; Rabinovitch, N.; Kim, S.-Y.; Sheppard, L.
2012-01-01
Improved understanding of the sources of air pollution that are most harmful could aid in developing more effective measures for protecting human health. The Denver Aerosol Sources and Health (DASH) study was designed to identify the sources of ambient fine particulate matter (PM2.5) that are most responsible for the adverse health effects of short-term exposure to PM 2.5. Daily 24-hour PM2.5 sampling began in July 2002 at a residential monitoring site in Denver, Colorado, using both Teflon and quartz filter samplers. Sampling is planned to continue through 2008. Chemical speciation is being carried out for mass, inorganic ionic compounds (sulfate, nitrate and ammonium), and carbonaceous components, including elemental carbon, organic carbon, temperature-resolved organic carbon fractions and a large array of organic compounds. In addition, water soluble metals were measured daily for 12 months in 2003. A receptor-based source apportionment approach utilizing positive matrix factorization (PMF) will be used to identify PM 2.5 source contributions for each 24-hour period. Based on a preliminary assessment using synthetic data, the proposed source apportionment should be able to identify many important sources on a daily basis, including secondary ammonium nitrate and ammonium sulfate, diesel vehicle exhaust, road dust, wood combustion and vegetative debris. Meat cooking, gasoline vehicle exhaust and natural gas combustion were more challenging for PMF to accurately identify due to high detection limits for certain organic molecular marker compounds. Measurements of these compounds are being improved and supplemented with additional organic molecular marker compounds. The health study will investigate associations between daily source contributions and an array of health endpoints, including daily mortality and hospitalizations and measures of asthma control in asthmatic children. Findings from the DASH study, in addition to being of interest to policymakers, by identifying harmful PM2.5 sources may provide insights into mechanisms of PM effect. PMID:22723735
The Denver Aerosol Sources and Health (DASH) study: Overview and early findings
NASA Astrophysics Data System (ADS)
Vedal, S.; Hannigan, M. P.; Dutton, S. J.; Miller, S. L.; Milford, J. B.; Rabinovitch, N.; Kim, S.-Y.; Sheppard, L.
Improved understanding of the sources of air pollution that are most harmful could aid in developing more effective measures for protecting human health. The Denver Aerosol Sources and Health (DASH) study was designed to identify the sources of ambient fine particulate matter (PM 2.5) that are most responsible for the adverse health effects of short-term exposure to PM 2.5. Daily 24-h PM 2.5 sampling began in July 2002 at a residential monitoring site in Denver, Colorado, using both Teflon and quartz filter samplers. Sampling is planned to continue through 2008. Chemical speciation is being carried out for mass, inorganic ionic compounds (sulfate, nitrate and ammonium), and carbonaceous components, including elemental carbon, organic carbon, temperature-resolved organic carbon fractions and a large array of organic compounds. In addition, water-soluble metals were measured daily for 12 months in 2003. A receptor-based source apportionment approach utilizing positive matrix factorization (PMF) will be used to identify PM 2.5 source contributions for each 24-h period. Based on a preliminary assessment using synthetic data, the proposed source apportionment should be able to identify many important sources on a daily basis, including secondary ammonium nitrate and ammonium sulfate, diesel vehicle exhaust, road dust, wood combustion and vegetative debris. Meat cooking, gasoline vehicle exhaust and natural gas combustion were more challenging for PMF to accurately identify due to high detection limits for certain organic molecular marker compounds. Measurements of these compounds are being improved and supplemented with additional organic molecular marker compounds. The health study will investigate associations between daily source contributions and an array of health endpoints, including daily mortality and hospitalizations and measures of asthma control in asthmatic children. Findings from the DASH study, in addition to being of interest to policymakers, by identifying harmful PM 2.5 sources may provide insights into mechanisms of PM effect.
Danylov, A A; Light, A R; Waldman, J; Erickson, N
2015-12-10
Measurements of the frequency stability of a far-infrared molecular laser have been made by mixing the harmonic of an ultrastable microwave source with a portion of the laser output signal in a terahertz (THz) Schottky diode balanced mixer. A 3 GHz difference-frequency signal was used in a frequency discriminator circuit to lock the laser to the microwave source. Comparisons of the short- and long-term laser frequency stability under free-running and locked conditions show a significant improvement with locking. Short-term frequency jitter was reduced by an order of magnitude, from approximately 40 to 4 kHz, and long-term drift was reduced by more than three orders of magnitude, from approximately 250 kHz to 80 Hz. The results, enabled by the efficient Schottky diode balanced mixer downconverter, demonstrate that ultrastable microwave-based frequency stabilization of THz optically pumped lasers (OPLs) will now be possible at frequencies extending well above 4.0 THz.
Multi-Decadal Change of Atmospheric Aerosols and Their Effect on Surface Radiation
NASA Technical Reports Server (NTRS)
Chin, Mian; Diehl, Thomas; Tan, Qian; Wild, Martin; Qian, Yun; Yu, Hongbin; Bian, Huisheng; Wang, Weiguo
2012-01-01
We present an investigation on multi-decadal changes of atmospheric aerosols and their effects on surface radiation using a global chemistry transport model along with the near-term to long-term data records. We focus on a 28-year time period of satellite era from 1980 to 2007, during which a suite of aerosol data from satellite observations and ground-based remote sensing and in-situ measurements have become available. We analyze the long-term global and regional aerosol optical depth and concentration trends and their relationship to the changes of emissions" and assess the role aerosols play in the multi-decadal change of solar radiation reaching the surface (known as "dimming" or "brightening") at different regions of the world, including the major anthropogenic source regions (North America, Europe, Asia) that have been experiencing considerable changes of emissions, dust and biomass burning regions that have large interannual variabilities, downwind regions that are directly affected by the changes in the source area, and remote regions that are considered to representing "background" conditions.
Measurement of volatile organic chemicals at selected sites in California
NASA Technical Reports Server (NTRS)
Singh, Hanwant B.; Salas, L.; Viezee, W.; Sitton, B.; Ferek, R.
1992-01-01
Urban air concentrations of 24 selected volatile organic chemicals that may be potentially hazardous to human health and environment were measured during field experiments conducted at two California locations, at Houston, and at Denver. Chemicals measured included chlorofluorocarbons, halomethanes, haloethanes, halopropanes, chloroethylenes, and aromatic hydrocarbons. With emphasis on California sites, data from these studies are analyzed and interpreted with respect to variabilities in ambient air concentrations, diurnal changes, relation to prevailing meteorology, sources and trends. Except in a few instances, mean concentrations are typically between 0 and 5 ppb. Significant variabilities in atmospheric concentrations associated with intense sources and adverse meteorological conditions are shown to exist. In addition to short-term variability, there is evidence of systematic diurnal and seasonal trends. In some instances it is possible to detect declining trends resulting from the effectiveness of control strategies.
NASA Astrophysics Data System (ADS)
Chatfield, R. B.; Browell, E. V.; Brune, W. H.; Crawford, J. H.; Esswein, R.; Fried, A.; Olson, J. R.; Shetter, R. E.; Singh, H. B.
2006-12-01
We propose and evaluate two related and surprisingly simple empirical estimators for the local chemical production term for photochemical ozone; each uses two moderate-technology chemical measurements and a measurement of ultraviolet light. We nickname the techniques POGO-FAN: Production of Ozone by Gauging Oxidation: Formaldehyde and NO. (1) A non-linear function of a single three-factor index-variable, j (HCHO=>rads) [HCHO] [NO] seems to provide a good estimator of the largest single term in the production of smog ozone, the HOO+NO term, over a very wide range of situations. (2) By considering empirical contour plots summarizing isopleths of HOO+NO using j (HCHO=>rads) [HCHO] and [NO] separately as coordinates, we provide a slightly more complex 2-d indicator of smog ozone production that additionally allows an estimate of the NOx-sensitivity or NOx-saturation (i.e., VOC-sensitivity) of sampled air parcels. ~85 to >90 % of the variance is explained. The correspondence to "EKMA" contour plots, estimating afternoon ozone based on morningtime organics and NOx mixes, is not coincidental. We utilize a broad set of urban plume, regionally polluted and cleaner NASA DC-8 PBL samples from the Intercontinental Transport Experiment-North America (INTEX-NA), in which each of the variables was measured, to help establish our relationship. The estimator is described in terms both both of asymptotic smog photochemistry theory; primarily this suggests appropriate statistical approaches which can capture some of the complex interrelations of lower-tropospheric smog mix through correlation of reactive mixture components. HCHO is not only an important source of HOO radicals, but it more important serves as a "gauge" of all photochemical processing of volatile organic compounds. It probably captures information related to coincident VOC sources of various compounds and parallels in photochemical processing. Constrained modeling of observed atmospheric concentrations suggests that the prime source of ozone from HOO+NO reaction and other peroxy radical ozone formation reactions (ROO+NO), thus all ozone production, are closely related. Additionally, modeling allows us to follow ozone production and NOx-sensitivity throughout the varying photolytic cycle.
Measurement of attenuation coefficients of the fundamental and second harmonic waves in water
NASA Astrophysics Data System (ADS)
Zhang, Shuzeng; Jeong, Hyunjo; Cho, Sungjong; Li, Xiongbing
2016-02-01
Attenuation corrections in nonlinear acoustics play an important role in the study of nonlinear fluids, biomedical imaging, or solid material characterization. The measurement of attenuation coefficients in a nonlinear regime is not easy because they depend on the source pressure and requires accurate diffraction corrections. In this work, the attenuation coefficients of the fundamental and second harmonic waves which come from the absorption of water are measured in nonlinear ultrasonic experiments. Based on the quasilinear theory of the KZK equation, the nonlinear sound field equations are derived and the diffraction correction terms are extracted. The measured sound pressure amplitudes are adjusted first for diffraction corrections in order to reduce the impact on the measurement of attenuation coefficients from diffractions. The attenuation coefficients of the fundamental and second harmonics are calculated precisely from a nonlinear least squares curve-fitting process of the experiment data. The results show that attenuation coefficients in a nonlinear condition depend on both frequency and source pressure, which are much different from a linear regime. In a relatively lower drive pressure, the attenuation coefficients increase linearly with frequency. However, they present the characteristic of nonlinear growth in a high drive pressure. As the diffraction corrections are obtained based on the quasilinear theory, it is important to use an appropriate source pressure for accurate attenuation measurements.
77 FR 19740 - Water Sources for Long-Term Recirculation Cooling Following a Loss-of-Coolant Accident
Federal Register 2010, 2011, 2012, 2013, 2014
2012-04-02
... NUCLEAR REGULATORY COMMISSION [NRC-2010-0249] Water Sources for Long-Term Recirculation Cooling... Regulatory Guide (RG) 1.82, ``Water Sources for Long-Term Recirculation Cooling Following a Loss-of-Coolant... regarding the sumps and suppression pools that provide water sources for emergency core cooling, containment...
Unsupervised Segmentation of Head Tissues from Multi-modal MR Images for EEG Source Localization.
Mahmood, Qaiser; Chodorowski, Artur; Mehnert, Andrew; Gellermann, Johanna; Persson, Mikael
2015-08-01
In this paper, we present and evaluate an automatic unsupervised segmentation method, hierarchical segmentation approach (HSA)-Bayesian-based adaptive mean shift (BAMS), for use in the construction of a patient-specific head conductivity model for electroencephalography (EEG) source localization. It is based on a HSA and BAMS for segmenting the tissues from multi-modal magnetic resonance (MR) head images. The evaluation of the proposed method was done both directly in terms of segmentation accuracy and indirectly in terms of source localization accuracy. The direct evaluation was performed relative to a commonly used reference method brain extraction tool (BET)-FMRIB's automated segmentation tool (FAST) and four variants of the HSA using both synthetic data and real data from ten subjects. The synthetic data includes multiple realizations of four different noise levels and several realizations of typical noise with a 20% bias field level. The Dice index and Hausdorff distance were used to measure the segmentation accuracy. The indirect evaluation was performed relative to the reference method BET-FAST using synthetic two-dimensional (2D) multimodal magnetic resonance (MR) data with 3% noise and synthetic EEG (generated for a prescribed source). The source localization accuracy was determined in terms of localization error and relative error of potential. The experimental results demonstrate the efficacy of HSA-BAMS, its robustness to noise and the bias field, and that it provides better segmentation accuracy than the reference method and variants of the HSA. They also show that it leads to a more accurate localization accuracy than the commonly used reference method and suggest that it has potential as a surrogate for expert manual segmentation for the EEG source localization problem.
USDA-ARS?s Scientific Manuscript database
With over 4 million ha irrigated cropland, the Lower Mississippi River Valley (LMRV) is a highly productive agricultural region where irrigation practices are similar and the Mississippi River Valley alluvial aquifer (MRVA) is a primary source of on-demand irrigation. Owing to agricultural exports, ...
Classes of Legitimate Evidence for Identifying Effective Teaching.
ERIC Educational Resources Information Center
Wagner, Paul A.
A criterion for selecting sources of evidence to evaluate effective teaching is described. It is suggested that teaching effectiveness is not measured solely in terms of cognitive change in students but in the extent to which academics practice teaching in accordance with the moral dictates of the profession. In developing a teacher effectiveness…
ERIC Educational Resources Information Center
Troyer, Jennifer L.; Sause, Wendy L.
2011-01-01
Purpose of the Study: Two consumer-derived measures of nursing home quality that have been underutilized by researchers are consumer complaints to the state certification agency between inspections and complaints to the Long-Term Care Ombudsman Program. This article describes these complaints, considers facility-level predictors of complaints, and…
NASA Astrophysics Data System (ADS)
Zolfaghari, Abolfazl; Jeon, Seongkyul; Stepanick, Christopher K.; Lee, ChaBum
2017-06-01
This paper presents a novel method for measuring two-degree-of-freedom (DOF) motion of flexure-based nanopositioning systems based on optical knife-edge sensing (OKES) technology, which utilizes the interference of two superimposed waves: a geometrical wave from the primary source of light and a boundary diffraction wave from the secondary source. This technique allows for two-DOF motion measurement of the linear and pitch motions of nanopositioning systems. Two capacitive sensors (CSs) are used for a baseline comparison with the proposed sensor by simultaneously measuring the motions of the nanopositioning system. The experimental results show that the proposed sensor closely agrees with the fundamental linear motion of the CS. However, the two-DOF OKES technology was shown to be approximately three times more sensitive to the pitch motion than the CS. The discrepancy in the two sensor outputs is discussed in terms of measuring principle, linearity, bandwidth, control effectiveness, and resolution.
The HEMP QSO Monitoring Project
NASA Astrophysics Data System (ADS)
Welsh, William F.; Robinson, E. L.
2000-02-01
Many AGN are highly variable sources. Some of these show a pronounced time delay between variations seen in their optical continuum and in their emission lines. ``Echo mapping'' is a technique that uses these time delays to measure the geometry and kinematics of the gas inside the AGN, near the supermassive black hole. The technique is immensely powerful, but the results so far have been modest due to relatively low quality data. We have initiated a long--term project to echo map QSOs. We will examine nearby (but intrinsically faint) QSOs as well as QSOs at high redshift. The high--z QSOs present a problem: it is not known ahead of time which of these are variable sources. Thus we have started a campaign to monitor about 60 high-redshift QSOs for the purpose of determining their variability characteristics. We request SSTO time on the 0.9m telescope for long--term monitoring of high--redshift QSOs to: (i) test their suitability as viable echo mapping candidates; and (ii) measure (for the first time) their variability properties, which is of intrinsic value itself.
The Study of High-Speed Surface Dynamics Using a Pulsed Proton Beam
NASA Astrophysics Data System (ADS)
Buttler, William; Stone, Benjamin; Oro, David; Dimonte, Guy; Preston, Dean; Cherne, Frank; Germann, Timothy; Terrones, Guillermo; Tupa, Dale
2011-06-01
Los Alamos National Laboratory is presently engaged in development and implementation of ejecta source term and transport models for integration into LANL hydrodynamic computer codes. Experimental support for the effort spans a broad array of activities, including ejecta source term measurements from machine roughened Sn surfaces shocked by HE or flyer plates. Because the underlying postulate for ejecta formation is that ejecta are characterized by Richtmyer-Meshkov instability (RMI) phenomena, a key element of the theory and modeling effort centers on validation and verification RMI experiments at the LANSCE Proton Radiography Facility (pRad) to compare with modeled ejecta measurements. Here we present experimental results used to define and validate a physics based ejecta model together with remarkable, unexpected results of Sn instability growth in vacuum and gasses, and Sn and Cu RM growth that reveals the sensitivity of the RM instability to the yield strength of the material, Cu. The motivation of this last subject, RM growth linked to material strength, is to probe the shock pressure regions over which ejecta begins to form. Presenter
Periodic cycles of social outbursts of activity
NASA Astrophysics Data System (ADS)
Berestycki, H.; Rossi, L.; Rodríguez, N.
2018-01-01
We study the long-time behavior of a 2 × 2 continuous dynamical system with a time-periodic source term which is either of cooperative-type or activator-inhibitor type. This system was recently introduced in the literature [2] to model the dynamics of social outbursts and consists of an explicit field measuring the level of activity and an implicit field measuring the effective tension. The system can be used to represent a general type of phenomena in which one variable exhibits self-excitement once the other variable has reached a critical value. The time-periodic source term allows one to analyze the effect that periodic external shocks to the system play in the dynamics of the outburst of activity. For cooperative systems we prove that for small shocks the level of activity dies down whereas, as the intensity of the shocks increases, the level of activity converges to a positive periodic solution (excited cycle). We further show that in some cases there is multiplicity of excited cycles. We derive a subset of these results for the activator-inhibitor system.
NASA Astrophysics Data System (ADS)
Kryjevskaia, Mila; Stetzer, MacKenzie R.; Heron, Paula R. L.
2013-06-01
In a previous paper that focused on the transmission of periodic waves at the boundary between two media, we documented difficulties with the basic concepts of wavelength, frequency, and propagation speed, and with the relationship v=fλ. In this paper, we report on student attempts to apply this relationship in problems involving two-source and thin-film interference. In both cases, interference arises from differences in the path lengths traveled by two waves. We found that some students (up to 40% on certain questions) had difficulty with a task that is fundamental to understanding these phenomena: expressing a physical distance, such as the separation between two sources, in terms of the wavelength of a periodic wave. We administered a series of questions to try to identify factors that influence student performance. We concluded that most incorrect responses stemmed from erroneous judgment about the type of reasoning required, not an inability to do said reasoning. A number of students do not seem to treat the spacing of moving wave fronts as analogous to immutable measurement tools (e.g., rulers).
Multi-Detector Analysis System for Spent Nuclear Fuel Characterization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reber, Edward Lawrence; Aryaeinejad, Rahmat; Cole, Jerald Donald
1999-09-01
The Spent Nuclear Fuel (SNF) Non-Destructive Analysis (NDA) program at INEEL is developing a system to characterize SNF for fissile mass, radiation source term, and fissile isotopic content. The system is based on the integration of the Fission Assay Tomography System (FATS) and the Gamma-Neutron Analysis Technique (GNAT) developed under programs supported by the DOE Office of Non-proliferation and National Security. Both FATS and GNAT were developed as separate systems to provide information on the location of special nuclear material in weapons configuration (FATS role), and to measure isotopic ratios of fissile material to determine if the material was frommore » a weapon (GNAT role). FATS is capable of not only determining the presence and location of fissile material but also the quantity of fissile material present to within 50%. GNAT determines the ratios of the fissile and fissionable material by coincidence methods that allow the two prompt (immediately) produced fission fragments to be identified. Therefore, from the combination of FATS and GNAT, MDAS is able to measure the fissile material, radiation source term, and fissile isotopics content.« less
Botha, R; Newman, R T; Maleka, P P
2016-09-01
Radon activity concentrations (in water and in air) were measured at 13 selected locations at the Avalon Springs thermal spa resort in Montagu (Western Cape, South Africa) to estimate the associated effective dose received by employees and visitors. A RAD-7 detector (DURRIDGE), based on alpha spectrometry, and electret detectors (E-PERM®Radelec) were used for these radon measurements. The primary source of radon was natural thermal waters from the hot spring, which were pumped to various locations on the resort, and consequently a range of radon in-water analyses were performed. Radon in-water activity concentration as a function of time (short term and long term measurements) and spatial distributions (different bathing pools, etc.) were studied. The mean radon in-water activity concentrations were found to be 205 ± 6 Bq L (source), 112 ± 5 Bq L (outdoor pool) and 79 ± 4 Bq L (indoor pool). Radon in-air activity concentrations were found to range between 33 ± 4 Bq m (at the outside bar) to 523 ± 26 Bq m (building enclosing the hot spring's source). The most significant potential radiation exposure identified is that due to inhalation of air rich in radon and its progeny by the resort employees. The annual occupational effective dose due to the inhalation of radon progeny ranges from 0.16 ± 0.01 mSv to 0.40 ± 0.02 mSv. For the water samples collected, the Ra in-water activity concentrations from samples collected were below the lower detection limit (~0.7 Bq L) of the γ-ray detector system used. No significant radiological health risk can be associated with radon and progeny from the hot spring at the Avalon Springs resort.
Can control of soil erosion mitigate water pollution by sediments?
Rickson, R J
2014-01-15
The detrimental impact of sediment and associated pollutants on water quality is widely acknowledged, with many watercourses in the UK failing to meet the standard of 'good ecological status'. Catchment sediment budgets show that hill slope erosion processes can be significant sources of waterborne sediment, with rates of erosion likely to increase given predicted future weather patterns. However, linking on-site erosion rates with off-site impacts is complicated because of the limited data on soil erosion rates in the UK and the dynamic nature of the source-pathway-receptor continuum over space and time. Even so, soil erosion control measures are designed to reduce sediment production (source) and mobilisation/transport (pathway) on hill slopes, with consequent mitigation of pollution incidents in watercourses (receptors). The purpose of this paper is to review the scientific evidence of the effectiveness of erosion control measures used in the UK to reduce sediment loads of hill slope origin in watercourses. Although over 73 soil erosion mitigation measures have been identified from the literature, empirical data on erosion control effectiveness are limited. Baseline comparisons for the 18 measures where data do exist reveal erosion control effectiveness is highly variable over time and between study locations. Given the limitations of the evidence base in terms of geographical coverage and duration of monitoring, performance of the different measures cannot be extrapolated to other areas. This uncertainty in effectiveness has implications for implementing erosion/sediment risk reduction policies, where quantified targets are stipulated, as is the case in the EU Freshwater Fish and draft Soil Framework Directives. Also, demonstrating technical effectiveness of erosion control measures alone will not encourage uptake by land managers: quantifying the costs and benefits of adopting erosion mitigation is equally important, but these are uncertain and difficult to express in monetary terms. Copyright © 2013 Elsevier B.V. All rights reserved.
Rosmalen, Judith G M; Kema, Ido P; Wüst, Stefan; van der Ley, Claude; Visser, Sipke T; Snieder, Harold; Bakker, Stephan J L
2014-09-01
Function of the hypothalamus-pituitary-adrenal (HPA) axis has been associated with several somatic and psychiatric health problems. The amount of free cortisol excreted in the urine during 24h (24-h UFC) has often been used as a proxy for HPA-axis function. Reference values for 24-h UFC and their stability in the short and long term, as well as sources of variability, are largely lacking. This study was performed in a general population cohort. Participants collected 24-h UFC on two consecutive days (T1), and repeated this collection approximately 2 years later (T2). Cortisol in urine was measured using LC-MS/MS. Height and weight were measured at the research facilities; glomerular filtration rate was estimated using creatinine clearance. Psychological distress (General Health Questionnaire), smoking, alcohol use and exercise were measured by means of questionnaires. 24-h UFC stability on a day-to-day basis was 0.69 (T1, N=1192) and 0.72 (T2, N=963) (both p<0.001). Long-term stability as indicated by correlation between 2-day averages of T1 and T2 was 0.60 (N=972, p<0.001). Multivariable linear regression analysis revealed that 24-h UFC was predicted by urine volume (standardized beta 0.282 (T1, N=1556) and 0.276 (T2, N=1244); both p<0.001) and glomerular filtration rate (standardized beta 0.137 (T1) and 0.179 (T2); both p<0.001), while also sex explained a small part (standardized beta for female sex -0.057 (T1) and -0.080 (T2); both p<0.05). 24-h UFC is moderately stable both in the short and the long term. The effects of urine volume and glomerular filtration rate on 24-h UFC are much stronger than those of sex. Copyright © 2014 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Franssens, Matthias; Flament, Pascal; Deboudt, Karine; Weis, Dominique; Perdrix, Espéranza
2004-09-01
To demonstrate the ability of the lead isotope signature technique to evidence the spatial extent of an industrial Pb deposition plume at a local scale, dry deposition of lead in the urban environment of a Pb-Zn refinery was investigated, as a study case, using transient ("short-lived") isotopic signatures of the industrial source. Sampling campaigns were achieved in representative weather conditions, on an 8-h basis. Dry deposition rates measured downwind from refinery emissions (≈102-103 μg Pb m-2 h-1), cross-sectionally in a 3-5 km radius area around the plant, represent 10-100 times the urban background dry fallout, measured upwind, as well as fallout measured near other potential sources of anthropogenic Pb. The Pb-Zn refinery isotopic signature (approx. 1.100<206Pb/207Pb<1.135) is made identifiable, using the same set of Pb and Zn ores for 2 days before sampling and during field experiments, by agreement with the executive staff of the plant. This source signature is less radiogenic than signatures of urban background Pb aerosols (1.155<206Pb/207Pb<1.165) and minor sources of Pb aerosols (1.147<206Pb/207Pb<1.165). By a simple binary mixing model calculation, we established the extension of the industrial Pb deposition plume. Fifty to eighty percents of total lead settled by the dry deposition mode, 3-4 km away from the refinery, still have an industrial origin. That represents from 40 to 80 μg Pb m-2 h-1, in an area where the blood lead level exceeds 100 μg Pb l-1 for 30% of men and 12% of women living there. We demonstrate here that stable Pb isotope analysis is able to evidence the Pb dry deposition plume in stabilised aerodynamic conditions, using a short-lived source term, suggesting that this methodology is able to furnish valuable data to validate industrial Pb aerosols dispersion models, at the urban scale.
IRAS low resolution spectra of 26 symbiotic stars
NASA Technical Reports Server (NTRS)
Stencel, Robert E.; Brugel, Edward W.; Goodwill, Michael E.
1990-01-01
Data related to the spectral scans for 26 symbiotic stars are described which were extracted from the IRAS low resolution database. Data from the 8-15- and 15-23-micron bands are merged in a program that scales the longer wavelength and produces a weighted average of the spectral scans for each source. The survey shows that active dust producers can probably be isolated and some theories related to the presence of dust emission features are discussed in terms of source variability for measurements made with low resolution spectra.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yang, Aileen, E-mail: Yang@uu.nl; Institute for Risk Assessment Sciences, Division Environmental Epidemiology, Utrecht University, P.O. Box 80.178, 3508TD Utrecht; Hoek, Gerard
Oxidative potential (OP) of ambient particulate matter (PM) has been suggested as a health-relevant exposure metric. In order to use OP for exposure assessment, information is needed about how well central site OP measurements and modeled average OP at the home address reflect temporal and spatial variation of personal OP. We collected 96-hour personal, home outdoor and indoor PM{sub 2.5} samples from 15 volunteers living either at traffic, urban or regional background locations in Utrecht, the Netherlands. OP was also measured at one central reference site to account for temporal variations. OP was assessed using electron spin resonance (OP{sup ESR})more » and dithiothreitol (OP{sup DTT}). Spatial variation of average OP at the home address was modeled using land use regression (LUR) models. For both OP{sup ESR} and OP{sup DTT}, temporal correlations of central site measurements with home outdoor measurements were high (R>0.75), and moderate to high (R=0.49–0.70) with personal measurements. The LUR model predictions for OP correlated significantly with the home outdoor concentrations for OP{sup DTT} and OP{sup ESR} (R=0.65 and 0.62, respectively). LUR model predictions were moderately correlated with personal OP{sup DTT} measurements (R=0.50). Adjustment for indoor sources, such as vacuum cleaning and absence of fume-hood, improved the temporal and spatial agreement with measured personal exposure for OP{sup ESR}. OP{sup DTT} was not associated with any indoor sources. Our study results support the use of central site OP for exposure assessment of epidemiological studies focusing on short-term health effects. - Highlights: • Oxidative potential (OP) of PM was proposed as a health-relevant exposure metric. • We evaluated the relationship between measured and modeled outdoor and personal OP. • Temporal correlations of central site with personal OP are moderate to high. • Adjusting for indoor sources improved the agreement with personal OP. • Our results support the use of central site OP for short-term health effect studies.« less
NASA Astrophysics Data System (ADS)
Becker, S.; Halsall, C. J.; Tych, W.; Kallenborn, R.; Schlabach, M.; Manø, S.
2009-01-01
An extensive database of organochlorine (OC) pesticide concentrations measured at the Norwegian Arctic Monitoring Station was analysed to assess longer-term trends in the Arctic atmosphere. Dynamic Harmonic Regression (DHR) is employed to investigate the seasonal and cyclical behaviour of chlordanes, DDTs and hexachlorobenzene (HCB), and to isolate underlying inter-annual trends. Although a simple comparison of annual mean concentrations (1994-2005) suggest a decline for all of the OCs investigated, the longer-term trends identified by DHR only show a significant decline for p,p'-DDT. Indeed, HCB shows an increase from 2003-2005. This is thought to be due to changes in source types and the presence of impurities in current use pesticides, together with retreating sea ice affecting air-water exchange. Changes in source types were revealed by using isomeric ratios for the chlordanes and DDTs. Declining trends in ratios of trans-chlordane/cis-chlordane (TC/CC) indicate a shift from primary sources, to more ''weathered'' secondary sources, whereas an increasing trend in o,p'-DDT/p,p'-DDT ratios indicate a shift from use of technical DDT to dicofol. Continued monitoring of these OC pesticides is required to fully understand the influence of a changing climate on the behaviour and environmental cycling of these chemicals in the Arctic as well as possible impacts from ''new'' sources.
NASA Astrophysics Data System (ADS)
Becker, S.; Halsall, C. J.; Tych, W.; Kallenborn, R.; Schlabach, M.; Manø, S.
2012-05-01
An extensive database of organochlorine (OC) pesticide concentrations measured at the Norwegian Arctic monitoring station at Ny-Ålesund, Svalbard, was analysed to assess longer-term trends in the Arctic atmosphere. Dynamic Harmonic Regression (DHR) is employed to investigate the seasonal and cyclical behaviour of chlordanes, DDTs and hexachlorobenzene (HCB), and to isolate underlying inter-annual trends. Although a simple comparison of annual mean concentrations (1994-2005) suggest a decline for all of the OCs investigated, the longer-term trends identified by DHR only show a significant decline for p,p'-DDT. Indeed, HCB shows an increase from 2003-2005. This is thought to be due to changes in source types and the presence of impurities in current use pesticides, together with retreating sea ice affecting air-water exchange. Changes in source types were revealed by using isomeric ratios for the chlordanes and DDTs. Declining trends in ratios of trans-chlordane/cis-chlordane (TC/CC) indicate a shift from primary sources, to more "weathered" secondary sources, whereas an increasing trend in o,p'-DDT/p,p'-DDT ratios indicate a shift from use of technical DDT to dicofol. Continued monitoring of these OC pesticides is required to fully understand the influence of a changing climate on the behaviour and environmental cycling of these chemicals in the Arctic as well as possible impacts from "new" sources.
NASA Astrophysics Data System (ADS)
Park, Junghyun; Hayward, Chris; Stump, Brian W.
2018-06-01
Ground truth sources in Utah during 2003-2013 are used to assess the contribution of temporal atmospheric conditions to infrasound detection and the predictive capabilities of atmospheric models. Ground truth sources consist of 28 long duration static rocket motor burn tests and 28 impulsive rocket body demolitions. Automated infrasound detections from a hybrid of regional seismometers and infrasound arrays use a combination of short-term time average/long-term time average ratios and spectral analyses. These detections are grouped into station triads using a Delaunay triangulation network and then associated to estimate phase velocity and azimuth to filter signals associated with a particular source location. The resulting range and azimuth distribution from sources to detecting stations varies seasonally and is consistent with predictions based on seasonal atmospheric models. Impulsive signals from rocket body detonations are observed at greater distances (>700 km) than the extended duration signals generated by the rocket burn test (up to 600 km). Infrasound energy attenuation associated with the two source types is quantified as a function of range and azimuth from infrasound amplitude measurements. Ray-tracing results using Ground-to-Space atmospheric specifications are compared to these observations and illustrate the degree to which the time variations in characteristics of the observations can be predicted over a multiple year time period.
Radiological analysis of plutonium glass batches with natural/enriched boron
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rainisch, R.
2000-06-22
The disposition of surplus plutonium inventories by the US Department of Energy (DOE) includes the immobilization of certain plutonium materials in a borosilicate glass matrix, also referred to as vitrification. This paper addresses source terms of plutonium masses immobilized in a borosilicate glass matrix where the glass components include both natural boron and enriched boron. The calculated source terms pertain to neutron and gamma source strength (particles per second), and source spectrum changes. The calculated source terms corresponding to natural boron and enriched boron are compared to determine the benefits (decrease in radiation source terms) for to the use ofmore » enriched boron. The analysis of plutonium glass source terms shows that a large component of the neutron source terms is due to (a, n) reactions. The Americium-241 and plutonium present in the glass emit alpha particles (a). These alpha particles interact with low-Z nuclides like B-11, B-10, and O-17 in the glass to produce neutrons. The low-Z nuclides are referred to as target particles. The reference glass contains 9.4 wt percent B{sub 2}O{sub 3}. Boron-11 was found to strongly support the (a, n) reactions in the glass matrix. B-11 has a natural abundance of over 80 percent. The (a, n) reaction rates for B-10 are lower than for B-11 and the analysis shows that the plutonium glass neutron source terms can be reduced by artificially enriching natural boron with B-10. The natural abundance of B-10 is 19.9 percent. Boron enriched to 96-wt percent B-10 or above can be obtained commercially. Since lower source terms imply lower dose rates to radiation workers handling the plutonium glass materials, it is important to know the achievable decrease in source terms as a result of boron enrichment. Plutonium materials are normally handled in glove boxes with shielded glass windows and the work entails both extremity and whole-body exposures. Lowering the source terms of the plutonium batches will make the handling of these materials less difficult and will reduce radiation exposure to operating workers.« less
Public health interventions and behaviour change: reviewing the grey literature.
Franks, H; Hardiker, N R; McGrath, M; McQuarrie, C
2012-01-01
This study identified and reviewed grey literature relating to factors facilitating and inhibiting effective interventions in three areas: the promotion of mental health and well-being, the improvement of food and nutrition, and interventions seeking to increase engagement in physical activity. Sourcing, reviewing and analysis of relevant grey literature. Evidence was collected from a variety of non-traditional sources. Thirty-six pieces of documentary evidence across the three areas were selected for in-depth appraisal and review. A variety of approaches, often short-term, were used both as interventions and outcome measures. Interventions tended to have common outcomes, enabling the identification of themes. These included improvements in participant well-being as well as identification of barriers to, and promoters of, success. Most interventions demonstrated some positive impact, although some did not. This was particularly the case for more objective measures of change, such as physiological measurements, particularly when used to evaluate short-term interventions. Objective health measurement as part of an intervention may act as a catalyst for future behaviour change. Time is an important factor that could either promote or impede the success of interventions for both participants and facilitators. Likewise, the importance of involving all stakeholders, including participants, when planning health promoting interventions was established as an important indicator of success. Despite its limited scope, this review suggests that interventions can be more efficient and effective. For example, larger-scale, longer-term interventions could be more efficient, whilst outcomes relating to the implementation and beyond could provide a clearer picture of effectiveness. Additionally, interventions and evaluations must be flexible, evolve in partnership with local communities, and reflect local need and context. Copyright © 2011 The Royal Society for Public Health. Published by Elsevier Ltd. All rights reserved.
Conceptual measurement framework for help-seeking for mental health problems
Rickwood, Debra; Thomas, Kerry
2012-01-01
Background Despite a high level of research, policy, and practice interest in help-seeking for mental health problems and mental disorders, there is currently no agreed and commonly used definition or conceptual measurement framework for help-seeking. Methods A systematic review of research activity in the field was undertaken to investigate how help-seeking has been conceptualized and measured. Common elements were used to develop a proposed conceptual measurement framework. Results The database search revealed a very high level of research activity and confirmed that there is no commonly applied definition of help-seeking and no psychometrically sound measures that are routinely used. The most common element in the help-seeking research was a focus on formal help-seeking sources, rather than informal sources, although studies did not assess a consistent set of professional sources; rather, each study addressed an idiosyncratic range of sources of professional health and community care. Similarly, the studies considered help-seeking for a range of mental health problems and no consistent terminology was applied. The most common mental health problem investigated was depression, followed by use of generic terms, such as mental health problem, psychological distress, or emotional problem. Major gaps in the consistent measurement of help-seeking were identified. Conclusion It is evident that an agreed definition that supports the comparable measurement of help-seeking is lacking. Therefore, a conceptual measurement framework is proposed to fill this gap. The framework maintains that the essential elements for measurement are: the part of the help-seeking process to be investigated and respective time frame, the source and type of assistance, and the type of mental health concern. It is argued that adopting this framework will facilitate progress in the field by providing much needed conceptual consistency. Results will then be able to be compared across studies and population groups, and this will significantly benefit understanding of policy and practice initiatives aimed at improving access to and engagement with services for people with mental health concerns. PMID:23248576
Energy Spectra of Abundant Cosmic-ray Nuclei in Sources, According to the ATIC Experiment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Panov, A. D.; Sokolskaya, N. V.; Zatsepin, V. I., E-mail: panov@dec1.sinp.msu.ru
One of the main results of the ATIC (Advanced Thin Ionization Calorimeter) experiment is a collection of energy spectra of abundant cosmic-ray nuclei: protons, He, C, O, Ne, Mg, Si, and Fe measured in terms of energy per particle in the energy range from 50 GeV to tens of teraelectronvolts. In this paper, the ATIC energy spectra of abundant primary nuclei are back-propagated to the spectra in sources in terms of magnetic rigidity using a leaky-box approximation of three different GALPROP-based diffusion models of propagation that fit the latest B/C data of the AMS-02 experiment. It is shown that themore » results of a comparison of the slopes of the spectra in sources are weakly model dependent; therefore the differences of spectral indices are reliable data. A regular growth of the steepness of spectra in sources in the range of magnetic rigidity of 50–1350 GV is found for a charge range from helium to iron. This conclusion is statistically reliable with significance better than 3.2 standard deviations. The results are discussed and compared to the data of other modern experiments.« less
ISS Ambient Air Quality: Updated Inventory of Known Aerosol Sources
NASA Technical Reports Server (NTRS)
Meyer, Marit
2014-01-01
Spacecraft cabin air quality is of fundamental importance to crew health, with concerns encompassing both gaseous contaminants and particulate matter. Little opportunity exists for direct measurement of aerosol concentrations on the International Space Station (ISS), however, an aerosol source model was developed for the purpose of filtration and ventilation systems design. This model has successfully been applied, however, since the initial effort, an increase in the number of crewmembers from 3 to 6 and new processes on board the ISS necessitate an updated aerosol inventory to accurately reflect the current ambient aerosol conditions. Results from recent analyses of dust samples from ISS, combined with a literature review provide new predicted aerosol emission rates in terms of size-segregated mass and number concentration. Some new aerosol sources have been considered and added to the existing array of materials. The goal of this work is to provide updated filtration model inputs which can verify that the current ISS filtration system is adequate and filter lifetime targets are met. This inventory of aerosol sources is applicable to other spacecraft, and becomes more important as NASA considers future long term exploration missions, which will preclude the opportunity for resupply of filtration products.
NASA Astrophysics Data System (ADS)
Perez, Pedro B.; Hamawi, John N.
2017-09-01
Nuclear power plant radiation protection design features are based on radionuclide source terms derived from conservative assumptions that envelope expected operating experience. Two parameters that significantly affect the radionuclide concentrations in the source term are failed fuel fraction and effective fission product appearance rate coefficients. Failed fuel fraction may be a regulatory based assumption such as in the U.S. Appearance rate coefficients are not specified in regulatory requirements, but have been referenced to experimental data that is over 50 years old. No doubt the source terms are conservative as demonstrated by operating experience that has included failed fuel, but it may be too conservative leading to over-designed shielding for normal operations as an example. Design basis source term methodologies for normal operations had not advanced until EPRI published in 2015 an updated ANSI/ANS 18.1 source term basis document. Our paper revisits the fission product appearance rate coefficients as applied in the derivation source terms following the original U.S. NRC NUREG-0017 methodology. New coefficients have been calculated based on recent EPRI results which demonstrate the conservatism in nuclear power plant shielding design.
NASA Astrophysics Data System (ADS)
Zhou, X.; Albertson, J. D.
2016-12-01
Natural gas is considered as a bridge fuel towards clean energy due to its potential lower greenhouse gas emission comparing with other fossil fuels. Despite numerous efforts, an efficient and cost-effective approach to monitor fugitive methane emissions along the natural gas production-supply chain has not been developed yet. Recently, mobile methane measurement has been introduced which applies a Bayesian approach to probabilistically infer methane emission rates and update estimates recursively when new measurements become available. However, the likelihood function, especially the error term which determines the shape of the estimate uncertainty, is not rigorously defined and evaluated with field data. To address this issue, we performed a series of near-source (< 30 m) controlled methane release experiments using a specialized vehicle mounted with fast response methane analyzers and a GPS unit. Methane concentrations were measured at two different heights along mobile traversals downwind of the sources, and concurrent wind and temperature data are recorded by nearby 3-D sonic anemometers. With known methane release rates, the measurements were used to determine the functional form and the parameterization of the likelihood function in the Bayesian inference scheme under different meteorological conditions.
French, N P; Clancy, D; Davison, H C; Trees, A J
1999-10-01
The transmission and control of Neospora caninum infection in dairy cattle was examined using deterministic and stochastic models. Parameter estimates were derived from recent studies conducted in the UK and from the published literature. Three routes of transmission were considered: maternal vertical transmission with a high probability (0.95), horizontal transmission from infected cattle within the herd, and horizontal transmission from an independent external source. Putative infection via pooled colostrum was used as an example of within-herd horizontal transmission, and the recent finding that the dog is a definitive host of N. caninum supported the inclusion of an external independent source of infection. The predicted amount of horizontal transmission required to maintain infection at levels commonly observed in field studies in the UK and elsewhere, was consistent with that observed in studies of post-natal seroconversion (0.85-9.0 per 100 cow-years). A stochastic version of the model was used to simulate the spread of infection in herds of 100 cattle, with a mean infection prevalence similar to that observed in UK studies (around 20%). The distributions of infected and uninfected cattle corresponded closely to Normal distributions, with S.D.s of 6.3 and 7.0, respectively. Control measures were considered by altering birth, death and horizontal transmission parameters. A policy of annual culling of infected cattle very rapidly reduced the prevalence of infection, and was shown to be the most effective method of control in the short term. Not breeding replacements from infected cattle was also effective in the short term, particularly in herds with a higher turnover of cattle. However, the long-term effectiveness of these measures depended on the amount and source of horizontal infection. If the level of within-herd transmission was above a critical threshold, then a combination of reducing within-herd, and blocking external sources of transmission was required to permanently eliminate infection.
NASA Astrophysics Data System (ADS)
Petit, J.-E.; Favez, O.; Sciare, J.; Crenn, V.; Sarda-Estève, R.; Bonnaire, N.; Močnik, G.; Dupont, J.-C.; Haeffelin, M.; Leoz-Garziandia, E.
2015-03-01
Aerosol mass spectrometer (AMS) measurements have been successfully used towards a better understanding of non-refractory submicron (PM1) aerosol chemical properties based on short-term campaigns. The recently developed Aerosol Chemical Speciation Monitor (ACSM) has been designed to deliver quite similar artifact-free chemical information but for low cost, and to perform robust monitoring over long-term periods. When deployed in parallel with real-time black carbon (BC) measurements, the combined data set allows for a quasi-comprehensive description of the whole PM1 fraction in near real time. Here we present 2-year long ACSM and BC data sets, between mid-2011 and mid-2013, obtained at the French atmospheric SIRTA supersite that is representative of background PM levels of the region of Paris. This large data set shows intense and time-limited (a few hours) pollution events observed during wintertime in the region of Paris, pointing to local carbonaceous emissions (mainly combustion sources). A non-parametric wind regression analysis was performed on this 2-year data set for the major PM1 constituents (organic matter, nitrate, sulfate and source apportioned BC) and ammonia in order to better refine their geographical origins and assess local/regional/advected contributions whose information is mandatory for efficient mitigation strategies. While ammonium sulfate typically shows a clear advected pattern, ammonium nitrate partially displays a similar feature, but, less expectedly, it also exhibits a significant contribution of regional and local emissions. The contribution of regional background organic aerosols (OA) is significant in spring and summer, while a more pronounced local origin is evidenced during wintertime, whose pattern is also observed for BC originating from domestic wood burning. Using time-resolved ACSM and BC information, seasonally differentiated weekly diurnal profiles of these constituents were investigated and helped to identify the main parameters controlling their temporal variations (sources, meteorological parameters). Finally, a careful investigation of all the major pollution episodes observed over the region of Paris between 2011 and 2013 was performed and classified in terms of chemical composition and the BC-to-sulfate ratio used here as a proxy of the local/regional/advected contribution of PM. In conclusion, these first 2-year quality-controlled measurements of ACSM clearly demonstrate their great potential to monitor on a long-term basis aerosol sources and their geographical origin and provide strategic information in near real time during pollution episodes. They also support the capacity of the ACSM to be proposed as a robust and credible alternative to filter-based sampling techniques for long-term monitoring strategies.
McLerran, Larry; Skokov, Vladimir V.
2016-09-19
We modify the McLerran–Venugopalan model to include only a finite number of sources of color charge. In the effective action for such a system of a finite number of sources, there is a point-like interaction and a Coulombic interaction. The point interaction generates the standard fluctuation term in the McLerran–Venugopalan model. The Coulomb interaction generates the charge screening originating from well known evolution in x. Such a model may be useful for computing angular harmonics of flow measured in high energy hadron collisions for small systems. In this study we provide a basic formulation of the problem on a lattice.
Neutron crosstalk between liquid scintillators
DOE Office of Scientific and Technical Information (OSTI.GOV)
Verbeke, J. M.; Prasad, M. K.; Snyderman, N. J.
2015-05-01
We propose a method to quantify the fractions of neutrons scattering between liquid scintillators. Using a spontaneous fission source, this method can be utilized to quickly characterize an array of liquid scintillators in terms of crosstalk. The point model theory due to Feynman is corrected to account for these multiple scatterings. Using spectral information measured by the liquid scintillators, fractions of multiple scattering can be estimated, and mass reconstruction of fissile materials under investigation can be improved. Monte Carlo simulations of mono-energetic neutron sources were performed to estimate neutron crosstalk. A californium source in an array of liquid scintillators wasmore » modeled to illustrate the improvement of the mass reconstruction.« less
NASA Astrophysics Data System (ADS)
Jiang, Daijun; Li, Zhiyuan; Liu, Yikan; Yamamoto, Masahiro
2017-05-01
In this paper, we first establish a weak unique continuation property for time-fractional diffusion-advection equations. The proof is mainly based on the Laplace transform and the unique continuation properties for elliptic and parabolic equations. The result is weaker than its parabolic counterpart in the sense that we additionally impose the homogeneous boundary condition. As a direct application, we prove the uniqueness for an inverse problem on determining the spatial component in the source term by interior measurements. Numerically, we reformulate our inverse source problem as an optimization problem, and propose an iterative thresholding algorithm. Finally, several numerical experiments are presented to show the accuracy and efficiency of the algorithm.
The SPES surface ionization source
NASA Astrophysics Data System (ADS)
Manzolaro, M.; D'Agostini, F.; Monetti, A.; Andrighetto, A.
2017-09-01
Ion sources and target systems play a crucial role in isotope separation on line facilities, determining the main characteristics of the radioactive ion beams available for experiments. In the context of the selective production of exotic species (SPES) facility, a 40 MeV, 200 μA proton beam directly impinges a uranium carbide target, generating approximately 1013 fissions per second. The radioactive isotopes produced by the 238U fissions are delivered to the 1+ ion source by means of a tubular transfer line. Here they can be ionized and subsequently accelerated toward the experimental areas. In this work, the characterization of the surface ionization source currently adopted for the SPES facility is presented, taking as a reference ionization efficiency and transversal emittance measurements. The effects of long term operation at high temperature are also illustrated and discussed.
Measurement of erosion in helicon plasma thrusters using the VASIMR® VX-CR device
NASA Astrophysics Data System (ADS)
Del Valle Gamboa, Juan Ignacio; Castro-Nieto, Jose; Squire, Jared; Carter, Mark; Chang-Diaz, Franklin
2015-09-01
The helicon plasma source is one of the principal stages of the high-power VASIMR® electric propulsion system. The VASIMR® VX-CR experiment focuses solely on this stage, exploring the erosion and long-term operation effects of the VASIMR helicon source. We report on the design and operational parameters of the VX-CR experiment, and the development of modeling tools and characterization techniques allowing the study of erosion phenomena in helicon plasma sources in general, and stand-alone helicon plasma thrusters (HPTs) in particular. A thorough understanding of the erosion phenomena within HPTs will enable better predictions of their behavior as well as more accurate estimations of their expected lifetime. We present a simplified model of the plasma-wall interactions within HPTs based on current models of the plasma density distributions in helicon discharges. Results from this modeling tool are used to predict the erosion within the plasma-facing components of the VX-CR device. Experimental techniques to measure actual erosion, including the use of coordinate-measuring machines and microscopy, will be discussed.
NASA Astrophysics Data System (ADS)
Ringenberg, Hunter; Rogers, Dylan; Wei, Nathaniel; Krane, Michael; Wei, Timothy
2017-11-01
The objective of this study is to apply experimental data to theoretical framework of Krane (2013) in which the principal aeroacoustic source is expressed in terms of vocal fold drag, glottal jet dynamic head, and glottal exit volume flow, reconciling formal theoretical aeroacoustic descriptions of phonation with more traditional lumped-element descriptions. These quantities appear in the integral equations of motion for phonatory flow. In this way time resolved velocity field measurements can be used to compute time-resolved estimates of the relevant terms in the integral equations of motion, including phonation aeroacoustic source strength. A simplified 10x scale vocal fold model from Krane, et al. (2007) was used to examine symmetric, i.e. `healthy', oscillatory motion of the vocal folds. By using water as the working fluid, very high spatial and temporal resolution was achieved. Temporal variation of transglottal pressure was simultaneously measured with flow on the vocal fold model mid-height. Experiments were dynamically scaled to examine a range of frequencies corresponding to male and female voice. The simultaneity of the pressure and flow provides new insights into the aeroacoustics associated with vocal fold oscillations. Supported by NIH Grant No. 2R01 DC005642-11.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Evtushenko, Pavel E.; Klopf, John M.
Frequency domain measurements with Martin-Puplett interferometer is one of a few techniques capable of bunch length measurements at the level of ~ 100 fs. As the bunch length becomes shorter, it is important to know and be able to measure the limitations of the instrument in terms of shortest measurable bunch length. In this paper we describe an experiment using a blackbody source with the modified Martin-Puplett interferometer that is routine- ly used for bunch length measurements at the JLab FEL, as a way to estimate the shortest, measurable bunch length. The limitation comes from high frequency cut-off of themore » wire-grid polarizer currently used and is estimated to be 50 fs RMS. The measurements are made with the same Golay cell detector that is used for beam measure- ments. We demonstrate that, even though the blackbody source is many orders of magnitude less bright than the coherent transition or synchrotron radiation, it can be used for the measurements and gives a very good signal to noise ratio in combination with lock-in detection. We also compare the measurements made in air and in vacuum to characterize the very strong effect of the atmospheric absorption.« less
Measures and Relative Motions of Some Mostly F. G. W. Struve Doubles
NASA Astrophysics Data System (ADS)
Wiley, E. O.
2012-04-01
Measures of 59 pairs of double stars with long observational histories using "lucky imaging" techniques are reported. Relative motions of 59 pairs are investigated using histories of observation, scatter plots of relative motion, ordinary least-squares (OLS) and total proper motion analyses performed in "R," an open source programming language. A scatter plot of the coefficient of determinations derived from the OLS y|epoch and OLS x|epoch clearly separates common proper motion pairs from optical pairs and what are termed "long-period binary candidates." Differences in proper motion separate optical pairs from long-term binary candidates. An Appendix is provided that details how to use known rectilinear pairs as calibration pairs for the program REDUC.
Minimal-Drift Heading Measurement using a MEMS Gyro for Indoor Mobile Robots.
Hong, Sung Kyung; Park, Sungsu
2008-11-17
To meet the challenges of making low-cost MEMS yaw rate gyros for the precise self-localization of indoor mobile robots, this paper examines a practical and effective method of minimizing drift on the heading angle that relies solely on integration of rate signals from a gyro. The main idea of the proposed approach is consists of two parts; 1) self-identification of calibration coefficients that affects long-term performance, and 2) threshold filter to reject the broadband noise component that affects short-term performance. Experimental results with the proposed phased method applied to Epson XV3500 gyro demonstrate that it effectively yields minimal drift heading angle measurements getting over major error sources in the MEMS gyro output.
NASA Astrophysics Data System (ADS)
Griessbach, Sabine; Hoffmann, Lars; Höpfner, Michael; Riese, Martin; Spang, Reinhold
2013-09-01
The viability of a spectrally averaging model to perform radiative transfer calculations in the infrared including scattering by atmospheric particles is examined for the application of infrared limb remote sensing measurements. Here we focus on the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS) aboard the European Space Agency's Envisat. Various spectra for clear air and cloudy conditions were simulated with a spectrally averaging radiative transfer model and a line-by-line radiative transfer model for three atmospheric window regions (825-830, 946-951, 1224-1228 cm-1) and compared to each other. The results are rated in terms of the MIPAS noise equivalent spectral radiance (NESR). The clear air simulations generally agree within one NESR. The cloud simulations neglecting the scattering source term agree within two NESR. The differences between the cloud simulations including the scattering source term are generally below three and always below four NESR. We conclude that the spectrally averaging approach is well suited for fast and accurate infrared radiative transfer simulations including scattering by clouds. We found that the main source for the differences between the cloud simulations of both models is the cloud edge sampling. Furthermore we reasoned that this model comparison for clouds is also valid for atmospheric aerosol in general.
Park, Eun Sug; Hopke, Philip K; Oh, Man-Suk; Symanski, Elaine; Han, Daikwon; Spiegelman, Clifford H
2014-07-01
There has been increasing interest in assessing health effects associated with multiple air pollutants emitted by specific sources. A major difficulty with achieving this goal is that the pollution source profiles are unknown and source-specific exposures cannot be measured directly; rather, they need to be estimated by decomposing ambient measurements of multiple air pollutants. This estimation process, called multivariate receptor modeling, is challenging because of the unknown number of sources and unknown identifiability conditions (model uncertainty). The uncertainty in source-specific exposures (source contributions) as well as uncertainty in the number of major pollution sources and identifiability conditions have been largely ignored in previous studies. A multipollutant approach that can deal with model uncertainty in multivariate receptor models while simultaneously accounting for parameter uncertainty in estimated source-specific exposures in assessment of source-specific health effects is presented in this paper. The methods are applied to daily ambient air measurements of the chemical composition of fine particulate matter ([Formula: see text]), weather data, and counts of cardiovascular deaths from 1995 to 1997 for Phoenix, AZ, USA. Our approach for evaluating source-specific health effects yields not only estimates of source contributions along with their uncertainties and associated health effects estimates but also estimates of model uncertainty (posterior model probabilities) that have been ignored in previous studies. The results from our methods agreed in general with those from the previously conducted workshop/studies on the source apportionment of PM health effects in terms of number of major contributing sources, estimated source profiles, and contributions. However, some of the adverse source-specific health effects identified in the previous studies were not statistically significant in our analysis, which probably resulted because we incorporated parameter uncertainty in estimated source contributions that has been ignored in the previous studies into the estimation of health effects parameters. © The Author 2014. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
26 CFR 1.737-3 - Basis adjustments; Recovery rules.
Code of Federal Regulations, 2012 CFR
2012-04-01
... Properties A1, A2, and A3 is long-term, U.S.-source capital gain or loss. The character of gain on Property A4 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real property...-term, foreign-source capital gain ($3,000 total gain under section 737 × $2,000 net long-term, foreign...
26 CFR 1.737-3 - Basis adjustments; Recovery rules.
Code of Federal Regulations, 2013 CFR
2013-04-01
... Properties A1, A2, and A3 is long-term, U.S.-source capital gain or loss. The character of gain on Property A4 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real property...-term, foreign-source capital gain ($3,000 total gain under section 737 × $2,000 net long-term, foreign...
26 CFR 1.737-3 - Basis adjustments; Recovery rules.
Code of Federal Regulations, 2011 CFR
2011-04-01
... Properties A1, A2, and A3 is long-term, U.S.-source capital gain or loss. The character of gain on Property A4 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real property...-term, foreign-source capital gain ($3,000 total gain under section 737 × $2,000 net long-term, foreign...
26 CFR 1.737-3 - Basis adjustments; Recovery rules.
Code of Federal Regulations, 2014 CFR
2014-04-01
... Properties A1, A2, and A3 is long-term, U.S.-source capital gain or loss. The character of gain on Property A4 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real property...-term, foreign-source capital gain ($3,000 total gain under section 737 × $2,000 net long-term, foreign...
26 CFR 1.737-3 - Basis adjustments; Recovery rules.
Code of Federal Regulations, 2010 CFR
2010-04-01
... Properties A1, A2, and A3 is long-term, U.S.-source capital gain or loss. The character of gain on Property A4 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real property...-term, foreign-source capital gain ($3,000 total gain under section 737 × $2,000 net long-term, foreign...
NASA Astrophysics Data System (ADS)
Kille, N.; Chiu, R.; Frey, M.; Hase, F.; Kumar Sha, M.; Blumenstock, T.; Hannigan, J. W.; Volkamer, R. M.
2017-12-01
Methane (CH4) is a major greenhouse gas emitted from biogenic, thermogenic, and pyrogenic sources. Here we demonstrate a novel approach to separate sources of CH4 emissions based on a network of small portable sensors performing column measurements in the Northern Colorado Front Range (NCFR). In the study area CH4 is emitted from biogenic sources such as concentrated animal feeding operations (CAFOs) and natural gas production and storage. In March 2015 we deployed a network of five Fourier Transform Spectrometers (FTS) to characterize the regional scale methane dome in Colorado's Denver-Julesburg Basin based on excess vertical column measurements (the column enhancement inside the dome over background). Three EM27sun FTS measured CH4, oxygen (O2) and water vapor (H2O) columns at Eaton, CO (inside the dome) and at two boundary sites; the CU mobile SOF (Solar Occultation Flux) measured ethane (C2H6), ammonia (NH3), and H2O at Eaton, CO. The column averaged dry air mole fractions XCH4, XC2H6, and XNH3 were determined using O2 columns for air mass factor normalization, and background column was subtracted to derive excess vertical columns of DXCH4, DXC2H6, DXNH3 at Eaton, CO. Eaton is located both near CAFOs and at the northern edge of oil and natural gas production wells. Our approach for source apportioning methane employs a linear regression analysis that explains DXCH4 in terms of DXC2H6 as tracer for natural gas sources, and DXNH3 as tracer for CAFO emissions. The results of the source apportionment are compared with literature values of the NH3/CH4 and C2H6/CH4 ratio to evaluate the method of excess columns, which is independent of boundary layer height.
Nitrogen enrichment regulates calcium sources in forests
Hynicka, Justin D.; Pett-Ridge, Julie C.; Perakis, Steven
2016-01-01
Nitrogen (N) is a key nutrient that shapes cycles of other essential elements in forests, including calcium (Ca). When N availability exceeds ecosystem demands, excess N can stimulate Ca leaching and deplete Ca from soils. Over the long term, these processes may alter the proportion of available Ca that is derived from atmospheric deposition vs. bedrock weathering, which has fundamental consequences for ecosystem properties and nutrient supply. We evaluated how landscape variation in soil N, reflecting long-term legacies of biological N fixation, influenced plant and soil Ca availability and ecosystem Ca sources across 22 temperate forests in Oregon. We also examined interactions between soil N and bedrock Ca using soil N gradients on contrasting basaltic vs. sedimentary bedrock that differed 17-fold in underlying Ca content. We found that low-N forests on Ca-rich basaltic bedrock relied strongly on Ca from weathering, but that soil N enrichment depleted readily weatherable mineral Ca and shifted forest reliance toward atmospheric Ca. Forests on Ca-poor sedimentary bedrock relied more consistently on atmospheric Ca across all levels of soil N enrichment. The broad importance of atmospheric Ca was unexpected given active regional uplift and erosion that are thought to rejuvenate weathering supply of soil minerals. Despite different Ca sources to forests on basaltic vs. sedimentary bedrock, we observed consistent declines in plant and soil Ca availability with increasing N, regardless of the Ca content of underlying bedrock. Thus, traditional measures of Ca availability in foliage and soil exchangeable pools may poorly reflect long-term Ca sources that sustain soil fertility. We conclude that long-term soil N enrichment can deplete available Ca and cause forests to rely increasingly on Ca from atmospheric deposition, which may limit ecosystem Ca supply in an increasingly N-rich world.
Public Exposure from Indoor Radiofrequency Radiation in the City of Hebron, West Bank-Palestine.
Lahham, Adnan; Sharabati, Afefeh; ALMasri, Hussien
2015-08-01
This work presents the results of measured indoor exposure levels to radiofrequency (RF) radiation emitting sources in one of the major cities in the West Bank-the city of Hebron. Investigated RF emitters include FM, TV broadcasting stations, mobile telephony base stations, cordless phones [Digital Enhanced Cordless Telecommunications (DECT)], and wireless local area networks (WLAN). Measurements of power density were conducted in 343 locations representing different site categories in the city. The maximum total power density found at any location was about 2.3 × 10 W m with a corresponding exposure quotient of about 0.01. This value is well below unity, indicating compliance with the guidelines of the International Commission on Non-ionizing Radiation Protection (ICNIRP). The average total exposure from all RF sources was 0.08 × 10 W m. The relative contributions from different sources to the total exposure in terms of exposure quotient were evaluated and found to be 46% from FM radio, 26% from GSM900, 15% from DECT phones, 9% from WLAN, 3% from unknown sources, and 1% from TV broadcasting. RF sources located outdoors contribute about 73% to the population exposure indoors.
Radiometric Calibration Techniques for Signal-of-Opportunity Reflectometers
NASA Technical Reports Server (NTRS)
Piepmeier, Jeffrey R.; Shah, Rashmi; Deshpande, Manohar; Johnson, Carey
2014-01-01
Bi-static reflection measurements utilizing global navigation satellite service (GNSS) or other signals of opportunity (SoOp) can be used to sense ocean and terrestrial surface properties. End-to-end calibration of GNSS-R has been performed using well-characterized reflection surface (e.g., water), direct path antenna, and receiver gain characterization. We propose an augmented approach using on-board receiver electronics for radiometric calibration of SoOp reflectometers utilizing direct and reflected signal receiving antennas. The method calibrates receiver and correlator gains and offsets utilizing a reference switch and common noise source. On-board electronic calibration sources, such as reference switches, noise diodes and loop-back circuits, have shown great utility in stabilizing total power and correlation microwave radiometer and scatterometer receiver electronics in L-band spaceborne instruments. Application to SoOp instruments is likely to bring several benefits. For example, application to provide short and long time scale calibration stability of the direct path channel, especially in low signal-to-noise ratio configurations, is directly analogous to the microwave radiometer problem. The direct path channel is analogous to the loopback path in a scatterometer to provide a reference of the transmitted power, although the receiver is independent from the reflected path channel. Thus, a common noise source can be used to measure the gain ratio of the two paths. Using these techniques long-term (days to weeks) calibration stability of spaceborne L-band scatterometer and radiometer has been achieved better than 0.1. Similar long-term stability would likely be needed for a spaceborne reflectometer mission to measure terrestrial properties such as soil moisture.
Troyer, Jennifer L; Sause, Wendy
2013-01-01
Objective To test for an association between traditional nursing home quality measures and two sources of resident- and caregiver-derived nursing home complaints. Data Sources Nursing home complaints to the North Carolina Long-Term Care Ombudsman Program and state certification agency from October 2002 through September 2006 were matched with Online Survey Certification and Reporting data and Minimum Data Set Quality Indicators (MDS-QIs). Study Design We examine the association between the number of complaints filed against a facility and measures of inspection violations, staffing levels, and MDS-QIs. Data Extraction One observation per facility per quarter is constructed by matching quarterly data on complaints to OSCAR data from the same or most recent prior quarter and to MDS-QIs from the same quarter. One observation per inspection is obtained by matching OSCAR data to complaint totals from both the same and the immediate prior quarter. Principal Findings There is little relationship between MDS-QIs and complaints. Ombudsman complaints and inspection violations are generally unrelated, but there is a positive relationship between state certification agency complaints and inspection violations. Conclusions Ombudsman and state certification agency complaint data are resident- and caregiver-derived quality measures that are distinctive from and complement traditional quality measures. PMID:23216541
Sources and Deposition of Polycyclic Aromatic Hydrocarbons to Western U.S. National Parks
USENKO, SASCHA; MASSEY SIMONICH, STACI L.; HAGEMAN, KIMBERLY J.; SCHRLAU, JILL E.; GEISER, LINDA; CAMPBELL, DON H.; APPLEBY, PETER G.; LANDERS, DIXON H.
2010-01-01
Seasonal snowpack, lichens, and lake sediment cores were collected from fourteen lake catchments in eight western U.S. National Parks and analyzed for sixteen polycyclic aromatic hydrocarbons (PAHs) in order to determine their current and historical deposition, as well as to identify their potential sources. Seasonal snowpack was measured to determine the current wintertime atmospheric PAH deposition; lichens were measured to determine the long-term, year around deposition; and the temporal PAH deposition trends were reconstructed using lake sediment cores dated using 210Pb and 137Cs. The fourteen remote lake catchments ranged from low-latitude catchments (36.6° N) at high elevation (2900 masl) in Sequoia National Park, CA to high-latitude catchments (68.4° N) at low elevation (427 masl) in the Alaskan Arctic. Over 75% of the catchments demonstrated statistically significant temporal trends in ΣPAH sediment flux, depending on catchment proximity to source regions and topographic barriers. The ΣPAH concentrations and fluxes in seasonal snowpack, lichens, and surficial sediment were 3.6 to 60,000 times greater in the Snyder Lake catchment of Glacier National Park than the other 13 lake catchments. The PAH ratios measured in snow, lichen, and sediment were used to identify a local aluminum smelter as a major source of PAHs to the Snyder Lake catchment. These results suggest that topographic barriers influence the atmospheric transport and deposition of PAHs in high-elevation ecosystems and that PAH sources to these national park ecosystems range from local point sources to diffuse regional and global sources. PMID:20465303
NASA Astrophysics Data System (ADS)
Xiong, C. Y.; Chen, J.; Li, Q.; Liu, Y.; Gao, L.
2014-12-01
A three-wave laser polarimeter-interferometer, equipped with three independent far-infrared laser sources, has been developed on Joint-TEXT (J-TEXT) tokamak. The diagnostic system is capable of high-resolution temporal and phase measurement of the Faraday angle and line-integrated density. However, for long-term operation (>10 min), the free-running lasers can lead to large drifts of the intermediate frequencies (˜100-˜500 kHz/10 min) and decay of laser power (˜10%-˜20%/10 min), which act to degrade diagnostic performance. In addition, these effects lead to increased maintenance cost and limit measurement applicability to long pulse/steady state experiments. To solve this problem, a real-time feedback control method of the laser source is proposed. By accurately controlling the length of each laser cavity, both the intermediate frequencies and laser power can be simultaneously controlled: the intermediate frequencies are controlled according to the pre-set values, while the laser powers are maintained at an optimal level. Based on this approach, a real-time feedback control system has been developed and applied on J-TEXT polarimeter-interferometer. Long-term (theoretically no time limit) feedback of intermediate frequencies (maximum change less than ±12 kHz) and laser powers (maximum relative power change less than ±7%) has been successfully achieved.
Xiong, C Y; Chen, J; Li, Q; Liu, Y; Gao, L
2014-12-01
A three-wave laser polarimeter-interferometer, equipped with three independent far-infrared laser sources, has been developed on Joint-TEXT (J-TEXT) tokamak. The diagnostic system is capable of high-resolution temporal and phase measurement of the Faraday angle and line-integrated density. However, for long-term operation (>10 min), the free-running lasers can lead to large drifts of the intermediate frequencies (∼100-∼500 kHz/10 min) and decay of laser power (∼10%-∼20%/10 min), which act to degrade diagnostic performance. In addition, these effects lead to increased maintenance cost and limit measurement applicability to long pulse/steady state experiments. To solve this problem, a real-time feedback control method of the laser source is proposed. By accurately controlling the length of each laser cavity, both the intermediate frequencies and laser power can be simultaneously controlled: the intermediate frequencies are controlled according to the pre-set values, while the laser powers are maintained at an optimal level. Based on this approach, a real-time feedback control system has been developed and applied on J-TEXT polarimeter-interferometer. Long-term (theoretically no time limit) feedback of intermediate frequencies (maximum change less than ±12 kHz) and laser powers (maximum relative power change less than ±7%) has been successfully achieved.
Constructing a Measure of Private-pay Nursing Home Days.
Thomas, Kali S; Silver, Benjamin; Gozalo, Pedro L; Dosa, David; Grabowski, David C; Makineni, Rajesh; Mor, Vincent
2018-05-01
Nursing home (NH) care is financed through multiple sources. Although Medicaid is the predominant payer for NH care, over 20% of residents pay out-of-pocket for their care. Despite this large percentage, an accepted measure of private-pay NH occupancy has not been established and little is known about the types of facilities and the long-term care markets that cater to this population. To describe 2 novel measures of private-pay utilization in the NH setting, including the proportion of privately financed residents and resident days, and examine their construct validity. Retrospective descriptive analysis of US NHs in 2007-2009. We used Medicare claims, Medicare Enrollment records, and the Minimum Data Set to create measures of private-pay resident prevalence and proportion of privately financed NH days. We compared our estimates of private-pay utilization to payer data collected in the NH annual certification survey and evaluated the relationships of our measures with facility characteristics. Our measures of private-pay resident prevalence and private-pay days are highly correlated (r=0.83, P<0.001 and r=0.83, P<0.001, respectively) with the rate of "other payer" reported in the annual certification survey. We also observed a significantly higher proportion of private-pay residents and days in higher quality facilities. This new methodology provides estimates of private-pay resident prevalence and resident days. These measures were correlated with estimates using other data sources and validated against measures of facility quality. These data set the stage for additional work to examine questions related to NH payment, quality of care, and responses to changes in the long-term care market.
Techniques for analyses of trends in GRUAN data
NASA Astrophysics Data System (ADS)
Bodeker, G. E.; Kremser, S.
2015-04-01
The Global Climate Observing System (GCOS) Reference Upper Air Network (GRUAN) provides reference quality RS92 radiosonde measurements of temperature, pressure and humidity. A key attribute of reference quality measurements, and hence GRUAN data, is that each datum has a well characterized and traceable estimate of the measurement uncertainty. The long-term homogeneity of the measurement records, and their well characterized uncertainties, make these data suitable for reliably detecting changes in global and regional climate on decadal time scales. Considerable effort is invested in GRUAN operations to (i) describe and analyse all sources of measurement uncertainty to the extent possible, (ii) quantify and synthesize the contribution of each source of uncertainty to the total measurement uncertainty, and (iii) verify that the evaluated net uncertainty is within the required target uncertainty. However, if the climate science community is not sufficiently well informed on how to capitalize on this added value, the significant investment in estimating meaningful measurement uncertainties is largely wasted. This paper presents and discusses the techniques that will need to be employed to reliably quantify long-term trends in GRUAN data records. A pedagogical approach is taken whereby numerical recipes for key parts of the trend analysis process are explored. The paper discusses the construction of linear least squares regression models for trend analysis, boot-strapping approaches to determine uncertainties in trends, dealing with the combined effects of autocorrelation in the data and measurement uncertainties in calculating the uncertainty on trends, best practice for determining seasonality in trends, how to deal with co-linear basis functions, and interpreting derived trends. Synthetic data sets are used to demonstrate these concepts which are then applied to a first analysis of temperature trends in RS92 radiosonde upper air soundings at the GRUAN site at Lindenberg, Germany (52.21° N, 14.12° E).
Techniques for analyses of trends in GRUAN data
NASA Astrophysics Data System (ADS)
Bodeker, G. E.; Kremser, S.
2014-12-01
The Global Climate Observing System (GCOS) Reference Upper Air Network (GRUAN) provides reference quality RS92 radiosonde measurements of temperature, pressure and humidity. A key attribute of reference quality measurements, and hence GRUAN data, is that each datum has a well characterised and traceable estimate of the measurement uncertainty. The long-term homogeneity of the measurement records, and their well characterised uncertainties, make these data suitable for reliably detecting changes in global and regional climate on decadal time scales. Considerable effort is invested in GRUAN operations to (i) describe and analyse all sources of measurement uncertainty to the extent possible, (ii) quantify and synthesize the contribution of each source of uncertainty to the total measurement uncertainty, and (iii) verify that the evaluated net uncertainty is within the required target uncertainty. However, if the climate science community is not sufficiently well informed on how to capitalize on this added value, the significant investment in estimating meaningful measurement uncertainties is largely wasted. This paper presents and discusses the techniques that will need to be employed to reliably quantify long-term trends in GRUAN data records. A pedagogical approach is taken whereby numerical recipes for key parts of the trend analysis process are explored. The paper discusses the construction of linear least squares regression models for trend analysis, boot-strapping approaches to determine uncertainties in trends, dealing with the combined effects of autocorrelation in the data and measurement uncertainties in calculating the uncertainty on trends, best practice for determining seasonality in trends, how to deal with co-linear basis functions, and interpreting derived trends. Synthetic data sets are used to demonstrate these concepts which are then applied to a first analysis of temperature trends in RS92 radiosonde upper air soundings at the GRUAN site at Lindenberg, Germany (52.21° N, 14.12° E).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Blumberg, L.N.; Murphy, J.B.; Reusch, M.F.
1991-01-01
The orbit, tune, chromaticity and {beta} values for the Phase 1 XLS ring were computed by numerical integration of equations of motion using fields obtained from the coefficients of the 3-dimensional solution of Laplace's Equation evaluated by fits to magnetic measurements. The results are in good agreement with available data. The method has been extended to higher order fits of TOSCA generated fields in planes normal to the reference axis using the coil configuration proposed for the Superconducting X-Ray Lithography Source. Agreement with results from numerical integration through fields given directly by TOSCA is excellent. The formulation of the normalmore » multipole expansion presented by Brown and Servranckx has been extended to include skew multipole terms. The method appears appropriate for analysis of magnetic measurements of the SXLS. 8 refs. , 2 figs., 2 tabs.« less
Hong, Hongwei; Rahal, Mohamad; Demosthenous, Andreas; Bayford, Richard H
2009-10-01
Multi-frequency electrical impedance tomography (MF-EIT) systems require current sources that are accurate over a wide frequency range (1 MHz) and with large load impedance variations. The most commonly employed current source design in EIT systems is the modified Howland circuit (MHC). The MHC requires tight matching of resistors to achieve high output impedance and may suffer from instability over a wide frequency range in an integrated solution. In this paper, we introduce a new integrated current source design in CMOS technology and compare its performance with the MHC. The new integrated design has advantages over the MHC in terms of power consumption and area. The output current and the output impedance of both circuits were determined through simulations and measurements over the frequency range of 10 kHz to 1 MHz. For frequencies up to 1 MHz, the measured maximum variation of the output current for the integrated current source is 0.8% whereas for the MHC the corresponding value is 1.5%. Although the integrated current source has an output impedance greater than 1 MOmega up to 1 MHz in simulations, in practice, the impedance is greater than 160 kOmega up to 1 MHz due to the presence of stray capacitance.
USDA-ARS?s Scientific Manuscript database
Large uncertainties for landfill CH4 emissions due to spatial and temporal variabilities remain unresolved by short-term field campaigns and historic GHG inventory models. Using four field methods (aircraft-based mass balance, tracer correlation, vertical radial plume mapping, and static chambers) ...
Nonlinear forecasting as a way of distinguishing chaos from measurement error in time series
NASA Astrophysics Data System (ADS)
Sugihara, George; May, Robert M.
1990-04-01
An approach is presented for making short-term predictions about the trajectories of chaotic dynamical systems. The method is applied to data on measles, chickenpox, and marine phytoplankton populations, to show how apparent noise associated with deterministic chaos can be distinguished from sampling error and other sources of externally induced environmental noise.
Kelly Elder; Don Cline; Angus Goodbody; Paul Houser; Glen E. Liston; Larry Mahrt; Nick Rutter
2009-01-01
A short-term meteorological database has been developed for the Cold Land Processes Experiment (CLPX). This database includes meteorological observations from stations designed and deployed exclusively for CLPXas well as observations available from other sources located in the small regional study area (SRSA) in north-central Colorado. The measured weather parameters...
George E. Myers
1983-01-01
A number of commercial panel products, primarily particleboard and hardwood plywood, were tested for their formaldehyde emission behavior using desiccator, perforator, and dynamic chamber methods. The results were analyzed in terms of the source of formaldehyde observed in the tests (free vs. hydrolytically produced) and the potential utility of the testa as product...
Brauer, M; Hirtle, R; Lang, B; Ott, W
2000-01-01
Personal monitoring studies have indicated that environmental tobacco smoke (ETS) and cooking are major indoor particulate sources in residential and nonindustrial environments. Continuous monitoring of fine particles improves exposure assessment by characterizing the effect of time-varying indoor sources. We evaluated a portable nephelometer as a continuous monitor of indoor particulate levels. Simultaneous sampling with the nephelometer and PM2.5 impactors was undertaken to determine the relationship between particle light scattering extinction coefficient (sigma(sp)) and particle mass concentration in field and environmental chamber settings. Chamber studies evaluated nephelometer measurements of ETS and particles produced from toasting bread and frying foods. Field measurements were conducted in 20 restaurants and bars with different smoking restrictions, and in five residential kitchens. Additional measurements compared the nephelometer to a different mass measurement method, a piezobalance, in a well-characterized residence where various foods were cooked and ETS was produced. Since the piezobalance provides 2-min average mass concentration measurements, these comparisons tested the ability of the nephelometer to measure transient particle concentration peaks and decay rate curves. We found that sigma(sp) and particle mass were highly correlated (R2 values of 0.63-0.98) over a large concentration range (5-1600 microg/m3) and for different particle sources. Piezobalance and gravimetric comparisons with the nephelometer indicated similar sigma(sp) vs. mass slopes (5.6 and 4.7 m2/g for piezobalance and gravimetric comparisons of ETS, respectively). Somewhat different sigma(sp) vs. particle mass slopes (1.9-5.6 m2/g) were observed for the different particle sources, reflecting the influence of particle composition on light scattering. However, in similar indoor environments, the relationship between particle light scattering and mass concentration was consistent enough to use independent nephelometer measurements as estimates of short-term mass concentrations. A method to use nephelometer measurements to determine particulate source strengths is derived and an example application is described.
Kim, Sun Kyu; Burris, David R; Bryant-Genevier, Jonathan; Gorder, Kyle A; Dettenmaier, Erik M; Zellers, Edward T
2012-06-05
We demonstrate the use of two prototype Si-microfabricated gas chromatographs (μGC) for continuous, short-term measurements of indoor trichloroethylene (TCE) vapor concentrations related to the investigation of TCE vapor intrusion (VI) in two houses. In the first house, with documented TCE VI, temporal variations in TCE air concentrations were monitored continuously for up to 48 h near the primary VI entry location under different levels of induced differential pressure (relative to the subslab). Concentrations ranged from 0.23 to 27 ppb by volume (1.2-150 μg/m(3)), and concentration trends agreed closely with those determined from concurrent reference samples. The sensitivity and temporal resolution of the measurements were sufficiently high to detect transient fluctuations in concentration resulting from short-term changes in variables affecting the extent of VI. Spatial monitoring showed a decreasing TCE concentration gradient with increasing distance from the primary VI entry location. In the second house, with no TCE VI, spatial profiles derived from the μGC prototype data revealed an intentionally hidden source of TCE within a closet, demonstrating the capability for locating non-VI sources. Concentrations measured in this house ranged from 0.51 to 56 ppb (2.7-300 μg/m(3)), in good agreement with reference method values. This first field demonstration of μGC technology for automated, near-real-time, selective VOC monitoring at low- or subppb levels augurs well for its use in short- and long-term on-site analysis of indoor air in support of VI assessments.
Schiller, Q.; Tu, W.; Ali, A. F.; ...
2017-03-11
The most significant unknown regarding relativistic electrons in Earth’s outer Van Allen radiation belt is the relative contribution of loss, transport, and acceleration processes within the inner magnetosphere. Detangling each individual process is critical to improve the understanding of radiation belt dynamics, but determining a single component is challenging due to sparse measurements in diverse spatial and temporal regimes. However, there are currently an unprecedented number of spacecraft taking measurements that sample different regions of the inner magnetosphere. With the increasing number of varied observational platforms, system dynamics can begin to be unraveled. In this work, we employ in-situ measurementsmore » during the 13-14 January 2013 enhancement event to isolate transport, loss, and source dynamics in a one dimensional radial diffusion model. We then validate the results by comparing them to Van Allen Probes and THEMIS observations, indicating that the three terms have been accurately and individually quantified for the event. Finally, a direct comparison is performed between the model containing event-specific terms and various models containing terms parameterized by geomagnetic index. Models using a simple 3/Kp loss timescale show deviation from the event specific model of nearly two orders of magnitude within 72 hours of the enhancement event. However, models using alternative loss timescales closely resemble the event specific model.« less
Stochastic Short-term High-resolution Prediction of Solar Irradiance and Photovoltaic Power Output
DOE Office of Scientific and Technical Information (OSTI.GOV)
Melin, Alexander M.; Olama, Mohammed M.; Dong, Jin
The increased penetration of solar photovoltaic (PV) energy sources into electric grids has increased the need for accurate modeling and prediction of solar irradiance and power production. Existing modeling and prediction techniques focus on long-term low-resolution prediction over minutes to years. This paper examines the stochastic modeling and short-term high-resolution prediction of solar irradiance and PV power output. We propose a stochastic state-space model to characterize the behaviors of solar irradiance and PV power output. This prediction model is suitable for the development of optimal power controllers for PV sources. A filter-based expectation-maximization and Kalman filtering mechanism is employed tomore » estimate the parameters and states in the state-space model. The mechanism results in a finite dimensional filter which only uses the first and second order statistics. The structure of the scheme contributes to a direct prediction of the solar irradiance and PV power output without any linearization process or simplifying assumptions of the signal’s model. This enables the system to accurately predict small as well as large fluctuations of the solar signals. The mechanism is recursive allowing the solar irradiance and PV power to be predicted online from measurements. The mechanism is tested using solar irradiance and PV power measurement data collected locally in our lab.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schiller, Q.; Tu, W.; Ali, A. F.
The most significant unknown regarding relativistic electrons in Earth’s outer Van Allen radiation belt is the relative contribution of loss, transport, and acceleration processes within the inner magnetosphere. Detangling each individual process is critical to improve the understanding of radiation belt dynamics, but determining a single component is challenging due to sparse measurements in diverse spatial and temporal regimes. However, there are currently an unprecedented number of spacecraft taking measurements that sample different regions of the inner magnetosphere. With the increasing number of varied observational platforms, system dynamics can begin to be unraveled. In this work, we employ in-situ measurementsmore » during the 13-14 January 2013 enhancement event to isolate transport, loss, and source dynamics in a one dimensional radial diffusion model. We then validate the results by comparing them to Van Allen Probes and THEMIS observations, indicating that the three terms have been accurately and individually quantified for the event. Finally, a direct comparison is performed between the model containing event-specific terms and various models containing terms parameterized by geomagnetic index. Models using a simple 3/Kp loss timescale show deviation from the event specific model of nearly two orders of magnitude within 72 hours of the enhancement event. However, models using alternative loss timescales closely resemble the event specific model.« less
10 CFR 50.67 - Accident source term.
Code of Federal Regulations, 2014 CFR
2014-01-01
... occupancy of the control room under accident conditions without personnel receiving radiation exposures in... 10 Energy 1 2014-01-01 2014-01-01 false Accident source term. 50.67 Section 50.67 Energy NUCLEAR... Conditions of Licenses and Construction Permits § 50.67 Accident source term. (a) Applicability. The...
10 CFR 50.67 - Accident source term.
Code of Federal Regulations, 2012 CFR
2012-01-01
... occupancy of the control room under accident conditions without personnel receiving radiation exposures in... 10 Energy 1 2012-01-01 2012-01-01 false Accident source term. 50.67 Section 50.67 Energy NUCLEAR... Conditions of Licenses and Construction Permits § 50.67 Accident source term. (a) Applicability. The...
10 CFR 50.67 - Accident source term.
Code of Federal Regulations, 2010 CFR
2010-01-01
... occupancy of the control room under accident conditions without personnel receiving radiation exposures in... 10 Energy 1 2010-01-01 2010-01-01 false Accident source term. 50.67 Section 50.67 Energy NUCLEAR... Conditions of Licenses and Construction Permits § 50.67 Accident source term. (a) Applicability. The...
10 CFR 50.67 - Accident source term.
Code of Federal Regulations, 2013 CFR
2013-01-01
... occupancy of the control room under accident conditions without personnel receiving radiation exposures in... 10 Energy 1 2013-01-01 2013-01-01 false Accident source term. 50.67 Section 50.67 Energy NUCLEAR... Conditions of Licenses and Construction Permits § 50.67 Accident source term. (a) Applicability. The...
10 CFR 50.67 - Accident source term.
Code of Federal Regulations, 2011 CFR
2011-01-01
... occupancy of the control room under accident conditions without personnel receiving radiation exposures in... 10 Energy 1 2011-01-01 2011-01-01 false Accident source term. 50.67 Section 50.67 Energy NUCLEAR... Conditions of Licenses and Construction Permits § 50.67 Accident source term. (a) Applicability. The...
Double frequency of difference frequency signals for optical Doppler effect measuring velocity
NASA Astrophysics Data System (ADS)
Yang, Xiufang; Zhou, Renkui; Wei, W. L.; Wang, Xiaoming
2005-12-01
The mathematical model for measuring moving objects (including fluid body, rolled steel materials in the steel works, turbulent flow, vibration body, etc.) velocity or speed by non-contact method is established using light-wave Doppler effect in this paper. In terms of concrete conditions of different optical circuits, and with the correlated conditions substituted, it is easy to obtain the measurement velocity formulas related to optical circuits. An optical circuit layout of difference Doppler effect measuring velocity is suggested in this paper. The fine beam of light emitted by laser is divided into parallel two beam by spectroscope and mirror They are focused on the object point p by a condenser lens respectively. The object point p become a diffuse source. It scatter rays to every aspect. Some rays scattered by the diffuse source p are collected by a lens. Photoelectric detecter receive the lights collected by the lens. This optical circuit layout can realize the double frequency of difference frequency signals in a novel way.
Measurement and analysis of electromagnetic fields from trams, trains and hybrid cars.
Halgamuge, Malka N; Abeyrathne, Chathurika D; Mendis, Priyan
2010-10-01
Electricity is used substantially and sources of electric and magnetic fields are, unavoidably, everywhere. The transportation system is a source of these fields, to which a large proportion of the population is exposed. Hence, investigation of the effects of long-term exposure of the general public to low-frequency electromagnetic fields caused by the transportation system is critically important. In this study, measurements of electric and magnetic fields emitted from Australian trams, trains and hybrid cars were investigated. These measurements were carried out under different conditions, locations, and are summarised in this article. A few of the measured electric and magnetic field strengths were significantly lower than those found in prior studies. These results seem to be compatible with the evidence of the laboratory studies on the biological effects that are found in the literature, although they are far lower than international levels, such as those set up in the International Commission on Non-Ionising Radiation Protection guidelines.
ICP-MS measurement of iodine diffusion in IG-110 graphite for HTGR/VHTR
NASA Astrophysics Data System (ADS)
Carter, L. M.; Brockman, J. D.; Robertson, J. D.; Loyalka, S. K.
2016-05-01
Graphite functions as a structural material and as a barrier to fission product release in HTGR/VHTR designs, and elucidation of transport parameters for fission products in reactor-grade graphite is thus required for reactor source terms calculations. We measured iodine diffusion in spheres of IG-110 graphite using a release method based on Fickain diffusion kinetics. Two sources of iodine were loaded into the graphite spheres; molecular iodine (I2) and cesium iodide (CsI). Measurements of the diffusion coefficient were made over a temperature range of 873-1293 K. We have obtained the following Arrhenius expressions for iodine diffusion:DI , CsI infused =(6 ×10-12 2/s) exp(30,000 J/mol RT) And,DI , I2 infused =(4 ×10-10 m2/s) exp(-11,000 J/mol RT ) The results indicate that iodine diffusion in IG-110 graphite is not well-described by Fickan diffusion kinetics. To our knowledge, these are the first measurements of iodine diffusion in IG-110 graphite.
NASA Astrophysics Data System (ADS)
Gillaspy, J. D.; Chantler, C. T.; Paterson, D.; Hudson, L. T.; Serpa, F. G.; Takács, E.
2010-04-01
The first measurement of hydrogen-like vanadium x-ray Lyman alpha transitions has been made. The measurement was made on an absolute scale, fully independent of atomic structure calculations. Sufficient signal was obtained to reduce the statistical uncertainty to a small fraction of the total uncertainty budget. Potential sources of systematic error due to Doppler shifts were eliminated by performing the measurement on trapped ions. The energies for Ly α1 (1s-2p3/2) and Ly α2 (1s-2p1/2) are found to be 5443.95(25) eV and 5431.10(25) eV, respectively. These results are within approximately 1.5 σ (experimental) of the theoretical values 5443.63 eV and 5430.70 eV. The results are discussed in terms of their relation to the Lamb shift and the development of an x-ray wavelength standard based on a compact source of trapped highly charged ions.
NASA Astrophysics Data System (ADS)
Blanchard, J. W.; Sjolander, T. F.; King, J. P.; Ledbetter, M. P.; Levine, E. H.; Bajaj, V. S.; Budker, D.; Pines, A.
2015-12-01
Zero- to ultralow-field nuclear magnetic resonance (ZULF NMR) provides a new regime for the measurement of nuclear spin-spin interactions free from the effects of large magnetic fields, such as truncation of terms that do not commute with the Zeeman Hamiltonian. One such interaction, the magnetic dipole-dipole coupling, is a valuable source of spatial information in NMR, though many terms are unobservable in high-field NMR, and the coupling averages to zero under isotropic molecular tumbling. Under partial alignment, this information is retained in the form of so-called residual dipolar couplings. We report zero- to ultralow-field NMR measurements of residual dipolar couplings in acetonitrile-2-13C aligned in stretched polyvinyl acetate gels. This permits the investigation of dipolar couplings as a perturbation on the indirect spin-spin J coupling in the absence of an applied magnetic field. As a consequence of working at zero magnetic field, we observe terms of the dipole-dipole coupling Hamiltonian that are invisible in conventional high-field NMR. This technique expands the capabilities of zero- to ultralow-field NMR and has potential applications in precision measurement of subtle physical interactions, chemical analysis, and characterization of local mesoscale structure in materials.
Analysis of neutron and gamma-ray streaming along the maze of NRCAM thallium production target room.
Raisali, G; Hajiloo, N; Hamidi, S; Aslani, G
2006-08-01
Study of the shield performance of a thallium-203 production target room has been investigated in this work. Neutron and gamma-ray equivalent dose rates at various points of the maze are calculated by simulating the transport of streaming neutrons, and photons using Monte Carlo method. For determination of neutron and gamma-ray source intensities and their energy spectrum, we have applied SRIM 2003 and ALICE91 computer codes to Tl target and its Cu substrate for a 145 microA of 28.5 MeV protons beam. The MCNP/4C code has been applied with neutron source term in mode n p to consider both prompt neutrons and secondary gamma-rays. Then the code is applied for the prompt gamma-rays as the source term. The neutron-flux energy spectrum and equivalent dose rates for neutron and gamma-rays in various positions in the maze have been calculated. It has been found that the deviation between calculated and measured dose values along the maze is less than 20%.
PCB remediation in schools: a review.
Brown, Kathleen W; Minegishi, Taeko; Cummiskey, Cynthia Campisano; Fragala, Matt A; Hartman, Ross; MacIntosh, David L
2016-02-01
Growing awareness of polychlorinated biphenyls (PCBs) in legacy caulk and other construction materials of schools has created a need for information on best practices to control human exposures and comply with applicable regulations. A concise review of approaches and techniques for management of building-related PCBs is the focus of this paper. Engineering and administrative controls that block pathways of PCB transport, dilute concentrations of PCBs in indoor air or other exposure media, or establish uses of building space that mitigate exposure can be effective initial responses to identification of PCBs in a building. Mitigation measures also provide time for school officials to plan a longer-term remediation strategy and to secure the necessary resources. These longer-term strategies typically involve removal of caulk or other primary sources of PCBs as well as nearby masonry or other materials contaminated with PCBs by the primary sources. The costs of managing PCB-containing building materials from assessment through ultimate disposal can be substantial. Optimizing the efficacy and cost-effectiveness of remediation programs requires aligning a thorough understanding of sources and exposure pathways with the most appropriate mitigation and abatement methods.
Assessing and measuring wetland hydrology
Rosenberry, Donald O.; Hayashi, Masaki; Anderson, James T.; Davis, Craig A.
2013-01-01
Virtually all ecological processes that occur in wetlands are influenced by the water that flows to, from, and within these wetlands. This chapter provides the “how-to” information for quantifying the various source and loss terms associated with wetland hydrology. The chapter is organized from a water-budget perspective, with sections associated with each of the water-budget components that are common in most wetland settings. Methods for quantifying the water contained within the wetland are presented first, followed by discussion of each separate component. Measurement accuracy and sources of error are discussed for each of the methods presented, and a separate section discusses the cumulative error associated with determining a water budget for a wetland. Exercises and field activities will provide hands-on experience that will facilitate greater understanding of these processes.
NASA Astrophysics Data System (ADS)
Burman, Erik; Hansbo, Peter; Larson, Mats G.
2018-03-01
Tikhonov regularization is one of the most commonly used methods for the regularization of ill-posed problems. In the setting of finite element solutions of elliptic partial differential control problems, Tikhonov regularization amounts to adding suitably weighted least squares terms of the control variable, or derivatives thereof, to the Lagrangian determining the optimality system. In this note we show that the stabilization methods for discretely ill-posed problems developed in the setting of convection-dominated convection-diffusion problems, can be highly suitable for stabilizing optimal control problems, and that Tikhonov regularization will lead to less accurate discrete solutions. We consider some inverse problems for Poisson’s equation as an illustration and derive new error estimates both for the reconstruction of the solution from the measured data and reconstruction of the source term from the measured data. These estimates include both the effect of the discretization error and error in the measurements.
NASA Astrophysics Data System (ADS)
Winiarek, Victor; Bocquet, Marc; Saunier, Olivier; Mathieu, Anne
2012-03-01
A major difficulty when inverting the source term of an atmospheric tracer dispersion problem is the estimation of the prior errors: those of the atmospheric transport model, those ascribed to the representativity of the measurements, those that are instrumental, and those attached to the prior knowledge on the variables one seeks to retrieve. In the case of an accidental release of pollutant, the reconstructed source is sensitive to these assumptions. This sensitivity makes the quality of the retrieval dependent on the methods used to model and estimate the prior errors of the inverse modeling scheme. We propose to use an estimation method for the errors' amplitude based on the maximum likelihood principle. Under semi-Gaussian assumptions, it takes into account, without approximation, the positivity assumption on the source. We apply the method to the estimation of the Fukushima Daiichi source term using activity concentrations in the air. The results are compared to an L-curve estimation technique and to Desroziers's scheme. The total reconstructed activities significantly depend on the chosen method. Because of the poor observability of the Fukushima Daiichi emissions, these methods provide lower bounds for cesium-137 and iodine-131 reconstructed activities. These lower bound estimates, 1.2 × 1016 Bq for cesium-137, with an estimated standard deviation range of 15%-20%, and 1.9 - 3.8 × 1017 Bq for iodine-131, with an estimated standard deviation range of 5%-10%, are of the same order of magnitude as those provided by the Japanese Nuclear and Industrial Safety Agency and about 5 to 10 times less than the Chernobyl atmospheric releases.
CO2 fluxes from a tropical neighborhood: sources and sinks
NASA Astrophysics Data System (ADS)
Velasco, E.; Roth, M.; Tan, S.; Quak, M.; Britter, R.; Norford, L.
2011-12-01
Cities are the main contributors to the CO2 rise in the atmosphere. The CO2 released from the various emission sources is typically quantified by a bottom-up aggregation process that accounts for emission factors and fossil fuel consumption data. This approach does not consider the heterogeneity and variability of the urban emission sources, and error propagation can result in large uncertainties. In this context, direct measurements of CO2 fluxes that include all major and minor anthropogenic and natural sources and sinks from a specific district can be used to evaluate emission inventories. This study reports and compares CO2 fluxes measured directly using the eddy covariance method with emissions estimated by emissions factors and activity data for a residential neighborhood of Singapore, a highly populated and urbanized tropical city. The flux measurements were conducted during one year. No seasonal variability was found as a consequence of the constant climate conditions of tropical places; but a clear diurnal pattern with morning and late afternoon peaks in phase with the rush-hour traffic was observed. The magnitude of the fluxes throughout daylight hours is modulated by the urban vegetation, which is abundant in terms of biomass but not of land-cover (15%). Even though the carbon uptake by vegetation is significant, it does not exceed the anthropogenic emissions and the monitored district is a net CO2 source of 20.3 ton km-2 day-1 on average. The carbon uptake by vegetation is investigated as the difference between the estimated emissions and the measured fluxes during daytime.
Long term spectral irradiance measurements of a 1000-watt xenon arc lamp
NASA Technical Reports Server (NTRS)
Schneider, W. E.
1974-01-01
Spectral irradiance measurements over the range of 200 to 1060 nm were made on a 1000-watt xenon arc lamp over a period of 1500 hours. Four sets of measurements were made after periods of 70, 525, 1000, and 1500 hours of operation. The lamp (Hanovia Compact Xenon Arc Lamp) was mounted in the NASA Solar Irradiation System. When used in the System, the lamp is used as the radiating source for six test stations. Measurements were made of both the longterm stability (or variation of spectral irradiance as a function of time) and the actual spectral irradiance incident on the test specimen.
Supersonic jet noise - Its generation, prediction and effects on people and structures
NASA Technical Reports Server (NTRS)
Preisser, J. S.; Golub, R. A.; Seiner, J. M.; Powell, C. A.
1990-01-01
This paper presents the results of a study aimed at quantifying the effects of jet source noise reduction, increases in aircraft lift, and reduced aircraft thrust on the take-off noise associated with supersonic civil transports. Supersonic jet noise sources are first described, and their frequency and directivity dependence are defined. The study utilizes NASA's Aircraft Noise Prediction Program in a parametric study to weigh the relative benefits of several approaches to low noise. The baseline aircraft concept used in these predictions is the AST-205-1 powered by GE21/J11-B14A scaled engines. Noise assessment is presented in terms of effective perceived noise levels at the FAA's centerline and sideline measuring locations for current subsonic aircraft, and in terms of audiologically perceived sound of people and other indirect effects. The results show that significant noise benefit can be achieved through proper understanding and utilization of all available approaches.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kayser, Y., E-mail: yves.kayser@psi.ch; Paul Scherrer Institut, 5232 Villigen-PSI; Błachucki, W.
2014-04-15
The high-resolution von Hamos bent crystal spectrometer of the University of Fribourg was upgraded with a focused X-ray beam source with the aim of performing micro-sized X-ray fluorescence (XRF) measurements in the laboratory. The focused X-ray beam source integrates a collimating optics mounted on a low-power micro-spot X-ray tube and a focusing polycapillary half-lens placed in front of the sample. The performances of the setup were probed in terms of spatial and energy resolution. In particular, the fluorescence intensity and energy resolution of the von Hamos spectrometer equipped with the novel micro-focused X-ray source and a standard high-power water-cooled X-raymore » tube were compared. The XRF analysis capability of the new setup was assessed by measuring the dopant distribution within the core of Er-doped SiO{sub 2} optical fibers.« less
Measurement of cesium diffusion coefficients in graphite IG-110
NASA Astrophysics Data System (ADS)
Carter, L. M.; Brockman, J. D.; Loyalka, S. K.; Robertson, J. D.
2015-05-01
An understanding of the transport of fission products in High Temperature Gas-Cooled Reactors (HTGRs) is needed for operational safety as well as source term estimations. We have measured diffusion coefficients of Cs in IG-110 by using the release method, wherein we infused small graphite spheres with Cs and measured the release rates using ICP-MS. Diffusion behavior was investigated in the temperature range of 1100-1300 K. We have obtained: DCs = (1.0 ×10-7m2 /s) exp(-1.1/×105J /mol RT) and, compared our results with those available in the literature.
LCLS in—photon out: fluorescence measurement of neon using soft x-rays
Obaid, Razib; Buth, Christian; Dakovski, Georgi L.; ...
2018-01-09
Here, we measured the fluorescence photon yield of neon upon soft x-ray ionization (~1200 eV) from the x-ray free-electron laser at Linac Coherent Light Source, and demonstrated the usage of a grazing incidence spectrometer with a variable line spacing grating to perform x-ray fluorescence spectroscopy on a gas phase system. Our measurements also allowed us to estimate the focal size of the beam from the theoretical description developed, in terms of the rate equation approximation accounting for photoionization shake off of neutral neon and double auger decay of single core holes.
[Features of control of electromagnetic radiation emitted by personal computers].
Pal'tsev, Iu P; Buzov, A L; Kol'chugin, Iu I
1996-01-01
Measurements of PC electromagnetic irradiation show that the main sources are PC blocks emitting the waves of certain frequencies. Use of wide-range detectors measuring field intensity in assessment of PC electromagnetic irradiation gives unreliable results. More precise measurements by selective devices are required. Thus, it is expedient to introduce a term "spectral density of field intensity" and its maximal allowable level. In this case a frequency spectrum of PC electromagnetic irradiation is divided into 4 ranges, one of which is subjected to calculation of field intensity for each harmonic frequency, and others undergo assessment of spectral density of field intensity.
LCLS in—photon out: fluorescence measurement of neon using soft x-rays
NASA Astrophysics Data System (ADS)
Obaid, Razib; Buth, Christian; Dakovski, Georgi L.; Beerwerth, Randolf; Holmes, Michael; Aldrich, Jeff; Lin, Ming-Fu; Minitti, Michael; Osipov, Timur; Schlotter, William; Cederbaum, Lorenz S.; Fritzsche, Stephan; Berrah, Nora
2018-02-01
We measured the fluorescence photon yield of neon upon soft x-ray ionization (∼1200 eV) from the x-ray free-electron laser at Linac Coherent Light Source, and demonstrated the usage of a grazing incidence spectrometer with a variable line spacing grating to perform x-ray fluorescence spectroscopy on a gas phase system. Our measurements also allowed us to estimate the focal size of the beam from the theoretical description developed, in terms of the rate equation approximation accounting for photoionization shake off of neutral neon and double auger decay of single core holes.
Measurement of Fukushima Aerosol Debris in Sequim and Richland, WA and Ketchikan, AK
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miley, Harry S.; Bowyer, Ted W.; Engelmann, Mark D.
2013-05-01
Aerosol collections were initiated at several locations by PNNL shortly after the Great East Japan Earthquake of May 2011. Aerosol samples were transferred to laboratory high-resolution gamma spectrometers for analysis. Similar to treaty monitoring stations operating across the Northern hemisphere, iodine and other isotopes which could be volatilized at high temperature were detected. Though these locations are not far apart, they have significant variations with respect to water, mountain-range placement, and local topography. Variation in computed source terms will be shown to bound the variability of this approach to source estimation.
NASA Astrophysics Data System (ADS)
Duncan, J. M.; Band, L. E.; Groffman, P.
2017-12-01
Discharge, land use, and watershed management practices (stream restoration and stormwater control measures) have been found to be important determinants of nitrogen (N) export to receiving waters. We used long-term water quality stations from the Baltimore Ecosystem Study Long-Term Ecological Research (BES LTER) Site to quantify nitrogen export across streamflow conditions at the small watershed scale. We calculated nitrate and total nitrogen fluxes using methodology that allows for changes over time; weighted regressions on time, discharge, and seasonality. Here we tested the hypotheses that a) while the largest N stream fluxes occur during storm events, there is not a clear relationship between N flux and discharge and b) N export patterns are aseasonal in developed watersheds where sources are larger and retention capacity is lower. The goal is to scale understanding from small watersheds to larger ones. Developing a better understanding of hydrologic controls on nitrogen export is essential for successful adaptive watershed management at societally meaningful spatial scales.
Piecewise synonyms for enhanced UMLS source terminology integration.
Huang, Kuo-Chuan; Geller, James; Halper, Michael; Cimino, James J
2007-10-11
The UMLS contains more than 100 source vocabularies and is growing via the integration of others. When integrating a new source, the source terms already in the UMLS must first be found. The easiest approach to this is simple string matching. However, string matching usually does not find all concepts that should be found. A new methodology, based on the notion of piecewise synonyms, for enhancing the process of concept discovery in the UMLS is presented. This methodology is supported by first creating a general synonym dictionary based on the UMLS. Each multi-word source term is decomposed into its component words, allowing for the generation of separate synonyms for each word from the general synonym dictionary. The recombination of these synonyms into new terms creates an expanded pool of matching candidates for terms from the source. The methodology is demonstrated with respect to an existing UMLS source. It shows a 34% improvement over simple string matching.
Madjidi, Faramarz; Behroozy, Ali
2014-01-01
Exposure to visible light and near infrared (NIR) radiation in the wavelength region of 380 to 1400 nm may cause thermal retinal injury. In this analysis, the effective spectral radiance of a hot source is replaced by its temperature in the exposure limit values in the region of 380-1400 nm. This article describes the development and implementation of a computer code to predict those temperatures, corresponding to the exposure limits proposed by the American Conference of Governmental Industrial Hygienists (ACGIH). Viewing duration and apparent diameter of the source were inputs for the computer code. At the first stage, an infinite series was created for calculation of spectral radiance by integration with Planck's law. At the second stage for calculation of effective spectral radiance, the initial terms of this infinite series were selected and integration was performed by multiplying these terms by a weighting factor R(λ) in the wavelength region 380-1400 nm. At the third stage, using a computer code, the source temperature that can emit the same effective spectral radiance was found. As a result, based only on measuring the source temperature and accounting for the exposure time and the apparent diameter of the source, it is possible to decide whether the exposure to visible and NIR in any 8-hr workday is permissible. The substitution of source temperature for effective spectral radiance provides a convenient way to evaluate exposure to visible light and NIR.
ON THE CONNECTION OF THE APPARENT PROPER MOTION AND THE VLBI STRUCTURE OF COMPACT RADIO SOURCES
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moor, A.; Frey, S.; Lambert, S. B.
2011-06-15
Many of the compact extragalactic radio sources that are used as fiducial points to define the celestial reference frame are known to have proper motions detectable with long-term geodetic/astrometric very long baseline interferometry (VLBI) measurements. These changes can be as high as several hundred microarcseconds per year for certain objects. When imaged with VLBI at milliarcsecond (mas) angular resolution, these sources (radio-loud active galactic nuclei) typically show structures dominated by a compact, often unresolved 'core' and a one-sided 'jet'. The positional instability of compact radio sources is believed to be connected with changes in their brightness distribution structure. For themore » first time, we test this assumption in a statistical sense on a large sample rather than on only individual objects. We investigate a sample of 62 radio sources for which reliable long-term time series of astrometric positions as well as detailed 8 GHz VLBI brightness distribution models are available. We compare the characteristic direction of their extended jet structure and the direction of their apparent proper motion. We present our data and analysis method, and conclude that there is indeed a correlation between the two characteristic directions. However, there are cases where the {approx}1-10 mas scale VLBI jet directions are significantly misaligned with respect to the apparent proper motion direction.« less
Revising the Lubben Social Network Scale for use in residential long-term care settings.
Munn, Jean; Radey, Melissa; Brown, Kristin; Kim, Hyejin
2018-04-19
We revised the Lubben Social Network Scale (LSNS) to develop a measure of social support specific to residential long-term care (LTC) settings, the LSNS-LTC with five domains (i.e., family, friends, residents, volunteers, and staff). The authors modified the LSNS-18 to capture sources of social support specific to LTC, specifically relationships with residents, volunteers, and staff. We piloted the resultant 28-item measure with 64 LTC residents. Fifty-four respondents provided adequate information for analyses that included descriptive statistics and reliability coefficients. Twenty of the items performed well (had correlations >0.3, overall α = 0.85) and were retained. Three items required modification. The five items related to volunteers were eliminated due to extensive (>15%) missing data resulting in a proposed 23-item measure. We identified, and to some degree quantified, supportive relationships within the LTC environment, while developing a self-report tool to measure social support in these settings.
Ancient Glass: A Literature Search and its Role in Waste Management
DOE Office of Scientific and Technical Information (OSTI.GOV)
Strachan, Denis M.; Pierce, Eric M.
2010-07-01
When developing a performance assessment model for the long-term disposal of immobilized low-activity waste (ILAW) glass, it is desirable to determine the durability of glass forms over very long periods of time. However, testing is limited to short time spans, so experiments are performed under conditions that accelerate the key geochemical processes that control weathering. Verification that models currently being used can reliably calculate the long term behavior ILAW glass is a key component of the overall PA strategy. Therefore, Pacific Northwest National Laboratory was contracted by Washington River Protection Solutions, LLC to evaluate alternative strategies that can be usedmore » for PA source term model validation. One viable alternative strategy is the use of independent experimental data from archaeological studies of ancient or natural glass contained in the literature. These results represent a potential independent experiment that date back to approximately 3600 years ago or 1600 before the current era (bce) in the case of ancient glass and 106 years or older in the case of natural glass. The results of this literature review suggest that additional experimental data may be needed before the result from archaeological studies can be used as a tool for model validation of glass weathering and more specifically disposal facility performance. This is largely because none of the existing data set contains all of the information required to conduct PA source term calculations. For example, in many cases the sediments surrounding the glass was not collected and analyzed; therefore having the data required to compare computer simulations of concentration flux is not possible. This type of information is important to understanding the element release profile from the glass to the surrounding environment and provides a metric that can be used to calibrate source term models. Although useful, the available literature sources do not contain the required information needed to simulate the long-term performance of nuclear waste glasses in a near-surface or deep geologic repositories. The information that will be required include 1) experimental measurements to quantify the model parameters, 2) detailed analyses of altered glass samples, and 3) detailed analyses of the sediment surrounding the ancient glass samples.« less
Data-optimized source modeling with the Backwards Liouville Test–Kinetic method
Woodroffe, J. R.; Brito, T. V.; Jordanova, V. K.; ...
2017-09-14
In the standard practice of neutron multiplicity counting , the first three sampled factorial moments of the event triggered neutron count distribution were used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α,n) production and the induced fission source responsible for multiplication. Our study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra,more » Italy, sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.« less
Casas, Francisco J; Ortiz, David; Villa, Enrique; Cano, Juan L; Cagigas, Jaime; Pérez, Ana R; Aja, Beatriz; Terán, J Vicente; de la Fuente, Luisa; Artal, Eduardo; Hoyland, Roger; Génova-Santos, Ricardo
2015-08-05
This paper presents preliminary polarization measurements and systematic-error characterization of the Thirty Gigahertz Instrument receiver developed for the QUIJOTE experiment. The instrument has been designed to measure the polarization of Cosmic Microwave Background radiation from the sky, obtaining the Q, U, and I Stokes parameters of the incoming signal simultaneously. Two kinds of linearly polarized input signals have been used as excitations in the polarimeter measurement tests in the laboratory; these show consistent results in terms of the Stokes parameters obtained. A measurement-based systematic-error characterization technique has been used in order to determine the possible sources of instrumental errors and to assist in the polarimeter calibration process.
Schweren, Lizanne; Hoekstra, Pieter; van Lieshout, Marloes; Oosterlaan, Jaap; Lambregts-Rommelse, Nanda; Buitelaar, Jan; Franke, Barbara; Hartman, Catharina
2018-03-13
Methodological and ethical constraints have hampered studies into long-term lasting outcomes of stimulant treatment in individuals with attention-deficit/hyperactivity disorder (ADHD). Lasting effects may be beneficial (i.e. improved functioning even when treatment is temporarily ceased) or detrimental (i.e. worse functioning while off medication), but both hypotheses currently lack empirical support. Here we investigate whether stimulant treatment history predicts long-term development of ADHD symptoms, social-emotional functioning or cognition, measured after medication wash-out. ADHD symptoms, social-emotional functioning and cognitive test performance were measured twice, 6 years apart, in two ADHD groups (stimulant-treated versus not stimulant-treated between baseline and follow-up). Groups were closely matched on baseline clinical and demographic variables (n = 148, 58% male, age = 11.1). A matched healthy control group was included for reference. All but two outcome measures (emotional problems and prosocial behaviour) improved between baseline and follow-up. Improvement over time in the stimulant-treated group did not differ from improvement in the not stimulant-treated group on any outcome measure. Stimulant treatment is not associated with the long-term developmental course of ADHD symptoms, social-emotional functioning, motor control, timing or verbal working memory. Adolescence is characterised by clinical improvement regardless of stimulant treatment during that time. These findings are an important source to inform the scientific and public debate.
NASA Astrophysics Data System (ADS)
Braban, Christine; Tang, Sim; Poskitt, Janet; Van Dijk, Netty; Leeson, Sarah; Dragosits, Ulli; Hutchings, Torben; Twigg, Marsailidh; Di Marco, Chiara; Langford, Ben; Tremper, Anja; Nemitz, Eiko; Sutton, Mark
2017-04-01
Emissions of ammonia affect both rural and urban air quality primarily via reaction of ammonia in the atmosphere forming secondary ammonium salts in particulate matter (PM). Urban ammonia emissions come from a variety of sources including biological decomposition, human waste, industrial processes and combustion engines. In the UK, the only long-term urban ammonia measurement is a UK National Ammonia Monitoring Network site at London Cromwell Road, recording monthly average concentrations. Short term measurements have also been made in the past decade at Marylebone Road, North Kensington and on the BT Tower. Cromwell Road is a kerbside site operational since 1999. The Cromwell Road data indicates that ammonia concentrations may be increasing since 2010-2012 after a long period of decreasing. Data from the National Atmospheric Emissions Inventory indicates ammonia emissions from diesel fleet exhausts increasing over this time period but an overall net decrease in ammonia emissions. With changes in engine and exhaust technology to minimise pollutant emissions and the importance of ammonia as a precursor gas for secondary PM, there is a challenge to understand urban ammonia concentrations and subsequent impacts on urban air quality. In this paper the long term measurements are assessed in conjunction with the short-term measurements.The challenges to assess the relative importance of local versus long range ammonia emission are discussed.
Evaluation of Chemistry-Climate Model Results using Long-Term Satellite and Ground-Based Data
NASA Technical Reports Server (NTRS)
Stolarski, Richard S.
2005-01-01
Chemistry-climate models attempt to bring together our best knowledge of the key processes that govern the composition of the atmosphere and its response to changes in forcing. We test these models on a process by process basis by comparing model results to data from many sources. A more difficult task is testing the model response to changes. One way to do this is to use the natural and anthropogenic experiments that have been done on the atmosphere and are continuing to be done. These include the volcanic eruptions of El Chichon and Pinatubo, the solar cycle, and the injection of chlorine and bromine from CFCs and methyl bromide. The test of the model's response to these experiments is their ability to produce the long-term variations in ozone and the trace gases that affect ozone. We now have more than 25 years of satellite ozone data. We have more than 15 years of satellite and ground-based data of HC1, HN03, and many other gases. I will discuss the testing of models using long-term satellite data sets, long-term measurements from the Network for Detection of Stratospheric Change (NDSC) , long-term ground-based measurements of ozone.
NASA Astrophysics Data System (ADS)
Salameh, T.; Sauvage, S.; Afif, C.; Borbon, A.; Locoge, N.
2015-10-01
We applied the Positive Matrix Factorization model to two large datasets collected during two intensive measurement campaigns (summer 2011 and winter 2012) at a sub-urban site in Beirut, Lebanon, in order to identify NMHC sources and quantify their contribution to ambient levels. Six factors were identified in winter and five factors in summer. PMF-resolved source profiles were consistent with source profiles established by near-field measurements. The major sources were traffic-related emissions (combustion and gasoline evaporation) in winter and in summer accounting for 51 and 74 wt % respectively in agreement with the national emission inventory. The gasoline evaporation related to traffic source had a significant contribution regardless of the season (22 wt % in winter and 30 wt % in summer). The NMHC emissions from road transport are estimated from observations and PMF results, and compared to local and global emission inventories. The national road transport inventory shows lowest emissions than the ones from PMF but with a reasonable difference lower than 50 %. Global inventories show higher discrepancies with lower emissions up to a factor of 10 for the transportation sector. When combining emission inventory to our results, there is a strong evidence that control measures in Lebanon should be targeted on mitigating the NMHC emissions from the traffic-related sources. From a global perspective, an assessment of VOC anthropogenic emission inventories for the Middle East region as a whole seems necessary as these emissions could be much higher than expected at least from the road transport sector. Highlights: - PMF model was applied to identify major NMHC sources and their seasonal variation. - Gasoline evaporation accounts for more than 40 % both in winter and in summer. - NMHC urban emissions are dominated by traffic related sources in both seasons. - Agreement with the emission inventory regarding the relative contribution of the on-road mobile source but disagreement in terms of emission quantities suggesting an underestimation of the inventories.
This project will transform MESA Air from its current focus on PM2.5 into a multi-pollutant study that can meaningfully investigate the impact of traffic-derived air pollution on cardiovascular health using a source-to-exposure approach. We will integrate data on tr...
R. S. Ahl; S. W. Woods
2006-01-01
Changes in the extent, composition, and configuration of forest cover over time due to succession or disturbance processes can result in measurable changes in streamflow and water yield. Removal of forest cover generally increases streamflow due to reduced canopy interception and evapotranspiration. In watersheds where snow is the dominant source of water, yield...
The effect of storage temperature and duration on northern red oak acorn viability and vigour
Thomas L. Noland; Andree E. Morneault; Daniel C. Dey; Dave Deugo
2013-01-01
Three separate collections of Ontario sources of northern red oak (Quercus rubra L.) acorns were made to determine the effects of long-term cold storage at +2°C, -1°C, and -2°C on their viability and vigour. We measured acorn moisture content, percent germination during storage, speed of germination and total germination...
Electric dipole moments of light nuclei from {chi}EFT
DOE Office of Scientific and Technical Information (OSTI.GOV)
Higa, Renato
I present recent calculations of EDMs of light nuclei using chiral effective field theory techniques. At leading-order, we argue that they can be expressed in terms of six CP-violating low-energy constants. With our expressions, eventual non-zero measurements of EDMs of deuteron, helion, and triton can be combined to disentangle the different sources of CP-violation.
Electric dipole moments of light nuclei from χEFT
NASA Astrophysics Data System (ADS)
Higa, Renato
2013-03-01
I present recent calculations of EDMs of light nuclei using chiral effective field theory techniques. At leading-order, we argue that they can be expressed in terms of six CP-violating low-energy constants. With our expressions, eventual non-zero measurements of EDMs of deuteron, helion, and triton can be combined to disentangle the different sources of CP-violation.
Galaxy–galaxy lensing estimators and their covariance properties
Singh, Sukhdeep; Mandelbaum, Rachel; Seljak, Uros; ...
2017-07-21
Here, we study the covariance properties of real space correlation function estimators – primarily galaxy–shear correlations, or galaxy–galaxy lensing – using SDSS data for both shear catalogues and lenses (specifically the BOSS LOWZ sample). Using mock catalogues of lenses and sources, we disentangle the various contributions to the covariance matrix and compare them with a simple analytical model. We show that not subtracting the lensing measurement around random points from the measurement around the lens sample is equivalent to performing the measurement using the lens density field instead of the lens overdensity field. While the measurement using the lens densitymore » field is unbiased (in the absence of systematics), its error is significantly larger due to an additional term in the covariance. Therefore, this subtraction should be performed regardless of its beneficial effects on systematics. Comparing the error estimates from data and mocks for estimators that involve the overdensity, we find that the errors are dominated by the shape noise and lens clustering, which empirically estimated covariances (jackknife and standard deviation across mocks) that are consistent with theoretical estimates, and that both the connected parts of the four-point function and the supersample covariance can be neglected for the current levels of noise. While the trade-off between different terms in the covariance depends on the survey configuration (area, source number density), the diagnostics that we use in this work should be useful for future works to test their empirically determined covariances.« less
Galaxy–galaxy lensing estimators and their covariance properties
DOE Office of Scientific and Technical Information (OSTI.GOV)
Singh, Sukhdeep; Mandelbaum, Rachel; Seljak, Uros
Here, we study the covariance properties of real space correlation function estimators – primarily galaxy–shear correlations, or galaxy–galaxy lensing – using SDSS data for both shear catalogues and lenses (specifically the BOSS LOWZ sample). Using mock catalogues of lenses and sources, we disentangle the various contributions to the covariance matrix and compare them with a simple analytical model. We show that not subtracting the lensing measurement around random points from the measurement around the lens sample is equivalent to performing the measurement using the lens density field instead of the lens overdensity field. While the measurement using the lens densitymore » field is unbiased (in the absence of systematics), its error is significantly larger due to an additional term in the covariance. Therefore, this subtraction should be performed regardless of its beneficial effects on systematics. Comparing the error estimates from data and mocks for estimators that involve the overdensity, we find that the errors are dominated by the shape noise and lens clustering, which empirically estimated covariances (jackknife and standard deviation across mocks) that are consistent with theoretical estimates, and that both the connected parts of the four-point function and the supersample covariance can be neglected for the current levels of noise. While the trade-off between different terms in the covariance depends on the survey configuration (area, source number density), the diagnostics that we use in this work should be useful for future works to test their empirically determined covariances.« less
Galaxy-galaxy lensing estimators and their covariance properties
NASA Astrophysics Data System (ADS)
Singh, Sukhdeep; Mandelbaum, Rachel; Seljak, Uroš; Slosar, Anže; Vazquez Gonzalez, Jose
2017-11-01
We study the covariance properties of real space correlation function estimators - primarily galaxy-shear correlations, or galaxy-galaxy lensing - using SDSS data for both shear catalogues and lenses (specifically the BOSS LOWZ sample). Using mock catalogues of lenses and sources, we disentangle the various contributions to the covariance matrix and compare them with a simple analytical model. We show that not subtracting the lensing measurement around random points from the measurement around the lens sample is equivalent to performing the measurement using the lens density field instead of the lens overdensity field. While the measurement using the lens density field is unbiased (in the absence of systematics), its error is significantly larger due to an additional term in the covariance. Therefore, this subtraction should be performed regardless of its beneficial effects on systematics. Comparing the error estimates from data and mocks for estimators that involve the overdensity, we find that the errors are dominated by the shape noise and lens clustering, which empirically estimated covariances (jackknife and standard deviation across mocks) that are consistent with theoretical estimates, and that both the connected parts of the four-point function and the supersample covariance can be neglected for the current levels of noise. While the trade-off between different terms in the covariance depends on the survey configuration (area, source number density), the diagnostics that we use in this work should be useful for future works to test their empirically determined covariances.
Measuring Nursing Care Time and Tasks in Long-Term Services and Supports: One Size Does Not Fit All
Sochalski, Julie A.; Foust, Janice B.; Zubritsky, Cynthia D.; Hirschman, Karen B.; Abbott, Katherine M.; Naylor, Mary D.
2015-01-01
Background Although nursing care personnel comprise the majority of staff in long-term care services and supports (LTSS), a method for measuring the provision of nursing care has not yet been developed. Purpose/Methods We sought to understand the challenges of measuring nursing care across different types of LTSS using a qualitative approach that included the triangulation of data from three unique sources. Results Six primary challenges to measuring nursing care across LTSS emerged: level of detail about time of day, amount of time, or type of tasks varied by type of nursing and organization; time and tasks were documented in clinical records and administrative databases; data existed both on paper and electronically; several sources of information were needed to create the fullest picture of nursing care; data was inconsistently available for contracted providers; documentation of informal caregiving was unavailable. Differences were observed for assisted living facilities and home and community based services compared to nursing homes and across organizations within a setting. A commonality across settings and organizations was the availability of an electronically stored care plan specifying individual needs but not necessarily how these would be met. Conclusions The findings demonstrate the variability of data availability and specificity across three distinct LTSS settings. This study is an initial step toward establishing a process for measuring the provision of nursing care across LTSS to be able to explore the range of nursing care needs of LTSS recipients and how these needs are fulfilled. PMID:22902975
NASA Astrophysics Data System (ADS)
Warneke, C.; Geiger, F.; Edwards, P. M.; Dube, W.; Pétron, G.; Kofler, J.; Zahn, A.; Brown, S. S.; Graus, M.; Gilman, J.; Lerner, B.; Peischl, J.; Ryerson, T. B.; de Gouw, J. A.; Roberts, J. M.
2014-05-01
The emissions of volatile organic compounds (VOCs) associated with oil and natural gas production in the Uinta Basin, Utah were measured at a ground site in Horse Pool and from a NOAA mobile laboratory with PTR-MS instruments. The VOC compositions in the vicinity of individual gas and oil wells and other point sources such as evaporation ponds, compressor stations and injection wells are compared to the measurements at Horse Pool. High mixing ratios of aromatics, alkanes, cycloalkanes and methanol were observed for extended periods of time and short-term spikes caused by local point sources. The mixing ratios during the time the mobile laboratory spent on the well pads were averaged. High mixing ratios were found close to all point sources, but gas wells using dry-gas collection, which means dehydration happens at the well, were clearly associated with higher mixing ratios than other wells. Another large source was the flowback pond near a recently hydraulically re-fractured gas well. The comparison of the VOC composition of the emissions from the oil and natural gas wells showed that wet gas collection wells compared well with the majority of the data at Horse Pool and that oil wells compared well with the rest of the ground site data. Oil wells on average emit heavier compounds than gas wells. The mobile laboratory measurements confirm the results from an emissions inventory: the main VOC source categories from individual point sources are dehydrators, oil and condensate tank flashing and pneumatic devices and pumps. Raw natural gas is emitted from the pneumatic devices and pumps and heavier VOC mixes from the tank flashings.
Nikoloski, Zoran
2015-01-01
Plants as sessile organisms cannot escape their environment and have to adapt to any changes in the availability of sunlight and nutrients. The quantification of synthesis costs of metabolites, in terms of consumed energy, is a prerequisite to understand trade-offs arising from energetic limitations. Here, we examine the energy consumption of amino acid synthesis in Arabidopsis thaliana. To quantify these costs in terms of the energy equivalent ATP, we introduce an improved cost measure based on flux balance analysis and apply it to three state-of-the-art metabolic reconstructions to ensure robust results. We present the first systematic in silico analysis of the effect of nitrogen supply (nitrate/ammonium) on individual amino acid synthesis costs as well as of the effect of photoautotrophic and heterotrophic growth conditions, integrating day/night-specific regulation. Our results identify nitrogen supply as a key determinant of amino acid costs, in agreement with experimental evidence. In addition, the association of the determined costs with experimentally observed growth patterns suggests that metabolite synthesis costs are involved in shaping regulation of plant growth. Finally, we find that simultaneous uptake of both nitrogen sources can lead to efficient utilization of energy source, which may be the result of evolutionary optimization. PMID:25706533
Arnold, Anne; Sajitz-Hermstein, Max; Nikoloski, Zoran
2015-01-01
Plants as sessile organisms cannot escape their environment and have to adapt to any changes in the availability of sunlight and nutrients. The quantification of synthesis costs of metabolites, in terms of consumed energy, is a prerequisite to understand trade-offs arising from energetic limitations. Here, we examine the energy consumption of amino acid synthesis in Arabidopsis thaliana. To quantify these costs in terms of the energy equivalent ATP, we introduce an improved cost measure based on flux balance analysis and apply it to three state-of-the-art metabolic reconstructions to ensure robust results. We present the first systematic in silico analysis of the effect of nitrogen supply (nitrate/ammonium) on individual amino acid synthesis costs as well as of the effect of photoautotrophic and heterotrophic growth conditions, integrating day/night-specific regulation. Our results identify nitrogen supply as a key determinant of amino acid costs, in agreement with experimental evidence. In addition, the association of the determined costs with experimentally observed growth patterns suggests that metabolite synthesis costs are involved in shaping regulation of plant growth. Finally, we find that simultaneous uptake of both nitrogen sources can lead to efficient utilization of energy source, which may be the result of evolutionary optimization.
Source Term Model for Vortex Generator Vanes in a Navier-Stokes Computer Code
NASA Technical Reports Server (NTRS)
Waithe, Kenrick A.
2004-01-01
A source term model for an array of vortex generators was implemented into a non-proprietary Navier-Stokes computer code, OVERFLOW. The source term models the side force created by a vortex generator vane. The model is obtained by introducing a side force to the momentum and energy equations that can adjust its strength automatically based on the local flow. The model was tested and calibrated by comparing data from numerical simulations and experiments of a single low profile vortex generator vane on a flat plate. In addition, the model was compared to experimental data of an S-duct with 22 co-rotating, low profile vortex generators. The source term model allowed a grid reduction of about seventy percent when compared with the numerical simulations performed on a fully gridded vortex generator on a flat plate without adversely affecting the development and capture of the vortex created. The source term model was able to predict the shape and size of the stream-wise vorticity and velocity contours very well when compared with both numerical simulations and experimental data. The peak vorticity and its location were also predicted very well when compared to numerical simulations and experimental data. The circulation predicted by the source term model matches the prediction of the numerical simulation. The source term model predicted the engine fan face distortion and total pressure recovery of the S-duct with 22 co-rotating vortex generators very well. The source term model allows a researcher to quickly investigate different locations of individual or a row of vortex generators. The researcher is able to conduct a preliminary investigation with minimal grid generation and computational time.
COMPARISON OF NONLINEAR DYNAMICS OPTIMIZATION METHODS FOR APS-U
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sun, Y.; Borland, Michael
Many different objectives and genetic algorithms have been proposed for storage ring nonlinear dynamics performance optimization. These optimization objectives include nonlinear chromaticities and driving/detuning terms, on-momentum and off-momentum dynamic acceptance, chromatic detuning, local momentum acceptance, variation of transverse invariant, Touschek lifetime, etc. In this paper, the effectiveness of several different optimization methods and objectives are compared for the nonlinear beam dynamics optimization of the Advanced Photon Source upgrade (APS-U) lattice. The optimized solutions from these different methods are preliminarily compared in terms of the dynamic acceptance, local momentum acceptance, chromatic detuning, and other performance measures.
Treatment of solid tumors by interstitial release of recoiling short-lived alpha emitters
NASA Astrophysics Data System (ADS)
Arazi, L.; Cooks, T.; Schmidt, M.; Keisari, Y.; Kelson, I.
2007-08-01
A new method utilizing alpha particles to treat solid tumors is presented. Tumors are treated with interstitial radioactive sources which continually release short-lived alpha emitting atoms from their surface. The atoms disperse inside the tumor, delivering a high dose through their alpha decays. We implement this scheme using thin wire sources impregnated with 224Ra, which release by recoil 220Rn, 216Po and 212Pb atoms. This work aims to demonstrate the feasibility of our method by measuring the activity patterns of the released radionuclides in experimental tumors. Sources carrying 224Ra activities in the range 10-130 kBq were used in experiments on murine squamous cell carcinoma tumors. These included gamma spectroscopy of the dissected tumors and major organs, Fuji-plate autoradiography of histological tumor sections and tissue damage detection by Hematoxylin-Eosin staining. The measurements focused on 212Pb and 212Bi. The 220Rn/216Po distribution was treated theoretically using a simple diffusion model. A simplified scheme was used to convert measured 212Pb activities to absorbed dose estimates. Both physical and histological measurements confirmed the formation of a 5-7 mm diameter necrotic region receiving a therapeutic alpha-particle dose around the source. The necrotic regions shape closely corresponded to the measured activity patterns. 212Pb was found to leave the tumor through the blood at a rate which decreased with tumor mass. Our results suggest that the proposed method, termed DART (diffusing alpha-emitters radiation therapy), may potentially be useful for the treatment of human patients.
Comparisons of thermospheric density data sets and models
NASA Astrophysics Data System (ADS)
Doornbos, Eelco; van Helleputte, Tom; Emmert, John; Drob, Douglas; Bowman, Bruce R.; Pilinski, Marcin
During the past decade, continuous long-term data sets of thermospheric density have become available to researchers. These data sets have been derived from accelerometer measurements made by the CHAMP and GRACE satellites and from Space Surveillance Network (SSN) tracking data and related Two-Line Element (TLE) sets. These data have already resulted in a large number of publications on physical interpretation and improvement of empirical density modelling. This study compares four different density data sets and two empirical density models, for the period 2002-2009. These data sources are the CHAMP (1) and GRACE (2) accelerometer measurements, the long-term database of densities derived from TLE data (3), the High Accuracy Satellite Drag Model (4) run by Air Force Space Command, calibrated using SSN data, and the NRLMSISE-00 (5) and Jacchia-Bowman 2008 (6) empirical models. In describing these data sets and models, specific attention is given to differences in the geo-metrical and aerodynamic satellite modelling, applied in the conversion from drag to density measurements, which are main sources of density biases. The differences in temporal and spa-tial resolution of the density data sources are also described and taken into account. With these aspects in mind, statistics of density comparisons have been computed, both as a function of solar and geomagnetic activity levels, and as a function of latitude and local solar time. These statistics give a detailed view of the relative accuracy of the different data sets and of the biases between them. The differences are analysed with the aim at providing rough error bars on the data and models and pinpointing issues which could receive attention in future iterations of data processing algorithms and in future model development.
NASA Astrophysics Data System (ADS)
Korsmeier, Michael; Donato, Fiorenza; Di Mauro, Mattia
2018-05-01
The cosmic-ray flux of antiprotons is measured with high precision by the space-borne particle spectrometers AMS-02. Its interpretation requires a correct description of the dominant production process for antiprotons in our Galaxy, namely, the interaction of cosmic-ray proton and helium with the interstellar medium. In light of new cross section measurements by the NA61 experiment of p +p →p ¯+X and the first ever measurement of p +He →p ¯+X by the LHCb experiment, we update the parametrization of proton-proton and proton-nucleon cross sections. We find that the LHCb p He data constrain a shape for the cross section at high energies and show for the first time how well the rescaling from the p p channel applies to a helium target. By using p p , p He and p C data we estimate the uncertainty on the Lorentz invariant cross section for p +He →p ¯+X . We use these new cross sections to compute the source term for all the production channels, considering also nuclei heavier than He both in cosmic rays and the interstellar medium. The uncertainties on the total source term are up to ±20 % and slightly increase below antiproton energies of 5 GeV. This uncertainty is dominated by the p +p →p ¯+X cross section, which translates into all channels since we derive them using the p p cross sections. The cross sections to calculate the source spectra from all relevant cosmic-ray isotopes are provided in Supplemental Material. We finally quantify the necessity of new data on antiproton production cross sections, and pin down the kinematic parameter space which should be covered by future data.
An Empirical Temperature Variance Source Model in Heated Jets
NASA Technical Reports Server (NTRS)
Khavaran, Abbas; Bridges, James
2012-01-01
An acoustic analogy approach is implemented that models the sources of jet noise in heated jets. The equivalent sources of turbulent mixing noise are recognized as the differences between the fluctuating and Favre-averaged Reynolds stresses and enthalpy fluxes. While in a conventional acoustic analogy only Reynolds stress components are scrutinized for their noise generation properties, it is now accepted that a comprehensive source model should include the additional entropy source term. Following Goldstein s generalized acoustic analogy, the set of Euler equations are divided into two sets of equations that govern a non-radiating base flow plus its residual components. When the base flow is considered as a locally parallel mean flow, the residual equations may be rearranged to form an inhomogeneous third-order wave equation. A general solution is written subsequently using a Green s function method while all non-linear terms are treated as the equivalent sources of aerodynamic sound and are modeled accordingly. In a previous study, a specialized Reynolds-averaged Navier-Stokes (RANS) solver was implemented to compute the variance of thermal fluctuations that determine the enthalpy flux source strength. The main objective here is to present an empirical model capable of providing a reasonable estimate of the stagnation temperature variance in a jet. Such a model is parameterized as a function of the mean stagnation temperature gradient in the jet, and is evaluated using commonly available RANS solvers. The ensuing thermal source distribution is compared with measurements as well as computational result from a dedicated RANS solver that employs an enthalpy variance and dissipation rate model. Turbulent mixing noise predictions are presented for a wide range of jet temperature ratios from 1.0 to 3.20.
Evaluation of Mobile Source Emissions and Trends
NASA Astrophysics Data System (ADS)
Dallmann, Timothy Ryan
Mobile sources contribute significantly to air pollution problems. Relevant pollutants include numerous gaseous and particle-phase species that can affect human health, ecosystems, and climate. Accurate inventories of emissions from these sources are needed to help understand possible adverse impacts, and to develop effective air quality management strategies. Unfortunately large uncertainties persist in the understanding of mobile source emissions, and how these emissions are changing over time. This dissertation aims to evaluate long-term trends in mobile source emissions in the United States, and to make detailed measurements of emissions from present-day fleets of on-road vehicles operating in California. Long-term trends in mobile source emissions of nitrogen oxides (NO x) and fine particulate matter (PM2.5) in the United States were investigated through development of a fuel-based emission inventory. Annual emissions from on- and off-road gasoline and diesel engines were quantified for the years 1996-2006. Diesel engines were found to be the dominant mobile source of NOx and PM2.5, and on-road diesel vehicles were identified as the single largest anthropogenic source of NOx emissions in the United States as of 2005. The importance of diesel engines as a source of exhaust particulate matter emissions has led to the recent introduction of advanced emission control technologies in the United States, such as diesel particle filters (DPF), which have been required since 2007 for all new on-road heavy-duty (HD) diesel engines. In addition to national requirements for the use of such control devices on new engines, California has mandated accelerated clean-up of statewide emissions from older in-use diesel engines. The plume capture method was further applied to measure emissions from a more diverse population of trucks observed at the Caldecott tunnel in summer 2010. Emissions from hundreds of individual trucks were measured, and emission factor distributions were characterized for nitric oxide (NO), nitrogen dioxide (NO2), carbon monoxide (CO), formaldehyde, BC, as well as optical properties of the emitted particles. Emission factor distributions for all species were skewed, with a small fraction of trucks contributing disproportionately to total emissions. These findings confirm that the use of catalyzed DPF systems is leading to increased primary NO2 emissions. Absorption and scattering cross-section emission factors were used to calculate the aerosol single scattering albedo (SSA, at 532 nm) for individual truck exhaust plumes, which averaged 0.14 +/- 0.03. This value of aerosol SSA is very low compared to typical values (0.90-0.99) observed in ambient air studies. It is indicative of a strongly light-absorbing aerosol, due to the high BC emissions that are a characteristic feature of diesel exhaust PM emissions. Measurements at the Caldecott tunnel also included efforts to quantify light-duty (LD) gasoline vehicle emission factors, and further investigation of the relative contributions of on-road gasoline and diesel engines to air pollutant emissions. Measurements of CO, NOx, PM2.5, BC, and organic aerosol (OA) were made in a tunnel traffic bore where LD vehicles account for >99% of total traffic. Measured pollutant concentrations were apportioned between LD gasoline vehicles and diesel trucks, and fleet-average emission factors were quantified for LD gasoline vehicles using a carbon balance method. Diesel trucks contributed 18 +/- 3, 22 +/- 5, 44 +/- 8% of measured NOx, OA, and BC concentrations, respectively, despite accounting for <1% of total vehicles. Emission factors and overall fuel consumption for gasoline and diesel engines were used to describe the relative contributions of these sources to overall on-road vehicle emissions. Gasoline engines were found to be the dominant source of CO, an insignificant source of BC, and a relatively minor source of on-road OA emissions at urban, state, and national scales. Measurements at the Caldecott tunnel also featured use of a new high-resolution time-of-flight aerosol mass spectrometer, which was used to characterize the chemical composition of PM emitted by gasoline and diesel vehicles. Measurements of PM in the exhaust of individual HD trucks show a predominance of cyclyoalkane-derived ion signals relative to saturated alkane ion signals in the truck exhaust OA spectra, indicating that lubricating oil, rather than diesel fuel, was the dominant source of OA emitted by diesel trucks. This conclusion is supported by the presence of lubricant-derived trace elements in truck exhaust, emitted relative to total OA at levels that correspond to their weight fractions in bulk oil. Furthermore, comparison of mass spectra for sampling periods with varying levels of diesel influence found a high degree of similarity in the chemical composition of OA emitted by gasoline and diesel engines, suggesting a common lubricating oil rather than fuel-derived source for OA emissions. (Abstract shortened by UMI.).
A Well-Balanced Path-Integral f-Wave Method for Hyperbolic Problems with Source Terms
2014-01-01
Systems of hyperbolic partial differential equations with source terms (balance laws) arise in many applications where it is important to compute accurate time-dependent solutions modeling small perturbations of equilibrium solutions in which the source terms balance the hyperbolic part. The f-wave version of the wave-propagation algorithm is one approach, but requires the use of a particular averaged value of the source terms at each cell interface in order to be “well balanced” and exactly maintain steady states. A general approach to choosing this average is developed using the theory of path conservative methods. A scalar advection equation with a decay or growth term is introduced as a model problem for numerical experiments. PMID:24563581
NASA Astrophysics Data System (ADS)
Efthimiou, George C.; Kovalets, Ivan V.; Venetsanos, Alexandros; Andronopoulos, Spyros; Argyropoulos, Christos D.; Kakosimos, Konstantinos
2017-12-01
An improved inverse modelling method to estimate the location and the emission rate of an unknown point stationary source of passive atmospheric pollutant in a complex urban geometry is incorporated in the Computational Fluid Dynamics code ADREA-HF and presented in this paper. The key improvement in relation to the previous version of the method lies in a two-step segregated approach. At first only the source coordinates are analysed using a correlation function of measured and calculated concentrations. In the second step the source rate is identified by minimizing a quadratic cost function. The validation of the new algorithm is performed by simulating the MUST wind tunnel experiment. A grid-independent flow field solution is firstly attained by applying successive refinements of the computational mesh and the final wind flow is validated against the measurements quantitatively and qualitatively. The old and new versions of the source term estimation method are tested on a coarse and a fine mesh. The new method appeared to be more robust, giving satisfactory estimations of source location and emission rate on both grids. The performance of the old version of the method varied between failure and success and appeared to be sensitive to the selection of model error magnitude that needs to be inserted in its quadratic cost function. The performance of the method depends also on the number and the placement of sensors constituting the measurement network. Of significant interest for the practical application of the method in urban settings is the number of concentration sensors required to obtain a ;satisfactory; determination of the source. The probability of obtaining a satisfactory solution - according to specified criteria -by the new method has been assessed as function of the number of sensors that constitute the measurement network.
Repeat immigration: A previously unobserved source of heterogeneity?
Aradhya, Siddartha; Scott, Kirk; Smith, Christopher D
2017-07-01
Register data allow for nuanced analyses of heterogeneities between sub-groups which are not observable in other data sources. One heterogeneity for which register data is particularly useful is in identifying unique migration histories of immigrant populations, a group of interest across disciplines. Years since migration is a commonly used measure of integration in studies seeking to understand the outcomes of immigrants. This study constructs detailed migration histories to test whether misclassified migrations may mask important heterogeneities. In doing so, we identify a previously understudied group of migrants called repeat immigrants, and show that they differ systematically from permanent immigrants. In addition, we quantify the degree to which migration information is misreported in the registers. The analysis is carried out in two steps. First, we estimate income trajectories for repeat immigrants and permanent immigrants to understand the degree to which they differ. Second, we test data validity by cross-referencing migration information with changes in income to determine whether there are inconsistencies indicating misreporting. From the first part of the analysis, the results indicate that repeat immigrants systematically differ from permanent immigrants in terms of income trajectories. Furthermore, income trajectories differ based on the way in which years since migration is calculated. The second part of the analysis suggests that misreported migration events, while present, are negligible. Repeat immigrants differ in terms of income trajectories, and may differ in terms of other outcomes as well. Furthermore, this study underlines that Swedish registers provide a reliable data source to analyze groups which are unidentifiable in other data sources.
NASA Astrophysics Data System (ADS)
Vitillaro, Enzo
2017-03-01
The aim of this paper is to study the problem u_{tt}-Δ u+P(x,u_t)=f(x,u) quad & in (0,∞)×Ω, u=0 & on (0,∞)× Γ_0, u_{tt}+partial_ν u-Δ_Γ u+Q(x,u_t)=g(x,u)quad & on (0,∞)× Γ_1, u(0,x)=u_0(x),quad u_t(0,x)=u_1(x) & in overline Ω, where {Ω} is a open bounded subset of R^N with C 1 boundary ({N ≥ 2}), {Γ = partialΩ}, {(Γ0,Γ1)} is a measurable partition of {Γ}, {Δ_{Γ}} denotes the Laplace-Beltrami operator on {Γ}, {ν} is the outward normal to {Ω}, and the terms P and Q represent nonlinear damping terms, while f and g are nonlinear subcritical perturbations. In the paper a local Hadamard well-posedness result for initial data in the natural energy space associated to the problem is given. Moreover, when {Ω} is C 2 and {overline{Γ0} \\cap overline{Γ1} = emptyset}, the regularity of solutions is studied. Next a blow-up theorem is given when P and Q are linear and f and g are superlinear sources. Finally a dynamical system is generated when the source parts of f and g are at most linear at infinity, or they are dominated by the damping terms.
[Analysis of Multiplatform CO (Carbon Monoxide) Measurements During Trace-P Mission
NASA Technical Reports Server (NTRS)
Pougatchev, Nikita S.
2004-01-01
Carbon monoxide is considered mission critical (TRACE-P NRA) because it is one of the gases involved in controlling the oxidizing power of the atmosphere and, as a tracer gas, is valuable in interpreting mission data sets. Carbon monoxide exhibits interannual differences, suggesting relatively short-term imbalances in sources and sinks. Sources of CO are dominated by fossil fuel combustion, biomass burning, and the photochemical oxidation of CH4 and nonmethane hydrocarbons while reaction with OH is believed to be the major sink for atmospheric CO, with additional losses due to soil uptake. Uncertainties in the magnitude and distribution of both sources and sinks remain fairly large however, and additional data are required to refine the global budget. Seasonal changes and a northern hemispheric latitudinal gradient have been described for a variety of Pacific basin sites through long-term monitoring of surface background levels. Latitudinal variations have also recently been described at upper tropospheric altitudes over a multi-year period by. TRACE-P will provide an aircraft survey of CO over the northern Pacific in the northern spring when CO concentrations are at their seasonal maximum in the northern hemisphere (NH) and at their seasonal minimum in the southern hemisphere (SH). Previous GTE missions, Le., PEM West-B and PEM Tropics-B, ground-based, and satellite observations (MAPS, April 1994) give us a general picture of the distribution of CO over the northern Pacific during this season. Based on these measurements, background CO levels over remote ocean areas are anticipated to be in the range of 110 - 180 ppbv, while those closer to the Asian continent may rise as high as 600 ppbv. These measurements also reveal high spatial variability (both horizontal and vertical) as well as temporal variations in CO over the area planned for the TRACE-P mission. This variability is a result of multiple CO sources, the meteorological complexity of transport processes, and the photochemical aging of air masses. The influence of biomass burning in the southern Pacific should be relatively small since the mission coincides with the southern tropical wet season when agricultural burning is at its seasonal low. The proposed CO measurements taken during TRACE-P should therefore largely be a function of the impact of various NH sources, primarily Asian and predominantly fossil fuel combustion and biomass burning. These processes are also major sources of many other atmospheric pollutants, consequently making accurate and precise CO measurements is one of the highest TRACE-P priorities [TRACE-P NRA]. The TRACE-P mission emphasizes the dual objectives of assessing the magnitude of the transport of chemically and radiatively important gases such as CO from Asia to the western Pacific, and determining how emissions change and are modified during this transport.
Lee, Jaebeom; Lee, Young-Joo
2018-01-01
Management of the vertical long-term deflection of a high-speed railway bridge is a crucial factor to guarantee traffic safety and passenger comfort. Therefore, there have been efforts to predict the vertical deflection of a railway bridge based on physics-based models representing various influential factors to vertical deflection such as concrete creep and shrinkage. However, it is not an easy task because the vertical deflection of a railway bridge generally involves several sources of uncertainty. This paper proposes a probabilistic method that employs a Gaussian process to construct a model to predict the vertical deflection of a railway bridge based on actual vision-based measurement and temperature. To deal with the sources of uncertainty which may cause prediction errors, a Gaussian process is modeled with multiple kernels and hyperparameters. Once the hyperparameters are identified through the Gaussian process regression using training data, the proposed method provides a 95% prediction interval as well as a predictive mean about the vertical deflection of the bridge. The proposed method is applied to an arch bridge under operation for high-speed trains in South Korea. The analysis results obtained from the proposed method show good agreement with the actual measurement data on the vertical deflection of the example bridge, and the prediction results can be utilized for decision-making on railway bridge maintenance. PMID:29747421
Lee, Jaebeom; Lee, Kyoung-Chan; Lee, Young-Joo
2018-05-09
Management of the vertical long-term deflection of a high-speed railway bridge is a crucial factor to guarantee traffic safety and passenger comfort. Therefore, there have been efforts to predict the vertical deflection of a railway bridge based on physics-based models representing various influential factors to vertical deflection such as concrete creep and shrinkage. However, it is not an easy task because the vertical deflection of a railway bridge generally involves several sources of uncertainty. This paper proposes a probabilistic method that employs a Gaussian process to construct a model to predict the vertical deflection of a railway bridge based on actual vision-based measurement and temperature. To deal with the sources of uncertainty which may cause prediction errors, a Gaussian process is modeled with multiple kernels and hyperparameters. Once the hyperparameters are identified through the Gaussian process regression using training data, the proposed method provides a 95% prediction interval as well as a predictive mean about the vertical deflection of the bridge. The proposed method is applied to an arch bridge under operation for high-speed trains in South Korea. The analysis results obtained from the proposed method show good agreement with the actual measurement data on the vertical deflection of the example bridge, and the prediction results can be utilized for decision-making on railway bridge maintenance.
NASA Astrophysics Data System (ADS)
Edwards, L. L.; Harvey, T. F.; Freis, R. P.; Pitovranov, S. E.; Chernokozhin, E. V.
1992-10-01
The accuracy associated with assessing the environmental consequences of an accidental release of radioactivity is highly dependent on our knowledge of the source term characteristics and, in the case when the radioactivity is condensed on particles, the particle size distribution, all of which are generally poorly known. This paper reports on the development of a numerical technique that integrates the radiological measurements with atmospheric dispersion modeling. This results in a more accurate particle-size distribution and particle injection height estimation when compared with measurements of high explosive dispersal of (239)Pu. The estimation model is based on a non-linear least squares regression scheme coupled with the ARAC three-dimensional atmospheric dispersion models. The viability of the approach is evaluated by estimation of ADPIC model input parameters such as the ADPIC particle size mean aerodynamic diameter, the geometric standard deviation, and largest size. Additionally we estimate an optimal 'coupling coefficient' between the particles and an explosive cloud rise model. The experimental data are taken from the Clean Slate 1 field experiment conducted during 1963 at the Tonopah Test Range in Nevada. The regression technique optimizes the agreement between the measured and model predicted concentrations of (239)Pu by varying the model input parameters within their respective ranges of uncertainties. The technique generally estimated the measured concentrations within a factor of 1.5, with the worst estimate being within a factor of 5, very good in view of the complexity of the concentration measurements, the uncertainties associated with the meteorological data, and the limitations of the models. The best fit also suggest a smaller mean diameter and a smaller geometric standard deviation on the particle size as well as a slightly weaker particle to cloud coupling than previously reported.
Development of Performance Dashboards in Healthcare Sector: Key Practical Issues.
Ghazisaeidi, Marjan; Safdari, Reza; Torabi, Mashallah; Mirzaee, Mahboobeh; Farzi, Jebraeil; Goodini, Azadeh
2015-10-01
Static nature of performance reporting systems in health care sector has resulted in inconsistent, incomparable, time consuming, and static performance reports that are not able to transparently reflect a round picture of performance and effectively support healthcare managers' decision makings. So, the healthcare sector needs interactive performance management tools such as performance dashboards to measure, monitor, and manage performance more effectively. The aim of this article was to identify key issues that need to be addressed for developing high-quality performance dashboards in healthcare sector. A literature review was established to search electronic research databases, e-journals collections, and printed journals, books, dissertations, and theses for relevant articles. The search strategy interchangeably used the terms of "dashboard", "performance measurement system", and "executive information system" with the term of "design" combined with operator "AND". Search results (n=250) were adjusted for duplications, screened based on their abstract relevancy and full-text availability (n=147) and then assessed for eligibility (n=40). Eligible articles were included if they had explicitly focused on dashboards, performance measurement systems or executive information systems design. Finally, 28 relevant articles included in the study. Creating high-quality performance dashboards requires addressing both performance measurement and executive information systems design issues. Covering these two fields, identified contents were categorized to four main domains: KPIs development, Data Sources and data generation, Integration of dashboards to source systems, and Information presentation issues. This study implies the main steps to develop dashboards for the purpose of performance management. Performance dashboards developed on performance measurement and executive information systems principles and supported by proper back-end infrastructure will result in creation of dynamic reports that help healthcare managers to consistently measure the performance, continuously detect outliers, deeply analyze causes of poor performance, and effectively plan for the future.
Using NDACC column measurements of carbonyl sulfide to estimate its sources and sinks
NASA Astrophysics Data System (ADS)
Wang, Yuting; Marshall, Julia; Palm, Mathias; Deutscher, Nicholas; Roedenbeck, Christian; Warneke, Thorsten; Notholt, Justus; Baker, Ian; Berry, Joe; Suntharalingam, Parvadha; Jones, Nicholas; Mahieu, Emmanuel; Lejeune, Bernard; Hannigan, James; Conway, Stephanie; Strong, Kimberly; Campbell, Elliott; Wolf, Adam; Kremser, Stefanie
2016-04-01
Carbonyl sulfide (OCS) is taken up by plants during photosynthesis through a similar pathway as carbon dioxide (CO2), but is not emitted by respiration, and thus holds great promise as an additional constraint on the carbon cycle. It might act as a sort of tracer of photosynthesis, a way to separate gross primary productivity (GPP) from the net ecosystem exchange (NEE) that is typically derived from flux modeling. However the estimates of OCS sources and sinks still have significant uncertainties, which make it difficult to use OCS as a photosynthetic tracer, and the existing long-term surface-based measurements are sparse. The NDACC-IRWG measures the absorption of OCS in the atmosphere, and provides a potential long-term database of OCS total/partial columns, which can be used to evaluate OCS fluxes. We have retrieved OCS columns from several NDACC sites around the globe, and compared them to model simulation with OCS land fluxes based on the simple biosphere model (SiB). The disagreement between the measurements and the forward simulations indicates that (1) the OCS land fluxes from SiB are too low in the northern boreal region; (2) the ocean fluxes need to be optimized. A statistical linear flux model describing OCS is developed in the TM3 inversion system, and is used to estimate the OCS fluxes. We performed flux inversions using only NOAA OCS surface measurements as an observational constraint and with both surface and NDACC OCS column measurements, and assessed the differences. The posterior uncertainties of the inverted OCS fluxes decreased with the inclusion of NDACC data comparing to those using surface data only, and could be further reduced if more NDACC sites were included.
NASA Astrophysics Data System (ADS)
Sailhac, P.; Marquis, G.; Darnet, M.; Szalai, S.
2003-04-01
Surface self potential measurements (SP) are useful to characterize underground fluid flow or chemical reactions (as redox) and can be used in addition to NMR and electrical prospecting in hydrological investigations. Assuming that the SP anomalies have an electrokinetic origin, the source of SP data is the divergence of underground fluid flow; one important problem with surface SP data is then its interpretation in terms of fluid flow geometry. Some integral transform techniques have been shown to be powerful for SP interpretation (e.g. Fournier 1989, Patella, 1997; Sailhac &Marquis 2001). All these techniques are based upon Green’{ }s functions to characterize underground water flow, but they assume a constant electrical conductivity in the subsurface. This unrealistic approximation results in the appearance of non-electrokinetic sources at strong lateral electrical conductivity contrasts. We present here new Green’{ }s functions suitable for media of heterogeneous electrical conductivity. This new approach allows the joint interpretation of electrical resistivity tomography and SP measurements to detect electrokinetic sources caused by fluid flow. Tests on synthetic examples show that it gives more realistic results that when a constant electrical conductivity is assumed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Parworth, Caroline; Fast, Jerome D.; Mei, Fan
In this study the long-term trends of non-refractory submicrometer aerosol (NR-PM1) composition and mass concentration measured by an Aerosol Chemical Speciation Monitor (ACSM) at the U.S. Department of Energy’s Southern Great Plains (SGP) site are discussed. Over the period of 19 months (Nov. 20, 2010 – June 2012) highly time resolved (~30 min.) NR-PM1 data was recorded. Using this dataset the value-added product (VAP) of deriving organic aerosol components (OACOMP) is introduced. With this VAP, multivariate analysis of the measured organic mass spectral matrix can be performed on long term data to return organic aerosol (OA) factors that are associatedmore » with distinct sources, evolution processes, and physiochemical properties. Three factors were obtained from this VAP including two oxygenated OA (OOA) factors, differing in degrees of oxidation, and a biomass burning OA (BBOA) factor. Back trajectory analyses were performed to investigate possible sources of major NR-PM1 species at the SGP site. Organics dominated NR-PM1 mass concentration for the majority of the study with the exception of winter, when nitrate increased due to transport of precursor species from surrounding urban and agricultural areas and also due to cooler temperatures. Sulfate mass concentrations showed little seasonal variation with mixed regional and local sources. In the spring BBOA emissions increased and were mainly associated with local fires. Isoprene and carbon monoxide emission rates were computed by the Model of Emissions of Gases and Aerosols from Nature (MEGAN) to represent the spatial distribution of biogenic and anthropogenic sources, respectively. From this model there is evidence to support that biogenic emissions from the southeast contribute to SOA formation at the SGP site during the summer.« less
Paraskevopoulou, D; Liakakou, E; Gerasopoulos, E; Mihalopoulos, N
2015-09-15
To identify the sources of aerosols in Greater Athens Area (GAA), a total of 1510 daily samples of fine (PM 2.5) and coarse (PM 10-2,5) aerosols were collected at a suburban site (Penteli), during a five year period (May 2008-April 2013) corresponding to the period before and during the financial crisis. In addition, aerosol sampling was also conducted in parallel at an urban site (Thissio), during specific, short-term campaigns during all seasons. In all these samples mass and chemical composition measurements were performed, the latest only at the fine fraction. Particulate organic matter (POM) and ionic masses (IM) are the main contributors of aerosol mass, equally contributing by accounting for about 24% of the fine aerosol mass. In the IM, nss-SO4(-2) is the prevailing specie followed by NO3(-) and NH4(+) and shows a decreasing trend during the 2008-2013 period similar to that observed for PM masses. The contribution of water in fine aerosol is equally significant (21 ± 2%), while during dust transport, the contribution of dust increases from 7 ± 2% to 31 ± 9%. Source apportionment (PCA and PMF) and mass closure exercises identified the presence of six sources of fine aerosols: secondary photochemistry, primary combustion, soil, biomass burning, sea salt and traffic. Finally, from winter 2012 to winter 2013 the contribution of POM to the urban aerosol mass is increased by almost 30%, reflecting the impact of wood combustion (dominant fuel for domestic heating) to air quality in Athens, which massively started in winter 2013. Copyright © 2015 Elsevier B.V. All rights reserved.
Assessing risk of non-compliance of phosphorus standards for lakes in England and Wales
NASA Astrophysics Data System (ADS)
Duethmann, D.; Anthony, S.; Carvalho, L.; Spears, B.
2009-04-01
High population densities, use of inorganic fertilizer and intensive livestock agriculture have increased phosphorus loads to lakes, and accelerated eutrophication is a major pressure for many lakes. The EC Water Framework Directive (WFD) requires that good chemical and ecological quality is restored in all surface water bodies by 2015. Total phosphorus (TP) standards for lakes in England and Wales have been agreed recently, and our aim was to estimate what percentage of lakes in England and Wales is at risk of failing these standards. With measured lake phosphorus concentrations only being available for a small number of lakes, such an assessment had to be model based. The study also makes a source apportionment of phosphorus inputs into lakes. Phosphorus loads were estimated from a range of sources including agricultural loads, sewage effluents, septic tanks, diffuse urban sources, atmospheric deposition, groundwater and bank erosion. Lake phosphorus concentrations were predicted using the Vollenweider model, and the model framework was satisfactorily tested against available observed lake concentration data. Even though predictions for individual lakes remain uncertain, results for a population of lakes are considered as sufficiently robust. A scenario analysis was carried out to investigate to what extent reductions in phosphorus loads would increase the number of lakes achieving good ecological status in terms of TP standards. Applying the model to all lakes in England and Wales greater than 1 ha, it was calculated that under current conditions roughly two thirds of the lakes would fail the good ecological status with respect to phosphorus. According to our estimates, agricultural phosphorus loads represent the most frequent dominant source for the majority of catchments, but diffuse urban runoff also is important in many lakes. Sewage effluents are the most frequent dominant source for large lake catchments greater than 100 km². The evaluation in terms of total load can be misleading in terms of what sources need to be tackled by catchment management for most of the lakes. For example sewage effluents are responsible for the majority of the total load but are the dominant source in only a small number of larger lake catchments. If loads from all sources were halved this would potentially increase the number of complying lakes to two thirds but require substantial measures to reduce phosphorus inputs to lakes. For agriculture, required changes would have to go beyond improvements of agricultural practise, and need to include reducing the intensity of land use. The time required for many lakes to respond to reduced nutrient loading is likely to extend beyond the current timelines of the WFD due to internal loading and biological resistances.
26 CFR 1.737-1 - Recognition of precontribution gain.
Code of Federal Regulations, 2012 CFR
2012-04-01
... Property A1 and Property A2 is long-term, U.S.-source capital gain or loss. The character of gain on Property A3 is long-term, foreign-source capital gain. B contributes Property B, nondepreciable real... long-term, U.S.-source capital gain ($10,000 gain on Property A1 and $8,000 loss on Property A2) and $1...
Source term model evaluations for the low-level waste facility performance assessment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yim, M.S.; Su, S.I.
1995-12-31
The estimation of release of radionuclides from various waste forms to the bottom boundary of the waste disposal facility (source term) is one of the most important aspects of LLW facility performance assessment. In this work, several currently used source term models are comparatively evaluated for the release of carbon-14 based on a test case problem. The models compared include PRESTO-EPA-CPG, IMPACTS, DUST and NEFTRAN-II. Major differences in assumptions and approaches between the models are described and key parameters are identified through sensitivity analysis. The source term results from different models are compared and other concerns or suggestions are discussed.
Intercomparison of open-path trace gas measurements with two dual-frequency-comb spectrometers
Waxman, Eleanor M.; Cossel, Kevin C.; Truong, Gar-Wing; ...
2017-09-11
We present the first quantitative intercomparison between two open-path dual-comb spectroscopy (DCS) instruments which were operated across adjacent 2 km open-air paths over a 2-week period. We used DCS to measure the atmospheric absorption spectrum in the near infrared from 6023 to 6376 cm −1 (1568 to 1660 nm), corresponding to a 355 cm −1 bandwidth, at 0.0067 cm −1 sample spacing. The measured absorption spectra agree with each other to within 5 × 10 −4 in absorbance without any external calibration of either instrument. The absorption spectra are fit to retrieve path-integrated concentrations for carbon dioxide (CO 2), methane (CH 4), water (H 2O), and deuteratedmore » water (HDO). The retrieved dry mole fractions agree to 0.14 % (0.57 ppm) for CO 2, 0.35 % (7 ppb) for CH 4, and 0.40 % (36 ppm) for H 2O at ∼ 30 s integration time over the 2-week measurement campaign, which included 24 °C outdoor temperature variations and periods of strong atmospheric turbulence. This agreement is at least an order of magnitude better than conventional active-source open-path instrument intercomparisons and is particularly relevant to future regional flux measurements as it allows accurate comparisons of open-path DCS data across locations and time. We additionally compare the open-path DCS retrievals to a World Meteorological Organization (WMO)-calibrated cavity ring-down point sensor located along the path with good agreement. Short-term and long-term differences between the open-path DCS and point sensor are attributed, respectively, to spatial sampling discrepancies and to inaccuracies in the current spectral database used to fit the DCS data. Finally, the 2-week measurement campaign yields diurnal cycles of CO 2 and CH 4 that are consistent with the presence of local sources of CO 2 and absence of local sources of CH 4.« less
Intercomparison of open-path trace gas measurements with two dual-frequency-comb spectrometers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Waxman, Eleanor M.; Cossel, Kevin C.; Truong, Gar-Wing
We present the first quantitative intercomparison between two open-path dual-comb spectroscopy (DCS) instruments which were operated across adjacent 2 km open-air paths over a 2-week period. We used DCS to measure the atmospheric absorption spectrum in the near infrared from 6023 to 6376 cm −1 (1568 to 1660 nm), corresponding to a 355 cm −1 bandwidth, at 0.0067 cm −1 sample spacing. The measured absorption spectra agree with each other to within 5 × 10 −4 in absorbance without any external calibration of either instrument. The absorption spectra are fit to retrieve path-integrated concentrations for carbon dioxide (CO 2), methane (CH 4), water (H 2O), and deuteratedmore » water (HDO). The retrieved dry mole fractions agree to 0.14 % (0.57 ppm) for CO 2, 0.35 % (7 ppb) for CH 4, and 0.40 % (36 ppm) for H 2O at ∼ 30 s integration time over the 2-week measurement campaign, which included 24 °C outdoor temperature variations and periods of strong atmospheric turbulence. This agreement is at least an order of magnitude better than conventional active-source open-path instrument intercomparisons and is particularly relevant to future regional flux measurements as it allows accurate comparisons of open-path DCS data across locations and time. We additionally compare the open-path DCS retrievals to a World Meteorological Organization (WMO)-calibrated cavity ring-down point sensor located along the path with good agreement. Short-term and long-term differences between the open-path DCS and point sensor are attributed, respectively, to spatial sampling discrepancies and to inaccuracies in the current spectral database used to fit the DCS data. Finally, the 2-week measurement campaign yields diurnal cycles of CO 2 and CH 4 that are consistent with the presence of local sources of CO 2 and absence of local sources of CH 4.« less
Johnston, David W; Shields, Michael A; Siminski, Peter
2016-01-01
This paper estimates the long-term health effects of Vietnam-era military service using Australia's National conscription lotteries for identification. Our primary contribution is the quality and breadth of our health outcomes. We use several administrative sources, containing a near-universe of records on mortality (1994-2011), cancer diagnoses (1982-2008), and emergency hospital presentations (2005-2010). We also analyse a range of self-reported morbidity indicators (2006-2009). We find no significant long-term effects on mortality, cancer or emergency hospital visits. In contrast, we find significant detrimental effects on a number of morbidity measures. Hearing and mental health appear to be particularly affected. Copyright © 2015 Elsevier B.V. All rights reserved.
Jin, Virginia L; Schmer, Marty R; Stewart, Catherine E; Sindelar, Aaron J; Varvel, Gary E; Wienhold, Brian J
2017-07-01
Over the last 50 years, the most increase in cultivated land area globally has been due to a doubling of irrigated land. Long-term agronomic management impacts on soil organic carbon (SOC) stocks, soil greenhouse gas (GHG) emissions, and global warming potential (GWP) in irrigated systems, however, remain relatively unknown. Here, residue and tillage management effects were quantified by measuring soil nitrous oxide (N 2 O) and methane (CH 4 ) fluxes and SOC changes (ΔSOC) at a long-term, irrigated continuous corn (Zea mays L.) system in eastern Nebraska, United States. Management treatments began in 2002, and measured treatments included no or high stover removal (0 or 6.8 Mg DM ha -1 yr -1 , respectively) under no-till (NT) or conventional disk tillage (CT) with full irrigation (n = 4). Soil N 2 O and CH 4 fluxes were measured for five crop-years (2011-2015), and ΔSOC was determined on an equivalent mass basis to ~30 cm soil depth. Both area- and yield-scaled soil N 2 O emissions were greater with stover retention compared to removal and for CT compared to NT, with no interaction between stover and tillage practices. Methane comprised <1% of total emissions, with NT being CH 4 neutral and CT a CH 4 source. Surface SOC decreased with stover removal and with CT after 14 years of management. When ΔSOC, soil GHG emissions, and agronomic energy usage were used to calculate system GWP, all management systems were net GHG sources. Conservation practices (NT, stover retention) each decreased system GWP compared to conventional practices (CT, stover removal), but pairing conservation practices conferred no additional mitigation benefit. Although cropping system, management equipment/timing/history, soil type, location, weather, and the depth to which ΔSOC is measured affect the GWP outcomes of irrigated systems at large, this long-term irrigated study provides valuable empirical evidence of how management decisions can impact soil GHG emissions and surface SOC stocks. © 2017 John Wiley & Sons Ltd.
NASA Astrophysics Data System (ADS)
Starko, Darij; Craig, Walter
2018-04-01
Variations in redshift measurements of Type 1a supernovae and intensity observations from large sky surveys are an indicator of a component of acceleration in the rate of expansion of space-time. A key factor in the measurements is the intensity-distance relation for Maxwell's equations in Friedmann-Robertson-Walker (FRW) space-times. In view of future measurements of the decay of other fields on astronomical time and spatial scales, we determine the asymptotic behavior of the intensity-distance relationship for the solution of the wave equation in space-times with an FRW metric. This builds on previous work done on initial value problems for the wave equation in FRW space-time [Abbasi, B. and Craig, W., Proc. R. Soc. London, Ser. A 470, 20140361 (2014)]. In this paper, we focus on the precise intensity decay rates of the special cases for curvature k = 0 and k = -1, as well as giving a general derivation of the wave solution for -∞ < k < 0. We choose a Cauchy surface {(t, x) : t = t0 > 0} where t0 represents the time of an initial emission source, relative to the Big Bang singularity at t = 0. The initial data [g(x), h(x)] are assumed to be compactly supported; supp(g, h) ⊆ BR(0) and terms in the expression for the fundamental solution for the wave equation with the slowest decay rate are retained. The intensities calculated for coordinate time {t : t > 0} contain correction terms proportional to the ratio of t0 and the time differences ρ = t - t0. For the case of general curvature k, these expressions for the intensity reduce by scaling to the same form as for k = -1, from which we deduce the general formula. We note that for typical astronomical events such as Type 1a supernovae, the first order correction term for all curvatures -∞ < k < 0 is on the order of 10-4 smaller than the zeroth order term. These correction terms are small but may be significant in applications to alternative observations of cosmological space-time expansion rates.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kent Simmons, J.A.; Knap, A.H.
1991-04-01
The computer model Industrial Source Complex Short Term (ISCST) was used to study the stack emissions from a refuse incinerator proposed for the inland of Bermuda. The model predicts that the highest ground level pollutant concentrations will occur near Prospect, 800 m to 1,000 m due south of the stack. The authors installed a portable laboratory and instruments at Prospect to begin making air quality baseline measurements. By comparing the model's estimates of the incinerator contribution to the background levels measured at the site they predicted that stack emissions would not cause an increase in TSP or SO{sub 2}. Themore » incinerator will be a significant source of HCI to Bermuda air with ambient levels approaching air quality guidelines.« less
Dutton, Steven J.; Williams, Daniel E.; Garcia, Jessica K.; Vedal, Sverre; Hannigan, Michael P.
2009-01-01
Particulate matter less than 2.5 microns in diameter (PM2.5) has been shown to have a wide range of adverse health effects and consequently is regulated in accordance with the US-EPA’s National Ambient Air Quality Standards. PM2.5 originates from multiple primary sources and is also formed through secondary processes in the atmosphere. It is plausible that some sources form PM2.5 that is more toxic than PM2.5 from other sources. Identifying the responsible sources could provide insight into the biological mechanisms causing the observed health effects and provide a more efficient approach to regulation. This is the goal of the Denver Aerosol Sources and Health (DASH) study, a multi-year PM2.5 source apportionment and health study. The first step in apportioning the PM2.5 to different sources is to determine the chemical make-up of the PM2.5. This paper presents the methodology used during the DASH study for organic speciation of PM2.5. Specifically, methods are covered for solvent extraction of non-polar and semi-polar organic molecular markers using gas chromatography-mass spectrometry (GC-MS). Vast reductions in detection limits were obtained through the use of a programmable temperature vaporization (PTV) inlet along with other method improvements. Results are presented for the first 1.5 years of the DASH study revealing seasonal and source-related patterns in the molecular markers and their long-term correlation structure. Preliminary analysis suggests that point sources are not a significant contributor to the organic molecular markers measured at our receptor site. Several motor vehicle emission markers help identify a gasoline/diesel split in the ambient data. Findings show both similarities and differences when compared with other cities where similar measurements and assessments have been made. PMID:20161318
NASA Astrophysics Data System (ADS)
Kotthaus, S.; Grimmond, S.
2013-12-01
Global urbanisation brings increasingly dense and complex urban structures. To manage cities sustainably and smartly, currently and into the future under changing climates, urban climate research needs to advance in areas such as Central Business Districts (CBD) where human interactions with the environment are particularly concentrated. Measurement and modelling approaches may be pushed to their limits in dense urban settings, but if urban climate research is to contribute to the challenges of real cities those limits have to be addressed. The climate of cities is strongly governed by surface-atmosphere exchanges of energy, moisture and momentum. Observations of the relevant fluxes provide important information for improvement and evaluation of modelling approaches. Due to the CBD's heterogeneity, a very careful analysis of observations is required to understand the relevant processes. Current approaches used to interpret observations and set them in a wider context may need to be adapted for use in these more complex areas. Here, we present long-term observations of the radiation balance components and turbulent fluxes of latent heat, sensible heat and momentum in the city centre of London. This is one of the first measurement studies in a CBD covering multiple years with analysis at temporal scales from days to seasons. Data gathered at two sites in close vicinity, but with different measurement heights, are analysed to investigate the influence of source area characteristics on long-term radiation and turbulent fluxes. Challenges of source area modelling and the critical aspect of siting in such a complex environment are considered. Outgoing long- and short-wave radiation are impacted by the anisotropic nature of the urban surface and the high reflectance materials increasingly being used as building materials. Results highlight the need to consider the source area of radiometers in terms of diffuse and direct irradiance. Sensible heat fluxes (QH) are positive all year round, even at night. QH systematically exceeds input from net all-wave radiation (Q*), probably sustained by a both storage and anthropogenic heat fluxes (QF). Model estimates suggest QF can exceed the Q* nearly all year round. The positive QH inhibits stable conditions, but the stability classification is determined predominantly by the pattern of friction velocity over the rough urban surface. Turbulent latent heat flux variations are controlled (beyond the available energy) by rainfall due to the small vegetation cover. The Bowen ratio is mostly larger than one. Analysis of the eddy covariance footprint surface controls for the different land cover types by flow patterns for measurements at the two heights suggests the spatial variations of the sensible heat flux observed are partly related to changes in surface roughness, even at the local scale. Where the source areas are most homogeneous, flow conditions are vertically consistent - even if initial morphometric parameters suggested the measurements may be below the blending height. Turbulence statistics and momentum flux patterns prove useful for the interpretation of turbulent heat exchanges observed.
POI Summarization by Aesthetics Evaluation From Crowd Source Social Media.
Qian, Xueming; Li, Cheng; Lan, Ke; Hou, Xingsong; Li, Zhetao; Han, Junwei
2018-03-01
Place-of-Interest (POI) summarization by aesthetics evaluation can recommend a set of POI images to the user and it is significant in image retrieval. In this paper, we propose a system that summarizes a collection of POI images regarding both aesthetics and diversity of the distribution of cameras. First, we generate visual albums by a coarse-to-fine POI clustering approach and then generate 3D models for each album by the collected images from social media. Second, based on the 3D to 2D projection relationship, we select candidate photos in terms of the proposed crowd source saliency model. Third, in order to improve the performance of aesthetic measurement model, we propose a crowd-sourced saliency detection approach by exploring the distribution of salient regions in the 3D model. Then, we measure the composition aesthetics of each image and we explore crowd source salient feature to yield saliency map, based on which, we propose an adaptive image adoption approach. Finally, we combine the diversity and the aesthetics to recommend aesthetic pictures. Experimental results show that the proposed POI summarization approach can return images with diverse camera distributions and aesthetics.
Sources and Fate of Reactive Carbon over North America
NASA Astrophysics Data System (ADS)
Chen, X.; Millet, D. B.; Singh, H. B.; Wisthaler, A.
2016-12-01
We apply a high-resolution chemical transport model (GEOS-Chem CTM at 0.25°×0.3125°) to generate, a comprehensive gas-phase reactive carbon budget over North America. Based on state-of-science source inventories and known chemistry, we find in the model that biogenic sources dominate the overall reactive carbon budget, with 49, 15, 4, and 39 TgC, respectively, introduced to the North American atmosphere from the biosphere, anthropogenic sources, fires, and from methane oxidation in 2013. Biogenic and anthropogenic non-methane volatile organic compounds contribute 60% and 10%, respectively, to the total OH reactivity over the Southeast US, along with other contributions from methane and inorganics. Oxidation to CO and CO2 then represents the overwhelming fate of that reactive carbon, with 65, 15, 7 and 5 TgC, respectively, oxidized to produce CO/CO2, dry deposited, wet deposited and transported (net) out of North America. We confront this simulation with an ensemble of recent airborne measurements over North America (SEAC4RS, SENEX, DISCOVER-AQ, DC3) and interpret the model-measurement comparisons in terms of their implications for current understanding of atmospheric reactive carbon and the processes driving its distribution.
Source Term Model for Steady Micro Jets in a Navier-Stokes Computer Code
NASA Technical Reports Server (NTRS)
Waithe, Kenrick A.
2005-01-01
A source term model for steady micro jets was implemented into a non-proprietary Navier-Stokes computer code, OVERFLOW. The source term models the mass flow and momentum created by a steady blowing micro jet. The model is obtained by adding the momentum and mass flow created by the jet to the Navier-Stokes equations. The model was tested by comparing with data from numerical simulations of a single, steady micro jet on a flat plate in two and three dimensions. The source term model predicted the velocity distribution well compared to the two-dimensional plate using a steady mass flow boundary condition, which was used to simulate a steady micro jet. The model was also compared to two three-dimensional flat plate cases using a steady mass flow boundary condition to simulate a steady micro jet. The three-dimensional comparison included a case with a grid generated to capture the circular shape of the jet and a case without a grid generated for the micro jet. The case without the jet grid mimics the application of the source term. The source term model compared well with both of the three-dimensional cases. Comparisons of velocity distribution were made before and after the jet and Mach and vorticity contours were examined. The source term model allows a researcher to quickly investigate different locations of individual or several steady micro jets. The researcher is able to conduct a preliminary investigation with minimal grid generation and computational time.
Validation of an Open-Source Tool for Measuring Carotid Lumen Diameter and Intima-Media Thickness.
Manterola, Hugo Luis; Lo Vercio, Lucas; Díaz, Alejandro; Del Fresno, Mariana; Larrabide, Ignacio
2018-05-14
In low- and middle-income regions, a relatively large number of deaths occur from cardiovascular disease or stroke. Carotid intima-media thickness (cIMT) and carotid lumen diameter (cLD) are strong indicators of cardiovascular event risk and stenosis severity, respectively. The interactive open-source software described here, Cimtool, is based on active contours for measuring these indicators in clinical practice and thus helping in preventive diagnosis and treatment. Cimtool was validated using carotid phantoms and real images obtained using ultrasound. Expert users measured cIMT and cLD in regular practice and also with Cimtool. The results obtained with Cimtool were then compared with the results for the manual approach in terms of measurement agreement, time spent on the measurements and usability. Intra-observer variability when using Cimtool was also analyzed. Statistical analysis revealed strong agreement between the manual method and Cimtool (p > 0.01 for cIMT and cLD). The correlation coefficient for both cIMT and cLD measurements was r > 0.9. Moreover, this software allowed the users to spend considerably less time on each measurement (3.5 min per study versus 50 s with Cimtool on average). An open-source, interactive, validated tool for measuring cIMT and cLD clinically was thus developed. Compared with the manual approach, Cimtool's straightforward measurement flow allows the user to spend less time per measurement and has less standard deviation. The coefficients of variation for measurements and intra-observer variability were lower than those reported for recent automated approaches, even with low-quality images. Copyright © 2018 World Federation for Ultrasound in Medicine and Biology. Published by Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Obaid, Razib; Buth, Christian; Dakovski, Georgi L.
Here, we measured the fluorescence photon yield of neon upon soft x-ray ionization (~1200 eV) from the x-ray free-electron laser at Linac Coherent Light Source, and demonstrated the usage of a grazing incidence spectrometer with a variable line spacing grating to perform x-ray fluorescence spectroscopy on a gas phase system. Our measurements also allowed us to estimate the focal size of the beam from the theoretical description developed, in terms of the rate equation approximation accounting for photoionization shake off of neutral neon and double auger decay of single core holes.
Method for measuring multiple scattering corrections between liquid scintillators
DOE Office of Scientific and Technical Information (OSTI.GOV)
Verbeke, J. M.; Glenn, A. M.; Keefer, G. J.
2016-04-11
In this study, a time-of-flight method is proposed to experimentally quantify the fractions of neutrons scattering between scintillators. An array of scintillators is characterized in terms of crosstalk with this method by measuring a californium source, for different neutron energy thresholds. The spectral information recorded by the scintillators can be used to estimate the fractions of neutrons multiple scattering. With the help of a correction to Feynman's point model theory to account for multiple scattering, these fractions can in turn improve the mass reconstruction of fissile materials under investigation.
Final design of thermal diagnostic system in SPIDER ion source
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brombin, M., E-mail: matteo.brombin@igi.cnr.it; Dalla Palma, M.; Pasqualotto, R.
The prototype radio frequency source of the ITER heating neutral beams will be first tested in SPIDER test facility to optimize H{sup −} production, cesium dynamics, and overall plasma characteristics. Several diagnostics will allow to fully characterise the beam in terms of uniformity and divergence and the source, besides supporting a safe and controlled operation. In particular, thermal measurements will be used for beam monitoring and system protection. SPIDER will be instrumented with mineral insulated cable thermocouples, both on the grids, on other components of the beam source, and on the rear side of the beam dump water cooled elements.more » This paper deals with the final design and the technical specification of the thermal sensor diagnostic for SPIDER. In particular the layout of the diagnostic, together with the sensors distribution in the different components, the cables routing and the conditioning and acquisition cubicles are described.« less
Final design of thermal diagnostic system in SPIDER ion source
NASA Astrophysics Data System (ADS)
Brombin, M.; Dalla Palma, M.; Pasqualotto, R.; Pomaro, N.
2016-11-01
The prototype radio frequency source of the ITER heating neutral beams will be first tested in SPIDER test facility to optimize H- production, cesium dynamics, and overall plasma characteristics. Several diagnostics will allow to fully characterise the beam in terms of uniformity and divergence and the source, besides supporting a safe and controlled operation. In particular, thermal measurements will be used for beam monitoring and system protection. SPIDER will be instrumented with mineral insulated cable thermocouples, both on the grids, on other components of the beam source, and on the rear side of the beam dump water cooled elements. This paper deals with the final design and the technical specification of the thermal sensor diagnostic for SPIDER. In particular the layout of the diagnostic, together with the sensors distribution in the different components, the cables routing and the conditioning and acquisition cubicles are described.
Lineal energy calibration of mini tissue-equivalent gas-proportional counters (TEPC)
NASA Astrophysics Data System (ADS)
Conte, V.; Moro, D.; Grosswendt, B.; Colautti, P.
2013-07-01
Mini TEPCs are cylindrical gas proportional counters of 1 mm or less of sensitive volume diameter. The lineal energy calibration of these tiny counters can be performed with an external gamma-ray source. However, to do that, first a method to get a simple and precise spectral mark has to be found and then the keV/μm value of this mark. A precise method (less than 1% of uncertainty) to identify this markis described here, and the lineal energy value of this mark has been measured for different simulated site sizes by using a 137Cs gamma source and a cylindrical TEPC equipped with a precision internal 244Cm alpha-particle source, and filled with propane-based tissue-equivalent gas mixture. Mini TEPCs can be calibrated in terms of lineal energy, by exposing them to 137Cesium sources, with an overall uncertainty of about 5%.
NASA Astrophysics Data System (ADS)
Zhang, Xueqing; Sciacchitano, Andrea; Pröbsting, Stefan
2018-05-01
Particle image velocimetry for the experimental assessment of trailing edge noise sources has become focus of research in recent years. The present study investigates the feasibility of the noise prediction for high-lift devices based on time-resolved particle image velocimetry (PIV). The model under investigation is a NACA 0015 airfoil with a Gurney flap with a height of 6% of the chord length. The velocity fields around and downstream of the Gurney flap were measured by PIV and used to compute the corresponding pressure fields by solving the Poisson equation for incompressible flows. The reconstructed pressure fluctuations on the airfoil surface constitute the source term for Curle's aeroacoustic analogy, which was employed in both the distributed and compact formulation to estimate the noise emission from PIV. The results of the two formulations are compared with the simultaneous far-field microphone measurements in the temporal and spectral domains. Both formulations of Curle's analogy yield acoustic sound pressure levels in good agreement with the simultaneous microphone measurements for the tonal component. The estimated far-field sound power spectra (SPL) from the PIV measurements reproduce the peak at the vortex shedding frequency, which also agrees well with the acoustic measurements.
Soppa, Vanessa J; Schins, Roel P F; Hennig, Frauke; Nieuwenhuijsen, Mark J; Hellack, Bryan; Quass, Ulrich; Kaminski, Heinz; Sasse, Birgitta; Shinnawi, Samir; Kuhlbusch, Thomas A J; Hoffmann, Barbara
2017-10-01
Particulate air pollution is linked to adverse cardiovascular effects. The aim of the study was to investigate the effect of short-term exposure to indoor particles on blood pressure (BP). We analyzed the association of particle emissions from indoor sources (candle burning, toasting bread, frying sausages) with BP changes in 54 healthy volunteers in a randomized cross-over controlled exposure study. Particle mass concentration (PMC), size-specific particle number concentration (PNC) and lung-deposited particle surface area concentration (PSC) were measured during the 2h exposure. Systolic and diastolic blood pressure were measured before, during, directly, 2, 4 and 24h after exposure. We performed multiple mixed linear regression analyses of different particle metrics and BP. BP significantly increased with increasing PMC, PSC and PNC resulting from toasting bread. For example, an increase per 10µg/m 3 PM 10 and PM 2.5 , systolic BP increased at all time points with largest changes 1h after exposure initiation of 1.5mmHg (95%-CI: 1.1; 1.9) and of 2.2mmHg (95%-CI: 1.3; 3.1), respectively. Our study suggests an association of short-term exposure to fine and ultrafine particles emitted from toasting bread with increases in BP. Particles emitted from frying sausages and candle burning did not consistently affect BP. Copyright © 2017. Published by Elsevier Inc.
Contini, D; Donateo, A; Cesari, D; Belosi, F; Francioso, S
2010-09-01
Aerosol and gaseous pollution measurements were carried out at an urban background site in the south of Italy located near an industrial complex. Collection of 24 h samples of PM10 and PM2.5 and successive chemical quantification of metals were performed. Data were compared with measurements taken at a suburban background site, located at 25 km distance. The comparison showed the presence of an industrial contribution with a well defined chemical emission profile, similar, in terms of metals content, to urban emissions. As this made difficult the quantitative characterisation of the contribution of the two sources to atmospheric PM, a statistical method based on the treatment of data arising from high temporal resolution measurements was developed. Data were taken with a micrometeorological station based on an integrating nephelometer (Mie pDR-1200) for optical detection of PM2.5 concentration, with successive evaluation of vertical turbulent fluxes using the eddy-correlation method. Results show that the contribution from the two sources (urban emissions and industrial releases) have a very different behaviour, with the industrial contribution being present at high wind velocity with short concentration peaks (average duration 4 min) associated to strong positive and negative vertical fluxes. The estimated contribution to PM2.5 is 2.3% over long-term averages. The urban emissions are mainly present at low wind velocity, with longer concentration peaks in the morning and late evening hours, generally associated to small positive vertical fluxes. The characterisation of the contribution was performed using deposition velocity V(d) that is on average -3.5 mm s(-1) and has a diurnal pattern, with negligible values during the night and a minimum value of around -9 mm s(-1) late in the afternoon. Results show a correlation between V(d), friction velocity and wind velocity that could be the basis for a parameterisation of V(d) to be used in dispersion codes.
NASA Astrophysics Data System (ADS)
Bilbro, Griff L.; Hou, Danqiong; Yin, Hong; Trew, Robert J.
2009-02-01
We have quantitatively modeled the conduction current and charge storage of an HFET in terms its physical dimensions and material properties. For DC or small-signal RF operation, no adjustable parameters are necessary to predict the terminal characteristics of the device. Linear performance measures such as small-signal gain and input admittance can be predicted directly from the geometric structure and material properties assumed for the device design. We have validated our model at low-frequency against experimental I-V measurements and against two-dimensional device simulations. We discuss our recent extension of our model to include a larger class of electron velocity-field curves. We also discuss the recent reformulation of our model to facilitate its implementation in commercial large-signal high-frequency circuit simulators. Large signal RF operation is more complex. First, the highest CW microwave power is fundamentally bounded by a brief, reversible channel breakdown in each RF cycle. Second, the highest experimental measurements of efficiency, power, or linearity always require harmonic load pull and possibly also harmonic source pull. Presently, our model accounts for these facts with an adjustable breakdown voltage and with adjustable load impedances and source impedances for the fundamental frequency and its harmonics. This has allowed us to validate our model for large signal RF conditions by simultaneously fitting experimental measurements of output power, gain, and power added efficiency of real devices. We show that the resulting model can be used to compare alternative device designs in terms of their large signal performance, such as their output power at 1dB gain compression or their third order intercept points. In addition, the model provides insight into new device physics features enabled by the unprecedented current and voltage levels of AlGaN/GaN HFETs, including non-ohmic resistance in the source access regions and partial depletion of the 2DEG in the drain access region.
Long term measurements of light absorbing particles on tropical glaciers
NASA Astrophysics Data System (ADS)
Schmitt, C. G.; Sanchez Rodriguez, W.; Arnott, W. P.; All, J.; Schwarz, J. P.
2016-12-01
We present results of six years of measurements of light absorbing particles (LAP) on glaciers of the Cordillera Blanca mountain range in Peru. Tropical glaciers are important sources of water for human consumption, agriculture, and hydroelectric power in the region. Regular measurements in the dry season show that light absorbing particle concentrations are generally low (equivalent to the absorption equivalent of 5-30 nanograms of black carbon per gram of snow) during non-El Nino years while values increase substantially during the recent El Nino. Two years of monthly measurements at two glaciers show that fresh snow LAP concentration are very low while LAP levels increase dramatically during snow-less periods.
Surface Wave Dynamics in the Coastal Zone
2014-09-30
also collected from the Duck measurement site, operated by the USACE Field Research Facility at Duck , North Carolina. The collection and validation...similar analysis for 10 storm periods using wave data collected at Duck , North Carolina. The preparations consist of creating a dedicated unstructured...validated in the Southern North Sea and Duck validation studies. The shallow water source terms for wave breaking and triad interactions are being
ERIC Educational Resources Information Center
Florida State Advisory Council on Vocational and Technical Education, Tallahassee.
A study of 13 vocational and technical education programs in Florida was conducted which represented an attempt to identify valid and reliable sources of data whereby some vocational programs might be evaluated in specified terms of effectiveness. The programs selected for study were among those which require graduating students to pass licensing…
NASA Astrophysics Data System (ADS)
Jordan, Phil; Melland, Alice; Shore, Mairead; Mellander, Per-Erik; Shortle, Ger; Ryan, David; Crockford, Lucy; Macintosh, Katrina; Campbell, Julie; Arnscheidt, Joerg; Cassidy, Rachel
2014-05-01
A complete appraisal of material fluxes in flowing waters is really only possibly with high time resolution data synchronous with measurements of discharge. Defined by Kirchner et al. (2004; Hydrological Processes, 18/7) as the high-frequency wave of the future and with regard to disentangling signal noise from process pattern, this challenge has been met in terms of nutrient flux monitoring by automated bankside analysis. In Ireland over a ten-year period, time-series nutrient data collected on a sub-hourly basis in rivers have been used to distinguish fluxes from different catchment sources and pathways and to provide more certain temporal pictures of flux for the comparative definition of catchment nutrient dynamics. In catchments where nutrient fluxes are particularly high and exhibit a mix of extreme diffuse and point source influences, high time resolution data analysis indicates that there are no satisfactory statistical proxies for seasonal or annual flux predictions that use coarse datasets. Or at least exposes the limits of statistical approaches to catchment scale and hydrological response. This has profound implications for catchment monitoring programmes that rely on modelled relationships. However, using high resolution monitoring for long term assessments of catchment mitigation measures comes with further challenges. Sustaining continuous wet chemistry analysis at river stations is resource intensive in terms of capital, maintenance and quality assurance. Furthermore, big data capture requires investment in data management systems and analysis. These two institutional challenges are magnified when considering the extended time period required to identify the influences of land-based nutrient control measures on water based systems. Separating the 'climate signal' from the 'source signal' in river nutrient flux data is a major analysis challenge; more so when tackled with anything but higher resolution data. Nevertheless, there is scope to lower costs in bankside analysis through technology development, and the scientific advantages of these data are clear and exciting. When integrating its use with policy appraisal, it must be made clear that the advances in river process understanding from high resolution monitoring data capture come as a package with the ability to make more informed decisions through an investment in better information.
On estimating attenuation from the amplitude of the spectrally whitened ambient seismic field
NASA Astrophysics Data System (ADS)
Weemstra, Cornelis; Westra, Willem; Snieder, Roel; Boschi, Lapo
2014-06-01
Measuring attenuation on the basis of interferometric, receiver-receiver surface waves is a non-trivial task: the amplitude, more than the phase, of ensemble-averaged cross-correlations is strongly affected by non-uniformities in the ambient wavefield. In addition, ambient noise data are typically pre-processed in ways that affect the amplitude itself. Some authors have recently attempted to measure attenuation in receiver-receiver cross-correlations obtained after the usual pre-processing of seismic ambient-noise records, including, most notably, spectral whitening. Spectral whitening replaces the cross-spectrum with a unit amplitude spectrum. It is generally assumed that cross-terms have cancelled each other prior to spectral whitening. Cross-terms are peaks in the cross-correlation due to simultaneously acting noise sources, that is, spurious traveltime delays due to constructive interference of signal coming from different sources. Cancellation of these cross-terms is a requirement for the successful retrieval of interferometric receiver-receiver signal and results from ensemble averaging. In practice, ensemble averaging is replaced by integrating over sufficiently long time or averaging over several cross-correlation windows. Contrary to the general assumption, we show in this study that cross-terms are not required to cancel each other prior to spectral whitening, but may also cancel each other after the whitening procedure. Specifically, we derive an analytic approximation for the amplitude difference associated with the reversed order of cancellation and normalization. Our approximation shows that an amplitude decrease results from the reversed order. This decrease is predominantly non-linear at small receiver-receiver distances: at distances smaller than approximately two wavelengths, whitening prior to ensemble averaging causes a significantly stronger decay of the cross-spectrum.
Short-term dynamics of indoor and outdoor endotoxin exposure: Case of Santiago, Chile, 2012.
Barraza, Francisco; Jorquera, Héctor; Heyer, Johanna; Palma, Wilfredo; Edwards, Ana María; Muñoz, Marcelo; Valdivia, Gonzalo; Montoya, Lupita D
2016-01-01
Indoor and outdoor endotoxin in PM2.5 was measured for the very first time in Santiago, Chile, in spring 2012. Average endotoxin concentrations were 0.099 and 0.094 [EU/m(3)] for indoor (N=44) and outdoor (N=41) samples, respectively; the indoor-outdoor correlation (log-transformed concentrations) was low: R=-0.06, 95% CI: (-0.35 to 0.24), likely owing to outdoor spatial variability. A linear regression model explained 68% of variability in outdoor endotoxins, using as predictors elemental carbon (a proxy of traffic emissions), chlorine (a tracer of marine air masses reaching the city) and relative humidity (a modulator of surface emissions of dust, vegetation and garbage debris). In this study, for the first time a potential source contribution function (PSCF) was applied to outdoor endotoxin measurements. Wind trajectory analysis identified upwind agricultural sources as contributors to the short-term, outdoor endotoxin variability. Our results confirm an association between combustion particles from traffic and outdoor endotoxin concentrations. For indoor endotoxins, a predictive model was developed but it only explained 44% of endotoxin variability; the significant predictors were tracers of indoor PM2.5 dust (Si, Ca), number of external windows and number of hours with internal doors open. Results suggest that short-term indoor endotoxin variability may be driven by household dust/garbage production and handling. This would explain the modest predictive performance of published models that use answers to household surveys as predictors. One feasible alternative is to increase the sampling period so that household features would arise as significant predictors of long-term airborne endotoxin levels. Copyright © 2016 Elsevier Ltd. All rights reserved.
Gallastegi, Mara; Huss, Anke; Santa-Marina, Loreto; Aurrekoetxea, Juan J; Guxens, Mònica; Birks, Laura Ellen; Ibarluzea, Jesús; Guerra, David; Röösli, Martin; Jiménez-Zabala, Ana
2018-05-24
Radiofrequency (RF) fields are widely used and, while it is still unknown whether children are more vulnerable to this type of exposure, it is essential to explore their level of exposure in order to conduct adequate epidemiological studies. Personal measurements provide individualized information, but they are costly in terms of time and resources, especially in large epidemiological studies. Other approaches, such as estimation of time-weighted averages (TWAs) based on spot measurements could simplify the work. The aims of this study were to assess RF exposure in the Spanish INMA birth cohort by spot measurements and by personal measurements in the settings where children tend to spend most of their time, i.e., homes, schools and parks; to identify the settings and sources that contribute most to that exposure; and to explore if exposure assessment based on spot measurements is a valid proxy for personal exposure. When children were 8 years old, spot measurements were conducted in the principal settings of 104 participants: homes (104), schools and their playgrounds (26) and parks (79). At the same time, personal measurements were taken for a subsample of 50 children during 3 days. Exposure assessment based on personal and on spot measurements were compared both in terms of mean exposures and in exposure-dependent categories by means of Bland-Altman plots, Cohen's kappa and McNemar test. Median exposure levels ranged from 29.73 (in children's bedrooms) to 200.10 μW/m 2 (in school playgrounds) for spot measurements and were higher outdoors than indoors. Median personal exposure was 52.13 μW/m 2 and median levels of assessments based on spot measurements ranged from 25.46 to 123.21 μW/m 2 . Based on spot measurements, the sources that contributed most to the exposure were FM radio, mobile phone downlink and Digital Video Broadcasting-Terrestrial, while indoor and personal sources contributed very little (altogether <20%). Similar distribution was observed with personal measurements. There was a bias proportional to power density between personal measurements and estimates based on spot measurements, with the latter providing higher exposure estimates. Nevertheless, there were no systematic differences between those methodologies when classifying subjects into exposure categories. Personal measurements of total RF exposure showed low to moderate agreement with home and bedroom spot measurements and agreed better, though moderately, with TWA based on spot measurements in the main settings where children spend time (homes, schools and parks; Kappa = 0.46). Exposure assessment based on spot measurements could be a feasible proxy to rank personal RF exposure in children population, providing that all relevant locations are being measured. Copyright © 2018. Published by Elsevier Ltd.
NASA Astrophysics Data System (ADS)
Bonhoff, H. A.; Petersson, B. A. T.
2010-08-01
For the characterization of structure-borne sound sources with multi-point or continuous interfaces, substantial simplifications and physical insight can be obtained by incorporating the concept of interface mobilities. The applicability of interface mobilities, however, relies upon the admissibility of neglecting the so-called cross-order terms. Hence, the objective of the present paper is to clarify the importance and significance of cross-order terms for the characterization of vibrational sources. From previous studies, four conditions have been identified for which the cross-order terms can become more influential. Such are non-circular interface geometries, structures with distinctively differing transfer paths as well as a suppression of the zero-order motion and cases where the contact forces are either in phase or out of phase. In a theoretical study, the former four conditions are investigated regarding the frequency range and magnitude of a possible strengthening of the cross-order terms. For an experimental analysis, two source-receiver installations are selected, suitably designed to obtain strong cross-order terms. The transmitted power and the source descriptors are predicted by the approximations of the interface mobility approach and compared with the complete calculations. Neglecting the cross-order terms can result in large misinterpretations at certain frequencies. On average, however, the cross-order terms are found to be insignificant and can be neglected with good approximation. The general applicability of interface mobilities for structure-borne sound source characterization and the description of the transmission process thereby is confirmed.
Isotopic composition and neutronics of the Okelobondo natural reactor
NASA Astrophysics Data System (ADS)
Palenik, Christopher Samuel
The Oklo-Okelobondo and Bangombe uranium deposits, in Gabon, Africa host Earth's only known natural nuclear fission reactors. These 2 billion year old reactors represent a unique opportunity to study used nuclear fuel over geologic periods of time. The reactors in these deposits have been studied as a means by which to constrain the source term of fission product concentrations produced during reactor operation. The source term depends on the neutronic parameters, which include reactor operation duration, neutron flux and the neutron energy spectrum. Reactor operation has been modeled using a point-source computer simulation (Oak Ridge Isotope Generation and Depletion, ORIGEN, code) for a light water reactor. Model results have been constrained using secondary ionization mass spectroscopy (SIMS) isotopic measurements of the fission products Nd and Te, as well as U in uraninite from samples collected in the Okelobondo reactor zone. Based upon the constraints on the operating conditions, the pre-reactor concentrations of Nd (150 ppm +/- 75 ppm) and Te (<1 ppm) in uraninite were estimated. Related to the burnup measured in Okelobondo samples (0.7 to 13.8 GWd/MTU), the final fission product inventories of Nd (90 to 1200 ppm) and Te (10 to 110 ppm) were calculated. By the same means, the ranges of all other fission products and actinides produced during reactor operation were calculated as a function of burnup. These results provide a source term against which the present elemental and decay abundances at the fission reactor can be compared. Furthermore, they provide new insights into the extent to which a "fossil" nuclear reactor can be characterized on the basis of its isotopic signatures. In addition, results from the study of two other natural systems related to the radionuclide and fission product transport are included. A detailed mineralogical characterization of the uranyl mineralogy at the Bangombe uranium deposit in Gabon, Africa was completed to improve geochemical models of the solubility-limiting phase. A study of the competing effects of radiation damage and annealing in a U-bearing crystal of zircon shows that low temperature annealing in actinide-bearing phases is significant in the annealing of radiation damage.
Correcting STIS CCD Point-Source Spectra for CTE Loss
NASA Technical Reports Server (NTRS)
Goudfrooij, Paul; Bohlin, Ralph C.; Maiz-Apellaniz, Jesus
2006-01-01
We review the on-orbit spectroscopic observations that are being used to characterize the Charge Transfer Efficiency (CTE) of the STIS CCD in spectroscopic mode. We parameterize the CTE-related loss for spectrophotometry of point sources in terms of dependencies on the brightness of the source, the background level, the signal in the PSF outside the standard extraction box, and the time of observation. Primary constraints on our correction algorithm are provided by measurements of the CTE loss rates for simulated spectra (images of a tungsten lamp taken through slits oriented along the dispersion axis) combined with estimates of CTE losses for actual spectra of spectrophotometric standard stars in the first order CCD modes. For point-source spectra at the standard reference position at the CCD center, CTE losses as large as 30% are corrected to within approx.1% RMS after application of the algorithm presented here, rendering the Poisson noise associated with the source detection itself to be the dominant contributor to the total flux calibration uncertainty.
Aerosol Source Attributions and Source-Receptor Relationships Across the Northern Hemisphere
NASA Technical Reports Server (NTRS)
Bian, Huisheng; Chin, Mian; Kucsera, Tom; Pan, Xiaohua; Darmenov, Anton; Colarco, Peter; Torres, Omar; Shults, Michael
2014-01-01
Emissions and long-range transport of air pollution pose major concerns on air quality and climate change. To better assess the impact of intercontinental transport of air pollution on regional and global air quality, ecosystems, and near-term climate change, the UN Task Force on Hemispheric Transport of Air Pollution (HTAP) is organizing a phase II activity (HTAP2) that includes global and regional model experiments and data analysis, focusing on ozone and aerosols. This study presents the initial results of HTAP2 global aerosol modeling experiments. We will (a) evaluate the model results with surface and aircraft measurements, (b) examine the relative contributions of regional emission and extra-regional source on surface PM concentrations and column aerosol optical depth (AOD) over several NH pollution and dust source regions and the Arctic, and (c) quantify the source-receptor relationships in the pollution regions that reflect the sensitivity of regional aerosol amount to the regional and extra-regional emission reductions.
Room temperature single photon source using fiber-integrated hexagonal boron nitride
NASA Astrophysics Data System (ADS)
Vogl, Tobias; Lu, Yuerui; Lam, Ping Koy
2017-07-01
Single photons are a key resource for quantum optics and optical quantum information processing. The integration of scalable room temperature quantum emitters into photonic circuits remains to be a technical challenge. Here we utilize a defect center in hexagonal boron nitride (hBN) attached by Van der Waals force onto a multimode fiber as a single photon source. We perform an optical characterization of the source in terms of spectrum, state lifetime, power saturation and photostability. A special feature of our source is that it allows for easy switching between fiber-coupled and free space single photon generation modes. In order to prove the quantum nature of the emission we measure the second-order correlation function {{g}(2)}≤ft(τ \\right) . For both fiber-coupled and free space emission, the {{g}(2)}≤ft(τ \\right) dips below 0.5 indicating operation in the single photon regime. The results so far demonstrate the feasibility of 2D material single photon sources for scalable photonic quantum information processing.
Casas, Francisco J.; Ortiz, David; Villa, Enrique; Cano, Juan L.; Cagigas, Jaime; Pérez, Ana R.; Aja, Beatriz; Terán, J. Vicente; de la Fuente, Luisa; Artal, Eduardo; Hoyland, Roger; Génova-Santos, Ricardo
2015-01-01
This paper presents preliminary polarization measurements and systematic-error characterization of the Thirty Gigahertz Instrument receiver developed for the QUIJOTE experiment. The instrument has been designed to measure the polarization of Cosmic Microwave Background radiation from the sky, obtaining the Q, U, and I Stokes parameters of the incoming signal simultaneously. Two kinds of linearly polarized input signals have been used as excitations in the polarimeter measurement tests in the laboratory; these show consistent results in terms of the Stokes parameters obtained. A measurement-based systematic-error characterization technique has been used in order to determine the possible sources of instrumental errors and to assist in the polarimeter calibration process. PMID:26251906
Scene-based Shack-Hartmann wavefront sensor for light-sheet microscopy
NASA Astrophysics Data System (ADS)
Lawrence, Keelan; Liu, Yang; Dale, Savannah; Ball, Rebecca; VanLeuven, Ariel J.; Sornborger, Andrew; Lauderdale, James D.; Kner, Peter
2018-02-01
Light-sheet microscopy is an ideal imaging modality for long-term live imaging in model organisms. However, significant optical aberrations can be present when imaging into an organism that is hundreds of microns or greater in size. To measure and correct optical aberrations, an adaptive optics system must be incorporated into the microscope. Many biological samples lack point sources that can be used as guide stars with conventional Shack-Hartmann wavefront sensors. We have developed a scene-based Shack-Hartmann wavefront sensor for measuring the optical aberrations in a light-sheet microscopy system that does not require a point-source and can measure the aberrations for different parts of the image. The sensor has 280 lenslets inside the pupil, creates an image from each lenslet with a 500 micron field of view and a resolution of 8 microns, and has a resolution for the wavefront gradient of 75 milliradians per lenslet. We demonstrate the system on both fluorescent bead samples and zebrafish embryos.
Advanced turbo-prop airplane interior noise reduction-source definition
NASA Technical Reports Server (NTRS)
Magliozzi, B.; Brooks, B. M.
1979-01-01
Acoustic pressure amplitudes and phases were measured in model scale on the surface of a rigid semicylinder mounted in an acoustically treated wind tunnel near a prop-fan (an advanced turboprop with many swept blades) model. Operating conditions during the test simulated those of a prop-fan at 0.8 Mach number cruise. Acoustic pressure amplitude and phase contours were defined on the semicylinder surface. Measurements obtained without the semi-cylinder in place were used to establish the magnitude of pressure doubling for an aircraft fuselage located near a prop-fan. Pressure doubling effects were found to be 6dB at 90 deg incidence decreasing to no effect at grazing incidence. Comparisons of measurements with predictions made using a recently developed prop-fan noise prediction theory which includes linear and non-linear source terms showed good agreement in phase and in peak noise amplitude. Predictions of noise amplitude and phase contours, including pressure doubling effects derived from test, are included for a full scale prop-fan installation.
Interpretation of fast-ion signals during beam modulation experiments
Heidbrink, W. W.; Collins, C. S.; Stagner, L.; ...
2016-07-22
Fast-ion signals produced by a modulated neutral beam are used to infer fast-ion transport. The measured quantity is the divergence of perturbed fast-ion flux from the phase-space volume measured by the diagnostic, ∇•more » $$\\bar{Γ}$$. Since velocity-space transport often contributes to this divergence, the phase-space sensitivity of the diagnostic (or “weight function”) plays a crucial role in the interpretation of the signal. The source and sink make major contributions to the signal but their effects are accurately modeled by calculations that employ an exponential decay term for the sink. Recommendations for optimal design of a fast-ion transport experiment are given, illustrated by results from DIII-D measurements of fast-ion transport by Alfv´en eigenmodes. Finally, the signal-to-noise ratio of the diagnostic, systematic uncertainties in the modeling of the source and sink, and the non-linearity of the perturbation all contribute to the error in ∇•$$\\bar{Γ}$$.« less
Boron selenide semiconductor detectors for thermal neutron counting
NASA Astrophysics Data System (ADS)
Kargar, Alireza; Tower, Joshua; Cirignano, Leonard; Shah, Kanai
2013-09-01
Thermal neutron detectors in planar configuration were fabricated from B2Se3 (Boron Selenide) crystals grown at RMD Inc. All fabricated semiconductor devices were characterized for the current-voltage (I-V) characteristic and neutron counting measurement. In this study, the resistivity of crystals is reported and the collected pulse height spectra are presented for devices irradiated with the 241AmBe neutron source. Long-term stability of the B2Se3 devices for neutron detection under continuous bias and without being under continuous bias was investigated and the results are reported. The B2Se3 devices showed response to thermal neutrons of the 241AmBe source.
On the sound field radiated by a tuning fork
NASA Astrophysics Data System (ADS)
Russell, Daniel A.
2000-12-01
When a sounding tuning fork is brought close to the ear, and rotated about its long axis, four distinct maxima and minima are heard. However, when the same tuning fork is rotated while being held at arm's length from the ear only two maxima and minima are heard. Misconceptions concerning this phenomenon are addressed and the fundamental mode of the fork is described in terms of a linear quadrupole source. Measured directivity patterns in the near field and far field of several forks agree very well with theoretical predictions for a linear quadrupole. Other modes of vibration are shown to radiate as dipole and lateral quadrupole sources.
Elemental composition and size distribution of particulates in Cleveland, Ohio
NASA Technical Reports Server (NTRS)
King, R. B.; Fordyce, J. S.; Neustadter, H. E.; Leibecki, H. F.
1975-01-01
Measurements were made of the elemental particle size distribution at five contrasting urban environments with different source-type distributions in Cleveland, Ohio. Air quality conditions ranged from normal to air pollution alert levels. A parallel network of high-volume cascade impactors (5-state) were used for simultaneous sampling on glass fiber surfaces for mass determinations and on Whatman-41 surfaces for elemental analysis by neutron activation for 25 elements. The elemental data are assessed in terms of distribution functions and interrelationships and are compared between locations as a function of resultant wind direction in an attempt to relate the findings to sources.
Elemental composition and size distribution of particulates in Cleveland, Ohio
NASA Technical Reports Server (NTRS)
Leibecki, H. F.; King, R. B.; Fordyce, J. S.; Neustadter, H. E.
1975-01-01
Measurements have been made of the elemental particle size distribution at five contrasting urban environments with different source-type distributions in Cleveland, Ohio. Air quality conditions ranged from normal to air pollution alert levels. A parallel network of high-volume cascade impactors (5-stage) were used for simultaneous sampling on glass fiber surfaces for mass determinations and on Whatman-41 surfaces for elemental analysis by neutron activation for 25 elements. The elemental data are assessed in terms of distribution functions and interrelationships and are compared between locations as a function of resultant wind direction in an attempt to relate the findings to sources.
NASA Technical Reports Server (NTRS)
Horsham, Gary A. P.
1999-01-01
This paper presents a comprehensive review and assessment of a demonstration technology transfer and commercialization prouram called "CommTech". The pro-ram was conceived and initiated in early to mid-fiscal year 1995, and extended roughly three years into the future. Market research sources were used to initially gather primary technological problems and needs data from non-aerospace companies in three targeted industry sectors: environmental, surface transportation, and bioengineering. Company-supplied information served as input data to activate or start-up an internal, phased matchmaking process. This process was based on technical-level relationship exploration followed by business-level agreement negotiations. and culminated with project management and execution. Space Act Agreements represented near-term outputs. Company product or process commercialization derived from NASA Glenn support and measurable economic effects represented far-term outputs.
NASA Astrophysics Data System (ADS)
Piedrahita, Ricardo A.
The Denver Aerosol Sources and Health study (DASH) was a long-term study of the relationship between the variability in fine particulate mass and chemical constituents (PM2.5, particulate matter less than 2.5mum) and adverse health effects such as cardio-respiratory illnesses and mortality. Daily filter samples were chemically analyzed for multiple species. We present findings based on 2.8 years of DASH data, from 2003 to 2005. Multilinear Engine 2 (ME-2), a receptor-based source apportionment model was applied to the data to estimate source contributions to PM2.5 mass concentrations. This study relied on two different ME-2 models: (1) a 2-way model that closely reflects PMF-2; and (2) an enhanced model with meteorological data that used additional temporal and meteorological factors. The Coarse Rural Urban Sources and Health study (CRUSH) is a long-term study of the relationship between the variability in coarse particulate mass (PMcoarse, particulate matter between 2.5 and 10mum) and adverse health effects such as cardio-respiratory illnesses, pre-term births, and mortality. Hourly mass concentrations of PMcoarse and fine particulate matter (PM2.5) are measured using tapered element oscillating microbalances (TEOMs) with Filter Dynamics Measurement Systems (FDMS), at two rural and two urban sites. We present findings based on nine months of mass concentration data, including temporal trends, and non-parametric regressions (NPR) results, which were used to characterize the wind speed and wind direction relationships that might point to sources. As part of CRUSH, 1-year coarse and fine mode particulate matter filter sampling network, will allow us to characterize the chemical composition of the particulate matter collected and perform spatial comparisons. This work describes the construction and validation testing of four dichotomous filter samplers for this purpose. The use of dichotomous splitters with an approximate 2.5mum cut point, coupled with a 10mum cut diameter inlet head allows us to collect the separated size fractions that the collocated TEOMs collect continuously. Chemical analysis of the filters will include inorganic ions, organic compounds, EC, OC, and biological analyses. Side by side testing showed the cut diameters were in agreement with each other, and with a well characterized virtual impactor lent to the group by the University of Southern California. Error propagation was performed and uncertainty results were similar to the observed standard deviations.
NASA Technical Reports Server (NTRS)
Ohring, G.; Wielicki, B.; Spencer, R.; Emery, B.; Datla, R.
2004-01-01
Measuring the small changes associated with long-term global climate change from space is a daunting task. To address these problems and recommend directions for improvements in satellite instrument calibration some 75 scientists, including researchers who develop and analyze long-term data sets from satellites, experts in the field of satellite instrument calibration, and physicists working on state of the art calibration sources and standards met November 12 - 14, 2002 and discussed the issues. The workshop defined the absolute accuracies and long-term stabilities of global climate data sets that are needed to detect expected trends, translated these data set accuracies and stabilities to required satellite instrument accuracies and stabilities, and evaluated the ability of current observing systems to meet these requirements. The workshop's recommendations include a set of basic axioms or overarching principles that must guide high quality climate observations in general, and a roadmap for improving satellite instrument characterization, calibration, inter-calibration, and associated activities to meet the challenge of measuring global climate change. It is also recommended that a follow-up workshop be conducted to discuss implementation of the roadmap developed at this workshop.
David, Pierre-Marie
2017-04-01
Fragile states have been raising increasing concern among donors since the mid-2000s. The policies of the Global Fund to fight HIV/AIDS, Malaria, and Tuberculosis (GF) have not excluded fragile states, and this source has provided financing for these countries according to standardized procedures. They represent interesting cases for exploring the meaning and role of measurement in a globalized context. Measurement in the field of HIV/AIDS and its treatment has given rise to a private outsourcing of expertise and auditing, thereby creating a new form of value based on the social process of registration and the creation of realities produced by the intervention itself. These "scriptural economies" must be questioned in terms of the production of knowledge, but also in terms of social justice. Governing HIV/AIDS treatments by numbers in a fragile state is explored in this article through the experience of the Central African Republic (CAR) in terms of epidemiology and access to antiretroviral drugs. The unexpected effects of performance-based programs in this context underline the need for global health governance to be re-embedded into a social justice framework. © 2016 John Wiley & Sons Ltd.
Gardner, Benjamin
2015-01-01
The term 'habit' is widely used to predict and explain behaviour. This paper examines use of the term in the context of health-related behaviour, and explores how the concept might be made more useful. A narrative review is presented, drawing on a scoping review of 136 empirical studies and 8 literature reviews undertaken to document usage of the term 'habit', and methods to measure it. A coherent definition of 'habit', and proposals for improved methods for studying it, were derived from findings. Definitions of 'habit' have varied in ways that are often implicit and not coherently linked with an underlying theory. A definition is proposed whereby habit is a process by which a stimulus generates an impulse to act as a result of a learned stimulus-response association. Habit-generated impulses may compete or combine with impulses and inhibitions arising from other sources, including conscious decision-making, to influence responses, and need not generate behaviour. Most research on habit is based on correlational studies using self-report measures. Adopting a coherent definition of 'habit', and a wider range of paradigms, designs and measures to study it, may accelerate progress in habit theory and application.
Isotopic Recorders of Pollution in Heterogeneous Urban Areas
NASA Astrophysics Data System (ADS)
Pataki, D. E.; Cobley, L.; Smith, R. M.; Ehleringer, J. R.; Chritz, K.
2017-12-01
A significant difficulty in quantifying urban pollution lies in the extreme spatial and temporal heterogeneity of cities. Dense sources of both point and non-point source pollution as well as the dynamic role of human activities, which vary over very short time scales and small spatial scales, complicate efforts to establish long-term urban monitoring networks that are relevant at neighborhood, municipal, and regional scales. Fortunately, the natural abundance of isotopes of carbon, nitrogen, and other elements provides a wealth of information about the sources and fate of urban atmospheric pollution. In particular, soils and plant material integrate pollution sources and cycling over space and time, and have the potential to provide long-term records of pollution dynamics that extend back before atmospheric monitoring data are available. Similarly, sampling organic material at high spatial resolution can provide "isoscapes" that shed light on the spatial heterogeneity of pollutants in different urban parcels and neighborhoods, along roads of varying traffic density, and across neighborhoods of varying affluence and sociodemographic composition. We have compiled numerous datasets of the isotopic composition of urban organic matter that illustrate the potential for isotopic monitoring of urban areas as a means of understanding hot spots and hot moments in urban atmospheric biogeochemistry. Findings to date already reveal the critical role of affluence, economic activity, demographic change, and land management practices in influencing urban pollution sources and sinks, and suggest an important role of stable isotope and radioisotope measurements in urban atmospheric and biogeochemical monitoring.
A Semi-implicit Treatment of Porous Media in Steady-State CFD.
Domaingo, Andreas; Langmayr, Daniel; Somogyi, Bence; Almbauer, Raimund
There are many situations in computational fluid dynamics which require the definition of source terms in the Navier-Stokes equations. These source terms not only allow to model the physics of interest but also have a strong impact on the reliability, stability, and convergence of the numerics involved. Therefore, sophisticated numerical approaches exist for the description of such source terms. In this paper, we focus on the source terms present in the Navier-Stokes or Euler equations due to porous media-in particular the Darcy-Forchheimer equation. We introduce a method for the numerical treatment of the source term which is independent of the spatial discretization and based on linearization. In this description, the source term is treated in a fully implicit way whereas the other flow variables can be computed in an implicit or explicit manner. This leads to a more robust description in comparison with a fully explicit approach. The method is well suited to be combined with coarse-grid-CFD on Cartesian grids, which makes it especially favorable for accelerated solution of coupled 1D-3D problems. To demonstrate the applicability and robustness of the proposed method, a proof-of-concept example in 1D, as well as more complex examples in 2D and 3D, is presented.
Seeley, T D; Mikheyev, A S; Pagano, G J
2000-09-01
For more than 50 years, investigators of the honey bee's waggle dance have reported that richer food sources seem to elicit longer-lasting and livelier dances than do poorer sources. However, no one had measured both dance duration and liveliness as a function of food-source profitability. Using video analysis, we found that nectar foragers adjust both the duration (D) and the rate (R) of waggle-run production, thereby tuning the number of waggle runs produced per foraging trip (W, where W= DR) as a function of food-source profitability. Both duration and rate of waggle-run production increase with rising food-source profitability. Moreover, we found that a dancing bee adjusts the rate of waggle-run production (R) in relation to food-source profitability by adjusting the mean duration of the return-phase portion of her dance circuits. This finding raises the possibility that bees can use return-phase duration as an index of food-source profitability. Finally, dances having different levels of liveliness have different mean durations of the return phase, indicating that dance liveliness can be quantified in terms of the time interval between consecutive waggle runs.
NASA Astrophysics Data System (ADS)
Kadem, L.; Knapp, Y.; Pibarot, P.; Bertrand, E.; Garcia, D.; Durand, L. G.; Rieu, R.
2005-12-01
The effective orifice area (EOA) is the most commonly used parameter to assess the severity of aortic valve stenosis as well as the performance of valve substitutes. Particle image velocimetry (PIV) may be used for in vitro estimation of valve EOA. In the present study, we propose a new and simple method based on Howe’s developments of Lighthill’s aero-acoustic theory. This method is based on an acoustical source term (AST) to estimate the EOA from the transvalvular flow velocity measurements obtained by PIV. The EOAs measured by the AST method downstream of three sharp-edged orifices were in excellent agreement with the EOAs predicted from the potential flow theory used as the reference method in this study. Moreover, the AST method was more accurate than other conventional PIV methods based on streamlines, inflexion point or vorticity to predict the theoretical EOAs. The superiority of the AST method is likely due to the nonlinear form of the AST. There was also an excellent agreement between the EOAs measured by the AST method downstream of the three sharp-edged orifices as well as downstream of a bioprosthetic valve with those obtained by the conventional clinical method based on Doppler-echocardiographic measurements of transvalvular velocity. The results of this study suggest that this new simple PIV method provides an accurate estimation of the aortic valve flow EOA. This new method may thus be used as a reference method to estimate the EOA in experimental investigation of the performance of valve substitutes and to validate Doppler-echocardiographic measurements under various physiologic and pathologic flow conditions.
NASA Astrophysics Data System (ADS)
Zavala, Miguel; Molina, Luisa T.; Fortner, Edward; Knighton, Berk; Herndon, Scott; Yacovitch, Tara; Floerchinger, Cody; Roscioli, Joseph; Kolb, Charles; Mejia, Jose Antonio; Sarmiento, Jorge; Paramo, Victor Hugo; Zirath, Sergio; Jazcilevich, Aron
2014-05-01
Black carbon emitted from freight, public transport, and heavy duty trucks sources is linked with adverse effects on human health. In addition, the control of emissions of black carbon, an important short-lived climate forcing agent (SLCF), has recently been considered as one of the key strategies for mitigating regional near-term climate change. Despite the availability of new emissions control technologies for reducing emissions from diesel-powered mobile sources, their introduction is still not widespread in many urban areas and there is a need to characterize real-world emission rates of black carbon from this key source. The emissions of black carbon, organic carbon, and other gaseous and particle pollutants from diesel-powered mobile sources in Mexico were characterized by deploying a mobile laboratory equipped with real-time instrumentation in Mexico City as part of the SLCFs-Mexico 2013 project. From February 25-28 of 2013 the emissions from selected diesel-powered vehicles were measured in both controlled experiments and real-world on-road driving conditions. Sampled vehicles had several emissions levels technologies, including: EPA98, EPA03, EPA04, EURO3-5, and Hybrid. All vehicles were sampled using diesel fuel and several vehicles were measured using both diesel and biodiesel fuels. Additional measurements included the use of a remote sensing unit for the co-sampling of all tested vehicles, and the installation and operation of a Portable Emissions Measurements System (PEMS) for the measurement of emissions from a test vehicle. We will present inter-comparisons of the emission factors obtained among the various vehicle technologies that were sampled during the experiment as well as the inter-comparison of results from the various sampling platforms. The results can be used to
NASA Technical Reports Server (NTRS)
Panda, Jayanta; Seasholtz, Richard G.; Elam, Kristie A.
2002-01-01
To locate noise sources in high-speed jets, the sound pressure fluctuations p', measured at far field locations, were correlated with each of radial velocity v, density rho, and phov(exp 2) fluctuations measured from various points in jet plumes. The experiments follow the cause-and-effect method of sound source identification, where
Jovian Northern Ethane Aurora and the Solar Cycle
NASA Technical Reports Server (NTRS)
Kostiuk,T.; Livengood, T.; Fast, K.; Buhl, D.; Goldstein, J.; Hewagama, T.
1999-01-01
Thermal infrared auroral spectra from Jupiter's North polar region have been collected from 1979 to 1998 in a continuing study of long-term variability in the northern thermal IR aurora, using C2H6 emission lines near 12 microns as a probe. Data from Voyager I and 2 IRIS measurements and ground based spectral measurements were analyzed using the same model atmosphere to provide a consistent relative comparison. A retrieved equivalent mole fraction was used to compare the observed integrated emission. Short term (days), medium term (months) and long term (years) variability in the ethane emission was observed. The variability Of C2H6 emission intensities was compared to Jupiter's seasonal cycle and the solar activity cycle. A positive correlation appears to exist, with significantly greater emission and short term variability during solar maxima. Observations on 60 N latitude during increased solar activity in 1979, 1989, and most recently in 1998 show up to 5 times brighter integrated line emission of C2H6 near the north polar "hot spot" (150-210 latitude) than from the north quiescent region. Significantly lower enhancement was observed during periods of lower solar activity in 1982, 1983, 1993, and 1995. Possible sources and mechanisms for the enhancement and variability will be discussed.
Dancing Bees Improve Colony Foraging Success as Long-Term Benefits Outweigh Short-Term Costs
Schürch, Roger; Grüter, Christoph
2014-01-01
Waggle dancing bees provide nestmates with spatial information about high quality resources. Surprisingly, attempts to quantify the benefits of this encoded spatial information have failed to find positive effects on colony foraging success under many ecological circumstances. Experimental designs have often involved measuring the foraging success of colonies that were repeatedly switched between oriented dances versus disoriented dances (i.e. communicating vectors versus not communicating vectors). However, if recruited bees continue to visit profitable food sources for more than one day, this procedure would lead to confounded results because of the long-term effects of successful recruitment events. Using agent-based simulations, we found that spatial information was beneficial in almost all ecological situations. Contrary to common belief, the benefits of recruitment increased with environmental stability because benefits can accumulate over time to outweigh the short-term costs of recruitment. Furthermore, we found that in simulations mimicking previous experiments, the benefits of communication were considerably underestimated (at low food density) or not detected at all (at medium and high densities). Our results suggest that the benefits of waggle dance communication are currently underestimated and that different experimental designs, which account for potential long-term benefits, are needed to measure empirically how spatial information affects colony foraging success. PMID:25141306
Dancing bees improve colony foraging success as long-term benefits outweigh short-term costs.
Schürch, Roger; Grüter, Christoph
2014-01-01
Waggle dancing bees provide nestmates with spatial information about high quality resources. Surprisingly, attempts to quantify the benefits of this encoded spatial information have failed to find positive effects on colony foraging success under many ecological circumstances. Experimental designs have often involved measuring the foraging success of colonies that were repeatedly switched between oriented dances versus disoriented dances (i.e. communicating vectors versus not communicating vectors). However, if recruited bees continue to visit profitable food sources for more than one day, this procedure would lead to confounded results because of the long-term effects of successful recruitment events. Using agent-based simulations, we found that spatial information was beneficial in almost all ecological situations. Contrary to common belief, the benefits of recruitment increased with environmental stability because benefits can accumulate over time to outweigh the short-term costs of recruitment. Furthermore, we found that in simulations mimicking previous experiments, the benefits of communication were considerably underestimated (at low food density) or not detected at all (at medium and high densities). Our results suggest that the benefits of waggle dance communication are currently underestimated and that different experimental designs, which account for potential long-term benefits, are needed to measure empirically how spatial information affects colony foraging success.
Assessment of Noise and Associated Health Impacts at Selected Secondary Schools in Ibadan, Nigeria
Ana, Godson R. E. E.; Shendell, Derek G.; Brown, G. E.; Sridhar, M. K. C.
2009-01-01
Background. Most schools in Ibadan, Nigeria, are located near major roads (mobile line sources). We conducted an initial assessment of noise levels and adverse noise-related health and learning effects. Methods. For this descriptive, cross-sectional study, four schools were selected randomly from eight participating in overall project. We administered 200 questionnaires, 50 per school, assessing health and learning-related outcomes. Noise levels (A-weighted decibels, dBA) were measured with calibrated sound level meters. Traffic density was assessed for school with the highest measured dBA. Observational checklists assessed noise control parameters and building physical attributes. Results. Short-term, cross-sectional school-day noise levels ranged 68.3–84.7 dBA. Over 60% of respondents reported that vehicular traffic was major source of noise, and over 70% complained being disturbed by noise. Three schools reported tiredness, and one school lack of concentration, as the most prevalent noise-related health problems. Conclusion. Secondary school occupants in Ibadan, Nigeria were potentially affected by exposure to noise from mobile line sources. PMID:20041025
Probing Atom-Surface Interactions by Diffraction of Bose-Einstein Condensates
NASA Astrophysics Data System (ADS)
Bender, Helmar; Stehle, Christian; Zimmermann, Claus; Slama, Sebastian; Fiedler, Johannes; Scheel, Stefan; Buhmann, Stefan Yoshi; Marachevsky, Valery N.
2014-01-01
In this article, we analyze the Casimir-Polder interaction of atoms with a solid grating and the repulsive interaction between the atoms and the grating in the presence of an external laser source. The Casimir-Polder potential is evaluated exactly in terms of Rayleigh reflection coefficients and via an approximate Hamaker approach. The laser-tuned repulsive interaction is given in terms of Rayleigh transmission coefficients. The combined potential landscape above the solid grating is probed locally by diffraction of Bose-Einstein condensates. Measured diffraction efficiencies reveal information about the shape of the potential landscape in agreement with the theory based on Rayleigh decompositions.
Long Term 2 Second Round Source Water Monitoring and Bin Placement Memo
The Long Term 2 Enhanced Surface Water Treatment Rule (LT2ESWTR) applies to all public water systems served by a surface water source or public water systems served by a ground water source under the direct influence of surface water.
Shifting material source of Chinese Loess since ~2.7 Ma reflected by Sr isotopic composition.
Zhang, Wenfang; Chen, Jun; Li, Gaojun
2015-05-21
Deciphering the sources of eolian dust on the Chinese Loess Plateau (CLP) is fundamental to reconstruct paleo-wind patterns and paleo-environmental changes. Existing datasets show contradictory source evolutions of eolian dust on the CLP, both on orbital and tectonic timescales. Here, the silicate Sr and Nd isotopic compositions of a restricted grain size fraction (28-45 μm) were measured to trace the source evolution of the CLP since ~2.7 Ma. Our results revealed an unchanged source on orbital timescales but a gradual source shift from the Qilian Mountains to the Gobi Altay Mountains during the past 2.7 Ma. Both tectonic uplift and climate change may have played important roles for this shift. The later uplift of the Gobi Altay Mountains relative to the Qilian Mountains since 5 ± 3 Ma might be responsible for the increasing contribution of Gobi materials to the source deserts in Alxa arid lands. Enhanced winter monsoon may also facilitate transportation of Gobi materials from the Alxa arid lands to the CLP. The shifting source of Asian dust was also reflected in north Pacific sediments. The finding of this shifting source calls for caution when interpreting the long-term climate changes based on the source-sensitive proxies of the eolian deposits.
Uncertainty, variability, and earthquake physics in ground‐motion prediction equations
Baltay, Annemarie S.; Hanks, Thomas C.; Abrahamson, Norm A.
2017-01-01
Residuals between ground‐motion data and ground‐motion prediction equations (GMPEs) can be decomposed into terms representing earthquake source, path, and site effects. These terms can be cast in terms of repeatable (epistemic) residuals and the random (aleatory) components. Identifying the repeatable residuals leads to a GMPE with reduced uncertainty for a specific source, site, or path location, which in turn can yield a lower hazard level at small probabilities of exceedance. We illustrate a schematic framework for this residual partitioning with a dataset from the ANZA network, which straddles the central San Jacinto fault in southern California. The dataset consists of more than 3200 1.15≤M≤3 earthquakes and their peak ground accelerations (PGAs), recorded at close distances (R≤20 km). We construct a small‐magnitude GMPE for these PGA data, incorporating VS30 site conditions and geometrical spreading. Identification and removal of the repeatable source, path, and site terms yield an overall reduction in the standard deviation from 0.97 (in ln units) to 0.44, for a nonergodic assumption, that is, for a single‐source location, single site, and single path. We give examples of relationships between independent seismological observables and the repeatable terms. We find a correlation between location‐based source terms and stress drops in the San Jacinto fault zone region; an explanation of the site term as a function of kappa, the near‐site attenuation parameter; and a suggestion that the path component can be related directly to elastic structure. These correlations allow the repeatable source location, site, and path terms to be determined a priori using independent geophysical relationships. Those terms could be incorporated into location‐specific GMPEs for more accurate and precise ground‐motion prediction.
NASA Astrophysics Data System (ADS)
Loubet, Benjamin; Gonzaga, Lais; Buysse, Pauline; Ciuraru, Raluca; Lafouge, Florence; Decuq, Céline; Zurfluh, Olivier; Fortineau, Alain; Fanucci, Olivier; Sarda-Esteve, Roland; Zannoni, Nora; Truong, Francois; Boissard, Christophe; Gros, Valérie
2017-04-01
Volatile organic compounds (VOC) are essential drivers of atmospheric chemistry. Many VOCs are emitted from and deposited to ecosystems. While forests and grasslands have already been substantially studied, exchanges of VOCs with crops are less known, although these ecosystems represent more than 50% of the surface in France. In this study, we analyze sources and sinks of VOCs in a wheat field (at the ICOS FR-GRI site near Paris) at anthesis based on measurements of fluxes, concentration profiles and branch chambers. The VOCs were measured using a PTR-TOF-Qi-MS (where Qi stands for Quad Ion guide). Air was successively sampled through lines located at different heights within and above the canopy, of which one was used for Eddy Covariance and located near a sonic anemometer. Additional measurements included the standard ICOS meteorological data as well as leaf area index profiles and photosynthesis curves at several heights in the canopy. We report fluxes and profiles for more than 500 VOCs. The deposition velocities of depositing compounds are compared to the maximum exchange velocity and the ozone deposition velocity. The sources and sinks location and magnitude are evaluated by inverse Lagrangian modelling assuming no reaction and simple reaction schemes in the canopy. The sources and sinks of VOC in the canopy are interpreted in terms crop phenology and the potential for reaction with ozone and NOx is evaluated. This study takes place in the ADEME CORTEA COV3ER French project (http://www6.inra.fr/cov3er).
Matrix effect and recovery terminology issues in regulated drug bioanalysis.
Huang, Yong; Shi, Robert; Gee, Winnie; Bonderud, Richard
2012-02-01
Understanding the meaning of the terms used in the bioanalytical method validation guidance is essential for practitioners to implement best practice. However, terms that have several meanings or that have different interpretations exist within bioanalysis, and this may give rise to differing practices. In this perspective we discuss an important but often confusing term - 'matrix effect (ME)' - in regulated drug bioanalysis. The ME can be interpreted as either the ionization change or the measurement bias of the method caused by the nonanalyte matrix. The ME definition dilemma makes its evaluation challenging. The matrix factor is currently used as a standard method for evaluation of ionization changes caused by the matrix in MS-based methods. Standard additions to pre-extraction samples have been suggested to evaluate the overall effects of a matrix from different sources on the analytical system, because it covers ionization variation and extraction recovery variation. We also provide our personal views on the term 'recovery'.
On the gravitational potential and field anomalies due to thin mass layers
NASA Technical Reports Server (NTRS)
Ockendon, J. R.; Turcotte, D. L.
1977-01-01
The gravitational potential and field anomalies for thin mass layers are derived using the technique of matched asymptotic expansions. An inner solution is obtained using an expansion in powers of the thickness and it is shown that the outer solution is given by a surface distribution of mass sources and dipoles. Coefficients are evaluated by matching the inner expansion of the outer solution with the outer expansion of the inner solution. The leading term in the inner expansion for the normal gravitational field gives the Bouguer formula. The leading term in the expansion for the gravitational potential gives an expression for the perturbation to the geoid. The predictions given by this term are compared with measurements by satellite altimetry. The second-order terms in the expansion for the gravitational field are required to predict the gravity anomaly at a continental margin. The results are compared with observations.
NASA Astrophysics Data System (ADS)
Jones, Sue; Hoon, Stephen R.; Richardson, Nigel; Bennett, Michael
2016-04-01
The significant increase in global air travel which has occurred during the last fifty years has generated growing concern regarding the potential impacts associated with increasing emissions of particulate matter (PM) from aviation activity on health and the environment. PM within the airport environment, in particular, may be derived from a wide range of potential sources including aircraft; vehicles; ground support equipment and buildings. In order to investigate and remediate potential problem sources, it is important to be able to identify characteristic particulate 'fingerprints' which would allow source attribution, particularly respirable particulates. To date the identification of such 'fingerprints' has remained elusive but remains a key research priority for the aviation industry (Webb et al, 2008). In previous PM studies, environmental magnetism has been used as a successful technique for discriminating between different emission types and particulate sources in both urban and industrial environments (e.g. Hunt et al 1984; Lecoanet et al 2003, Jones et al 2015). Environmental magnetism is a non-destructive and relatively rapid technique involving the use of non-directional, rock magnetic measurements to characterise the mineral magnetic properties of natural and anthropogenic materials. In other studies scanning electron microscopy (SEM) has also been used as an effective characterisation technique for the investigation of grain size and morphology of PM derived from vehicle emissions (e.g. Bucko et al 2010) and fossil fuel combustion sources (Kim et al 2009). In this study, environmental magnetic measurements and SEM/EDAX have been used to characterise dusts from specific aircraft sources including engines, brakes and tyres. Furthermore, these methods have also been applied to runway (both hard and grass covered surfaces), taxiway and apron dusts collected during extensive environmental sampling at Manchester International Airport, UK in order to investigate source attribution. The results indicate that the dusts collected from the various aircraft sources (i.e. engines, brakes and tyres) are significantly different in terms of magnetic mineral type and grain size. Furthermore, particulates deposited at different locations on the runway surface show significant differentiation in magnetic grain size and mineralogy which when compared with the results from the different aircraft sources suggest that they may relate to emissions from different sources at various stages of the take/off landing cycle. Results of SEM/EDAX analysis show that aircraft engine, brake and tyre dust particulates vary significantly in terms of morphology and chemical composition. All sources include respirable (sub 10 micron) particulates. Engine dusts are carbon and silicon rich dominated by angular particulates. They have a distinctive chemical composition including Chromium, Cobalt and Nickel. Tyre dusts are predominantly carbon based dominated by spherical particulates and a unique presence of Zinc. Brake dusts, carbon and oxygen dominated and trace metals, include sub-angular particulates but an absence of the characteristic engine and tyre dusts metals. By combining SEM/EDAX measurements and magnetic measurements we are establishing potential fingerprints for particulates from ground based air transport activities to enable identification of potential health hazards. This will help inform management plans for reduction of associated risks to the environment and health. References Bucko, M., Magiera, T., Pesonen, L., Janus, B. (2010) 'Magnetic, geochemical and microstructural characteristics of road dust on roadsides with different traffic volumes - Case study from Finland' Water, Air and Soil Pollution 209, pp. 295-306. Hunt, A., Jones, J. and Oldfield, F. (1984) 'Magnetic measurements and heavy metals in atmospheric particulates of anthropogenic origin' The Science of the Total Environment 33, 129-139. Jones, S., Richardson, N., Bennett, M. and Hoon, S.R. (2015) The application of magnetic measurements for the characterization of atmospheric particulate pollution within the airport environment. Science of the Total Environment., 502 pp.385-390 Kim, W., Doh, S., Yu, Y. (2009) 'Anthropogenic contribution of magnetic particulates in urban roadside dust' Atmospheric Environment 43 (19) pp.3137-3144. Lecoanet, H., Leveque, F. and Ambrosi, J.P. (2003) 'Combination of magnetic parameters: an efficient way to discriminate soil-contamination sources (south France)' Environmental Pollution 122, 229-234. Webb, S., Whitefield, P.D., Miake-Lye, R.C., Timko, M.T. and Thrasher, T.G. (2008) 'ACRP Report 6: Research needs associated with particulate emissions at airports'. Transportation Research Board'.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hensley, F; Chofor, N; Schoenfeld, A
2016-06-15
Purpose: In the steep dose gradients in the vicinity of a radiation source and due to the properties of the changing photon spectra, dose measurements in Brachytherapy usually have large uncertainties. Working group DIN 6803-3 is presently discussing recommendations for practical brachytherapy dosimetry incorporating recent theoretical developments in the description of brachytherapy radiation fields as well as new detectors and phantom materials. The goal is to prepare methods and instruments to verify dose calculation algorithms and for clinical dose verification with reduced uncertainties. Methods: After analysis of the distance dependent spectral changes of the radiation field surrounding brachytherapy sources, themore » energy dependent response of typical brachytherapy detectors was examined with Monte Carlo simulations. A dosimetric formalism was developed allowing the correction of their energy dependence as function of source distance for a Co-60 calibrated detector. Water equivalent phantom materials were examined with Monte Carlo calculations for their influence on brachytherapy photon spectra and for their water equivalence in terms of generating equivalent distributions of photon spectra and absorbed dose to water. Results: The energy dependence of a detector in the vicinity of a brachytherapy source can be described by defining an energy correction factor kQ for brachytherapy in the same manner as in existing dosimetry protocols which incorporates volume averaging and radiation field distortion by the detector. Solid phantom materials were identified which allow precise positioning of a detector together with small correctable deviations from absorbed dose to water. Recommendations for the selection of detectors and phantom materials are being developed for different measurements in brachytherapy. Conclusion: The introduction of kQ for brachytherapy sources may allow more systematic and comparable dose measurements. In principle, the corrections can be verified or even determined by measurement in a water phantom and comparison with dose distributions calculated using the TG43 dosimetry formalism. Project is supported by DIN Deutsches Institut fuer Normung.« less
OGLE-2012-bLG-0950Lb: the First Planet Mass Measurement From Only Microlens Parallax and Lens Flux
NASA Technical Reports Server (NTRS)
Koshimoto, N.; Udalski, A.; Beaulieu, J. P.; Sumi, T.; Bennett, D. P.; Bond, I. A.; Rattenbury, N.; Fukui, A.; Bhattacharya, A.; Suzuki, D.
2016-01-01
We report the discovery of a microlensing planet OGLE-2012-BLG-0950Lb with a planet/host mass ratio Periapsis Approx. = 2 x10(exp. -4). A long term distortion detected in both MOA and OGLE light curve can be explained by themicrolens parallax due to the Earths orbital motion around the Sun. Although the finite source effect is not detected, we obtain the lens flux by the high resolution Keck AO observation. Combining the microlens parallax and the lens flux reveal the nature of the lens: a planet with mass of M(sub p) = 35(+17/-)M compared to Earth is orbiting around an M-dwarf with mass of M(sub host) = 0.56(+0.12/-0.16) M compared to the Sun with a planet-host projected separation of r1 = 2.7(+0.6/-0.7) au located at Luminosity Distance = 3.0(+0.8/-1.1) kpc from us. This is the first mass measurement from only microlens parallax and the lens flux without the finite source effect. In the coming space observation-era with Spitzer, K2, Euclid, and WFIRST, we expect many such events for which we will not be able to measure any finite source effect. This work demonstrates an ability of mass measurements in such events.
Dannemiller, Karen C.; Murphy, Johnna S.; Dixon, Sherry L.; Pennell, Kelly G.; Suuberg, Eric M.; Jacobs, David E.; Sandel, Megan
2013-01-01
Formaldehyde is a colorless, pungent gas commonly found in homes that is a respiratory irritant, sensitizer, carcinogen and asthma trigger. Typical household sources include plywood and particleboard, cleaners, cosmetics, pesticides, and others. Development of a fast and simple measurement technique could facilitate continued research on this important chemical. The goal of this research is to apply an inexpensive short-term measurement method to find correlations between formaldehyde sources and concentration, and formaldehyde concentration and asthma control. Formaldehyde was measured using 30-minute grab samples in length-of-stain detector tubes in homes (n=70) of asthmatics in the Boston, MA area. Clinical status and potential formaldehyde sources were determined. The geometric mean formaldehyde level was 35.1 ppb and ranged from 5–132 ppb. Based on one-way ANOVA, t-tests, and linear regression, predictors of log-transformed formaldehyde concentration included absolute humidity, season, and the presence of decorative laminates, fiberglass, or permanent press fabrics (p<0.05), as well as temperature and household cleaner use (p<0.10). The geometric mean formaldehyde concentration was 57% higher in homes of children with very poorly controlled asthma compared to homes of other asthmatic children (p=0.078). This study provides a simple method for measuring household formaldehyde and suggests that exposure is related to poorly controlled asthma. PMID:23278296
Diffusive-convective physical vapor transport of PbTe from a Te-rich solid source
NASA Technical Reports Server (NTRS)
Zoutendyk, J.; Akutagawa, W.
1982-01-01
Crystal growth of PbTe by physical vapor transport (sublimation) in a closed ampoule is governed by the vapor species in thermal equilibrium with the solid compound. Deviations from stoichiometry in the source material cause diffusion limitation of the transport rate, which can be modified by natural (gravity-driven) convection. Mass-transport experiments have been performed using Te-rich material wherein sublimation rates have been measured in order to study the effects of natural convection in diffusion-limited vapor transport. Linear velocities for both crystal growth and evaporation (back sublimation) have been measured for transport in the direction of gravity, horizontally, and opposite to gravity. The experimental results are discussed in terms of both the one-dimensional diffusive-advective model and current, more sophisticated theory which includes natural convection. There is some evidence that convection effects from radial temperature gradients and solutal density gradients have been observed.
NASA Astrophysics Data System (ADS)
Mulia, Iyan E.; Gusman, Aditya Riadi; Satake, Kenji
2017-12-01
Recently, there are numerous tsunami observation networks deployed in several major tsunamigenic regions. However, guidance on where to optimally place the measurement devices is limited. This study presents a methodological approach to select strategic observation locations for the purpose of tsunami source characterizations, particularly in terms of the fault slip distribution. Initially, we identify favorable locations and determine the initial number of observations. These locations are selected based on extrema of empirical orthogonal function (EOF) spatial modes. To further improve the accuracy, we apply an optimization algorithm called a mesh adaptive direct search to remove redundant measurement locations from the EOF-generated points. We test the proposed approach using multiple hypothetical tsunami sources around the Nankai Trough, Japan. The results suggest that the optimized observation points can produce more accurate fault slip estimates with considerably less number of observations compared to the existing tsunami observation networks.
Quantifying Transmission of Clostridium difficile within and outside Healthcare Settings
Olsen, Margaret A.; Dubberke, Erik R.; Galvani, Alison P.; Townsend, Jeffrey P.
2016-01-01
To quantify the effect of hospital and community-based transmission and control measures on Clostridium difficile infection (CDI), we constructed a transmission model within and between hospital, community, and long-term care-facility settings. By parameterizing the model from national databases and calibrating it to C. difficile prevalence and CDI incidence, we found that hospitalized patients with CDI transmit C. difficile at a rate 15 (95% CI 7.2–32) times that of asymptomatic patients. Long-term care facility residents transmit at a rate of 27% (95% CI 13%–51%) that of hospitalized patients, and persons in the community at a rate of 0.1% (95% CI 0.062%–0.2%) that of hospitalized patients. Despite lower transmission rates for asymptomatic carriers and community sources, these transmission routes have a substantial effect on hospital-onset CDI because of the larger reservoir of hospitalized carriers and persons in the community. Asymptomatic carriers and community sources should be accounted for when designing and evaluating control interventions. PMID:26982504
Satellite Remote Sensing: Aerosol Measurements
NASA Technical Reports Server (NTRS)
Kahn, Ralph A.
2013-01-01
Aerosols are solid or liquid particles suspended in the air, and those observed by satellite remote sensing are typically between about 0.05 and 10 microns in size. (Note that in traditional aerosol science, the term "aerosol" refers to both the particles and the medium in which they reside, whereas for remote sensing, the term commonly refers to the particles only. In this article, we adopt the remote-sensing definition.) They originate from a great diversity of sources, such as wildfires, volcanoes, soils and desert sands, breaking waves, natural biological activity, agricultural burning, cement production, and fossil fuel combustion. They typically remain in the atmosphere from several days to a week or more, and some travel great distances before returning to Earth's surface via gravitational settling or washout by precipitation. Many aerosol sources exhibit strong seasonal variability, and most experience inter-annual fluctuations. As such, the frequent, global coverage that space-based aerosol remote-sensing instruments can provide is making increasingly important contributions to regional and larger-scale aerosol studies.
Gopalakrishnan, V; Baskaran, R; Venkatraman, B
2016-08-01
A decision support system (DSS) is implemented in Radiological Safety Division, Indira Gandhi Centre for Atomic Research for providing guidance for emergency decision making in case of an inadvertent nuclear accident. Real time gamma dose rate measurement around the stack is used for estimating the radioactive release rate (source term) by using inverse calculation. Wireless gamma dose logging network is designed, implemented, and installed around the Madras Atomic Power Station reactor stack to continuously acquire the environmental gamma dose rate and the details are presented in the paper. The network uses XBee-Pro wireless modules and PSoC controller for wireless interfacing, and the data are logged at the base station. A LabView based program is developed to receive the data, display it on the Google Map, plot the data over the time scale, and register the data in a file to share with DSS software. The DSS at the base station evaluates the real time source term to assess radiation impact.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gopalakrishnan, V.; Baskaran, R.; Venkatraman, B.
A decision support system (DSS) is implemented in Radiological Safety Division, Indira Gandhi Centre for Atomic Research for providing guidance for emergency decision making in case of an inadvertent nuclear accident. Real time gamma dose rate measurement around the stack is used for estimating the radioactive release rate (source term) by using inverse calculation. Wireless gamma dose logging network is designed, implemented, and installed around the Madras Atomic Power Station reactor stack to continuously acquire the environmental gamma dose rate and the details are presented in the paper. The network uses XBee–Pro wireless modules and PSoC controller for wireless interfacing,more » and the data are logged at the base station. A LabView based program is developed to receive the data, display it on the Google Map, plot the data over the time scale, and register the data in a file to share with DSS software. The DSS at the base station evaluates the real time source term to assess radiation impact.« less
Cost-effective bidirectional digitized radio-over-fiber systems employing sigma delta modulation
NASA Astrophysics Data System (ADS)
Lee, Kyung Woon; Jung, HyunDo; Park, Jung Ho
2016-11-01
We propose a cost effective digitized radio-over-fiber (D-RoF) system employing a sigma delta modulation (SDM) and a bidirectional transmission technique using phase modulated downlink and intensity modulated uplink. SDM is transparent to different radio access technologies and modulation formats, and more suitable for a downlink of wireless system because a digital to analog converter (DAC) can be avoided at the base station (BS). Also, Central station and BS share the same light source by using a phase modulation for the downlink and an intensity modulation for the uplink transmission. Avoiding DACs and light sources have advantages in terms of cost reduction, power consumption, and compatibility with conventional wireless network structure. We have designed a cost effective bidirectional D-RoF system using a low pass SDM and measured the downlink and uplink transmission performance in terms of error vector magnitude, signal spectra, and constellations, which are based on the 10MHz LTE 64-QAM standard.
Characterization of air manganese exposure estimates for residents in two Ohio towns
Colledge, Michelle A.; Julian, Jaime R.; Gocheva, Vihra V.; Beseler, Cheryl L.; Roels, Harry A.; Lobdell, Danelle T.; Bowler, Rosemarie M.
2016-01-01
This study was conducted to derive receptor-specific outdoor exposure concentrations of total suspended particulate (TSP) and respirable (dae ≤ 10 μm) air manganese (air-Mn) for East Liverpool and Marietta (Ohio) in the absence of facility emissions data, but where long-term air measurements were available. Our “site-surface area emissions method” used U.S. Environmental Protection Agency’s (EPA) AERMOD (AMS/EPA Regulatory Model) dispersion model and air measurement data to estimate concentrations for residential receptor sites in the two communities. Modeled concentrations were used to create ratios between receptor points and calibrated using measured data from local air monitoring stations. Estimated outdoor air-Mn concentrations were derived for individual study subjects in both towns. The mean estimated long-term air-Mn exposure levels for total suspended particulate were 0.35 μg/m3 (geometric mean [GM]) and 0.88 μg/m3 (arithmetic mean [AM]) in East Liverpool (range: 0.014–6.32 μg/m3) and 0.17 μg/m3 (GM) and 0.21 μg/m3 (AM) in Marietta (range: 0.03–1.61 μg/m3). Modeled results compared well with averaged ambient air measurements from local air monitoring stations. Exposure to respirable Mn particulate matter (PM10; PM <10 μm) was higher in Marietta residents. Implications Few available studies evaluate long-term health outcomes from inhalational manganese (Mn) exposure in residential populations, due in part to challenges in measuring individual exposures. Local long-term air measurements provide the means to calibrate models used in estimating long-term exposures. Furthermore, this combination of modeling and ambient air sampling can be used to derive receptor-specific exposure estimates even in the absence of source emissions data for use in human health outcome studies. PMID:26211636
Fermi Large Area Telescope Second Source Catalog
NASA Astrophysics Data System (ADS)
Nolan, P. L.; Abdo, A. A.; Ackermann, M.; Ajello, M.; Allafort, A.; Antolini, E.; Atwood, W. B.; Axelsson, M.; Baldini, L.; Ballet, J.; Barbiellini, G.; Bastieri, D.; Bechtol, K.; Belfiore, A.; Bellazzini, R.; Berenji, B.; Bignami, G. F.; Blandford, R. D.; Bloom, E. D.; Bonamente, E.; Bonnell, J.; Borgland, A. W.; Bottacini, E.; Bouvier, A.; Brandt, T. J.; Bregeon, J.; Brigida, M.; Bruel, P.; Buehler, R.; Burnett, T. H.; Buson, S.; Caliandro, G. A.; Cameron, R. A.; Campana, R.; Cañadas, B.; Cannon, A.; Caraveo, P. A.; Casandjian, J. M.; Cavazzuti, E.; Ceccanti, M.; Cecchi, C.; Çelik, Ö.; Charles, E.; Chekhtman, A.; Cheung, C. C.; Chiang, J.; Chipaux, R.; Ciprini, S.; Claus, R.; Cohen-Tanugi, J.; Cominsky, L. R.; Conrad, J.; Corbet, R.; Cutini, S.; D'Ammando, F.; Davis, D. S.; de Angelis, A.; DeCesar, M. E.; DeKlotz, M.; De Luca, A.; den Hartog, P. R.; de Palma, F.; Dermer, C. D.; Digel, S. W.; Silva, E. do Couto e.; Drell, P. S.; Drlica-Wagner, A.; Dubois, R.; Dumora, D.; Enoto, T.; Escande, L.; Fabiani, D.; Falletti, L.; Favuzzi, C.; Fegan, S. J.; Ferrara, E. C.; Focke, W. B.; Fortin, P.; Frailis, M.; Fukazawa, Y.; Funk, S.; Fusco, P.; Gargano, F.; Gasparrini, D.; Gehrels, N.; Germani, S.; Giebels, B.; Giglietto, N.; Giommi, P.; Giordano, F.; Giroletti, M.; Glanzman, T.; Godfrey, G.; Grenier, I. A.; Grondin, M.-H.; Grove, J. E.; Guillemot, L.; Guiriec, S.; Gustafsson, M.; Hadasch, D.; Hanabata, Y.; Harding, A. K.; Hayashida, M.; Hays, E.; Hill, A. B.; Horan, D.; Hou, X.; Hughes, R. E.; Iafrate, G.; Itoh, R.; Jóhannesson, G.; Johnson, R. P.; Johnson, T. E.; Johnson, A. S.; Johnson, T. J.; Kamae, T.; Katagiri, H.; Kataoka, J.; Katsuta, J.; Kawai, N.; Kerr, M.; Knödlseder, J.; Kocevski, D.; Kuss, M.; Lande, J.; Landriu, D.; Latronico, L.; Lemoine-Goumard, M.; Lionetto, A. M.; Llena Garde, M.; Longo, F.; Loparco, F.; Lott, B.; Lovellette, M. N.; Lubrano, P.; Madejski, G. M.; Marelli, M.; Massaro, E.; Mazziotta, M. N.; McConville, W.; McEnery, J. E.; Mehault, J.; Michelson, P. F.; Minuti, M.; Mitthumsiri, W.; Mizuno, T.; Moiseev, A. A.; Mongelli, M.; Monte, C.; Monzani, M. E.; Morselli, A.; Moskalenko, I. V.; Murgia, S.; Nakamori, T.; Naumann-Godo, M.; Norris, J. P.; Nuss, E.; Nymark, T.; Ohno, M.; Ohsugi, T.; Okumura, A.; Omodei, N.; Orlando, E.; Ormes, J. F.; Ozaki, M.; Paneque, D.; Panetta, J. H.; Parent, D.; Perkins, J. S.; Pesce-Rollins, M.; Pierbattista, M.; Pinchera, M.; Piron, F.; Pivato, G.; Porter, T. A.; Racusin, J. L.; Rainò, S.; Rando, R.; Razzano, M.; Razzaque, S.; Reimer, A.; Reimer, O.; Reposeur, T.; Ritz, S.; Rochester, L. S.; Romani, R. W.; Roth, M.; Rousseau, R.; Ryde, F.; Sadrozinski, H. F.-W.; Salvetti, D.; Sanchez, D. A.; Saz Parkinson, P. M.; Sbarra, C.; Scargle, J. D.; Schalk, T. L.; Sgrò, C.; Shaw, M. S.; Shrader, C.; Siskind, E. J.; Smith, D. A.; Spandre, G.; Spinelli, P.; Stephens, T. E.; Strickman, M. S.; Suson, D. J.; Tajima, H.; Takahashi, H.; Takahashi, T.; Tanaka, T.; Thayer, J. G.; Thayer, J. B.; Thompson, D. J.; Tibaldo, L.; Tibolla, O.; Tinebra, F.; Tinivella, M.; Torres, D. F.; Tosti, G.; Troja, E.; Uchiyama, Y.; Vandenbroucke, J.; Van Etten, A.; Van Klaveren, B.; Vasileiou, V.; Vianello, G.; Vitale, V.; Waite, A. P.; Wallace, E.; Wang, P.; Werner, M.; Winer, B. L.; Wood, D. L.; Wood, K. S.; Wood, M.; Yang, Z.; Zimmer, S.
2012-04-01
We present the second catalog of high-energy γ-ray sources detected by the Large Area Telescope (LAT), the primary science instrument on the Fermi Gamma-ray Space Telescope (Fermi), derived from data taken during the first 24 months of the science phase of the mission, which began on 2008 August 4. Source detection is based on the average flux over the 24 month period. The second Fermi-LAT catalog (2FGL) includes source location regions, defined in terms of elliptical fits to the 95% confidence regions and spectral fits in terms of power-law, exponentially cutoff power-law, or log-normal forms. Also included are flux measurements in five energy bands and light curves on monthly intervals for each source. Twelve sources in the catalog are modeled as spatially extended. We provide a detailed comparison of the results from this catalog with those from the first Fermi-LAT catalog (1FGL). Although the diffuse Galactic and isotropic models used in the 2FGL analysis are improved compared to the 1FGL catalog, we attach caution flags to 162 of the sources to indicate possible confusion with residual imperfections in the diffuse model. The 2FGL catalog contains 1873 sources detected and characterized in the 100 MeV to 100 GeV range of which we consider 127 as being firmly identified and 1171 as being reliably associated with counterparts of known or likely γ-ray-producing source classes. We dedicate this paper to the memory of our colleague Patrick Nolan, who died on 2011 November 6. His career spanned much of the history of high-energy astronomy from space and his work on the Large Area Telescope (LAT) began nearly 20 years ago when it was just a concept. Pat was a central member in the operation of the LAT collaboration and he is greatly missed.
Delivery of Modular Lethality via a Parent-Child Concept
2015-02-01
time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the...downrange distance to the target, is the time of flight, is the distance of the thruster force from the body center of gravity, and is...velocity and time of flight can be estimated or measured in flight. These values can be collected in a term, , and the 2 components of lateral
Verification of Methods for Assessing the Sustainability of Monitored Natural Attenuation (MNA)
2013-01-01
sugars TOC total organic carbon TSR thermal source removal USACE U.S. Army Corps of Engineers USEPA U.S. Environmental Protection Agency USGS...the SZD function for long-term DNAPL dissolution simulations. However, the sustainability assessment was easily implemented using an alternative...neutral sugars [THNS]). Chapelle et al. (2009) suggested THAA and THNS as measures of the bioavailability of organic carbon based on an analysis of
Characterization and Reliability of Vertical N-Type Gallium Nitride Schottky Contacts
2016-09-01
barrier diode SEM scanning electron microscopy SiC silicon carbide SMU source measure unit xvi THIS PAGE INTENTIONALLY LEFT BLANK xvii...arguably the Schottky barrier diode (SBD). The SBD is a fundamental component in the majority of power electronic devices; specifically, those used in...Ishizuka, and Ueno demonstrated the long-term reliability of vertical metal-GaN Schottky barrier diodes through their analysis of the degradation
Long-Term Observation of Small and Medium-Scale Gravity Waves over the Brazilian Equatorial Region
NASA Astrophysics Data System (ADS)
Essien, Patrick; Buriti, Ricardo; Wrasse, Cristiano M.; Medeiros, Amauri; Paulino, Igo; Takahashi, Hisao; Campos, Jose Andre
2016-07-01
This paper reports the long term observations of small and medium-scale gravity waves over Brazilian equatorial region. Coordinated optical and radio measurements were made from OLAP at Sao Joao do Cariri (7.400S, 36.500W) to investigate the occurrences and properties and to characterize the regional mesospheric gravity wave field. All-sky imager measurements were made from the site. for almost 11 consecutive years (September 2000 to November 2010). Most of the waves propagated were characterized as small-scale gravity. The characteristics of the two waves events agreed well with previous gravity wave studies from Brazil and other sites. However, significant differences in the wave propagation headings indicate dissimilar source regions. The observed medium-scale gravity wave events constitute an important new dataset to study their mesospheric properties at equatorial latitudes. These data exhibited similar propagation headings to the short period events, suggesting they originated from the same source regions. It was also observed that some of the medium-scale were capable of propagating into the lower thermosphere where they may have acted directly as seeds for the Rayleigh-Taylor instability development. The wave events were primarily generated by meteorological processes since there was no correlation between the evolution of the wave events and solar cycle F10.7.
CSI-EPT in Presence of RF-Shield for MR-Coils.
Arduino, Alessandro; Zilberti, Luca; Chiampi, Mario; Bottauscio, Oriano
2017-07-01
Contrast source inversion electric properties tomography (CSI-EPT) is a recently developed technique for the electric properties tomography that recovers the electric properties distribution starting from measurements performed by magnetic resonance imaging scanners. This method is an optimal control approach based on the contrast source inversion technique, which distinguishes itself from other electric properties tomography techniques for its capability to recover also the local specific absorption rate distribution, essential for online dosimetry. Up to now, CSI-EPT has only been described in terms of integral equations, limiting its applicability to homogeneous unbounded background. In order to extend the method to the presence of a shield in the domain-as in the recurring case of shielded radio frequency coils-a more general formulation of CSI-EPT, based on a functional viewpoint, is introduced here. Two different implementations of CSI-EPT are proposed for a 2-D transverse magnetic model problem, one dealing with an unbounded domain and one considering the presence of a perfectly conductive shield. The two implementations are applied on the same virtual measurements obtained by numerically simulating a shielded radio frequency coil. The results are compared in terms of both electric properties recovery and local specific absorption rate estimate, in order to investigate the requirement of an accurate modeling of the underlying physical problem.
Characterization and Evolution of the Swift X-ray Telescope Instrumental Background
NASA Technical Reports Server (NTRS)
Hill, Joanne; Pagani, C.; Morris, D. C.; Racusin, J.; Grupe, D.; Vetere, L.; Stroh, M.; Falcone, A.; Kennea, J.; Burrows, D. N.;
2007-01-01
The X-ray telescope (XRT) on board the Swift Gamma Ray Burst Explorer has successfully operated since the spacecraft launch on 20 November 2004, automatically locating GRB afterglows, measuring their spectra and lightcurves and performing observations of high-energy sources. In this work we investigate the properties of the instrumental background, focusing on its dynamic behavior on both long and short timescales. The operational temperature of the CCD is the main factor that influences the XRT background level. After the failure of the Swift active on-board temperature control system, the XRT detector now operates at a temperature range between -75C and -45C thanks to a passive cooling Heat Rejection System. We report on the long-term effects on the background caused by radiation, consisting mainly of proton irradiation in Swift's low Earth orbit and on the short-term effects of transits through the South Atlantic Anomaly (SAA), which expose the detector to periods of intense proton flux. We have determined the fraction of the detector background that is due to the internal, instrumental background and the part that is due to unresolved astrophysical sources (the cosmic X-ray background) by investigating the degree of vignetting of the measured background and comparing it to the expected value from calibration data.
Effects of air pollution on human health and practical measures for prevention in Iran
Ghorani-Azam, Adel; Riahi-Zanjani, Bamdad; Balali-Mood, Mahdi
2016-01-01
Air pollution is a major concern of new civilized world, which has a serious toxicological impact on human health and the environment. It has a number of different emission sources, but motor vehicles and industrial processes contribute the major part of air pollution. According to the World Health Organization, six major air pollutants include particle pollution, ground-level ozone, carbon monoxide, sulfur oxides, nitrogen oxides, and lead. Long and short term exposure to air suspended toxicants has a different toxicological impact on human including respiratory and cardiovascular diseases, neuropsychiatric complications, the eyes irritation, skin diseases, and long-term chronic diseases such as cancer. Several reports have revealed the direct association between exposure to the poor air quality and increasing rate of morbidity and mortality mostly due to cardiovascular and respiratory diseases. Air pollution is considered as the major environmental risk factor in the incidence and progression of some diseases such as asthma, lung cancer, ventricular hypertrophy, Alzheimer's and Parkinson's diseases, psychological complications, autism, retinopathy, fetal growth, and low birth weight. In this review article, we aimed to discuss toxicology of major air pollutants, sources of emission, and their impact on human health. We have also proposed practical measures to reduce air pollution in Iran. PMID:27904610
Effects of air pollution on human health and practical measures for prevention in Iran.
Ghorani-Azam, Adel; Riahi-Zanjani, Bamdad; Balali-Mood, Mahdi
2016-01-01
Air pollution is a major concern of new civilized world, which has a serious toxicological impact on human health and the environment. It has a number of different emission sources, but motor vehicles and industrial processes contribute the major part of air pollution. According to the World Health Organization, six major air pollutants include particle pollution, ground-level ozone, carbon monoxide, sulfur oxides, nitrogen oxides, and lead. Long and short term exposure to air suspended toxicants has a different toxicological impact on human including respiratory and cardiovascular diseases, neuropsychiatric complications, the eyes irritation, skin diseases, and long-term chronic diseases such as cancer. Several reports have revealed the direct association between exposure to the poor air quality and increasing rate of morbidity and mortality mostly due to cardiovascular and respiratory diseases. Air pollution is considered as the major environmental risk factor in the incidence and progression of some diseases such as asthma, lung cancer, ventricular hypertrophy, Alzheimer's and Parkinson's diseases, psychological complications, autism, retinopathy, fetal growth, and low birth weight. In this review article, we aimed to discuss toxicology of major air pollutants, sources of emission, and their impact on human health. We have also proposed practical measures to reduce air pollution in Iran.
NASA Technical Reports Server (NTRS)
Yee, H. C.; Shinn, J. L.
1986-01-01
Some numerical aspects of finite-difference algorithms for nonlinear multidimensional hyperbolic conservation laws with stiff nonhomogenous (source) terms are discussed. If the stiffness is entirely dominated by the source term, a semi-implicit shock-capturing method is proposed provided that the Jacobian of the soruce terms possesses certain properties. The proposed semi-implicit method can be viewed as a variant of the Bussing and Murman point-implicit scheme with a more appropriate numerical dissipation for the computation of strong shock waves. However, if the stiffness is not solely dominated by the source terms, a fully implicit method would be a better choice. The situation is complicated by problems that are higher than one dimension, and the presence of stiff source terms further complicates the solution procedures for alternating direction implicit (ADI) methods. Several alternatives are discussed. The primary motivation for constructing these schemes was to address thermally and chemically nonequilibrium flows in the hypersonic regime. Due to the unique structure of the eigenvalues and eigenvectors for fluid flows of this type, the computation can be simplified, thus providing a more efficient solution procedure than one might have anticipated.
What do popular Spanish women's magazines say about caesarean section? A 21-year survey
Torloni, MR; Campos Mansilla, B; Merialdi, M; Betrán, AP
2014-01-01
Objectives Caesarean section (CS) rates are increasing worldwide and maternal request is cited as one of the main reasons for this trend. Women's preferences for route of delivery are influenced by popular media, including magazines. We assessed the information on CS presented in Spanish women's magazines. Design Systematic review. Setting Women's magazines printed from 1989 to 2009 with the largest national distribution. Sample Articles with any information on CS. Methods Articles were selected, read and abstracted in duplicate. Sources of information, scientific accuracy, comprehensiveness and women's testimonials were objectively extracted using a content analysis form designed for this study. Main outcome measures Accuracy, comprehensiveness and sources of information. Results Most (67%) of the 1223 selected articles presented exclusively personal opinion/birth stories, 12% reported the potential benefits of CS, 26% mentioned the short-term and 10% mentioned the long-term maternal risks, and 6% highlighted the perinatal risks of CS. The most frequent short-term risks were the increased time for maternal recovery (n = 86), frustration/feelings of failure (n = 83) and increased post-surgical pain (n = 71). The most frequently cited long-term risks were uterine rupture (n = 57) and the need for another CS in any subsequent pregnancy (n = 42). Less than 5% of the selected articles reported that CS could increase the risks of infection (n = 53), haemorrhage (n = 31) or placenta praevia/accreta in future pregnancies (n = 6). The sources of information were not reported by 68% of the articles. Conclusions The portrayal of CS in Spanish women's magazines is not sufficiently comprehensive and does not provide adequate important information to help the readership to understand the real benefits and risks of this route of delivery. PMID:24467797
Hobson, Keith A; Blight, Louise K; Arcese, Peter
2015-09-15
Measurements of naturally occurring stable isotopes in tissues of seabirds and their prey are a powerful tool for investigating long-term changes in marine foodwebs. Recent isotopic (δ(15)N, δ(13)C) evidence from feathers of Glaucous-winged Gulls (Larus glaucescens) has shown that over the last 150 years, this species shifted from a midtrophic marine diet to one including lower trophic marine prey and/or more terrestrial or freshwater foods. However, long-term isotopic patterns of δ(15)N and δ(13)C cannot distinguish between the relative importance of lower trophic-level marine foods and terrestrial sources. We examined 48 feather stable-hydrogen (δ(2)H) and -sulfur (δ(34)S) isotope values from this same 150-year feather set and found additional isotopic evidence supporting the hypothesis that gulls shifted to terrestrial and/or freshwater prey. Mean feather δ(2)H and δ(34)S values (± SD) declined from the earliest period (1860-1915; n = 12) from -2.5 ± 21.4 ‰ and 18.9 ± 2.7 ‰, respectively, to -35.5 ± 15.5 ‰ and 14.8 ± 2.4 ‰, respectively, for the period 1980-2009 (n = 12). We estimated a shift of ∼ 30% increase in dependence on terrestrial/freshwater sources. These results are consistent with the hypothesis that gulls increased terrestrial food inputs in response to declining forage fish availability.
White Paper on SBUV/2 Solar Irradiance Measurements
NASA Technical Reports Server (NTRS)
Hilsenrath, Ernest; DeLand, Matthew T.; Cebula, Richard P.
1996-01-01
The importance of solar irradiance measurements by the Solar Backscatter Ultraviolet, Model 2 (SBUV/2) instruments on NOAA's operational satellites is described. These measurements are necessary accurately monitor the long-term changes in the global column ozone amount, the altitude distribution of ozone in the upper stratosphere, and the degree to which ozone changes are caused by anthropogenic sources. Needed to accomplish these goals are weekly solar irradiance measurements at the operational ozone wavelengths, daily measurements of the Mg II proxy index, instrument-specific Mg II scale factors, and daily measurements of the solar spectral irradiance at photochemically important wavelengths. Two solar measurement schedules are provided: (1) a baseline schedule for all instruments except the NOAA-14 instrument and (2) a modified schedule for the NOAA-14 SBUV/2 instrument. This latter schedule is needed due to the NOAA-14 grating drive problems.
Characterization of Industrial Emission Sources and Photochemistry in Houston, Texas
NASA Astrophysics Data System (ADS)
Washenfelder, R. A.; Atlas, E. L.; Degouw, J.; Flocke, F. M.; Fried, A.; Frost, G. J.; Holloway, J.; Richter, D.; Ryerson, T. B.; Schauffler, S.; Trainer, M.; Walega, J.; Warneke, C.; Weibring, P.; Zheng, W.
2009-12-01
The Houston-Galveston urban area contains a number of large industrial petrochemical emission sources that produce volatile organic compounds and nitrogen oxides. These co-located emissions result in rapid and efficient ozone production downwind. Unlike a single large power plant, the industrial complexes consist of numerous sources that can be difficult to quantify in emission inventories. During September - October 2006, the NOAA WP-3 aircraft conducted research flights as part of the second Texas Air Quality Study (TexAQS II). We examine measurements of NOx, SO2, and speciated hydrocarbons from the Houston Ship Channel, which contains a dense concentration of industrial petrochemical sources, and isolated petrochemical facilities. These measurements are used to derive source emission estimates, which are then compared to available emission inventories. We find that high hydrocarbon emissions are typical for the Houston Ship Channel and isolated petrochemical facilities. Ethene and propene are found to be major contributors to ozone formation. Ratios of C2H4 / NOx and C3H6 / NOx exceed emission inventory values by factors of 10 - 50. These findings are consistent with the first TexAQS study in 2000. We examine trends in C2H4 / NOx and C3H6 / NOx ratios between 2000 and 2006, and determine that day-to-day variability and within-plume variability exceeds any long-term reduction in ethene and propene emissions for the isolated petrochemical sources. We additionally examine downwind photochemical products formed by these alkenes.
NASA Astrophysics Data System (ADS)
Chino, Masamichi; Terada, Hiroaki; Nagai, Haruyasu; Katata, Genki; Mikami, Satoshi; Torii, Tatsuo; Saito, Kimiaki; Nishizawa, Yukiyasu
2016-08-01
The Fukushima Daiichi nuclear power reactor units that generated large amounts of airborne discharges during the period of March 12-21, 2011 were identified individually by analyzing the combination of measured 134Cs/137Cs depositions on ground surfaces and atmospheric transport and deposition simulations. Because the values of 134Cs/137Cs are different in reactor units owing to fuel burnup differences, the 134Cs/137Cs ratio measured in the environment was used to determine which reactor unit ultimately contaminated a specific area. Atmospheric dispersion model simulations were used for predicting specific areas contaminated by each dominant release. Finally, by comparing the results from both sources, the specific reactor units that yielded the most dominant atmospheric release quantities could be determined. The major source reactor units were Unit 1 in the afternoon of March 12, 2011, Unit 2 during the period from the late night of March 14 to the morning of March 15, 2011. These results corresponded to those assumed in our previous source term estimation studies. Furthermore, new findings suggested that the major source reactors from the evening of March 15, 2011 were Units 2 and 3 and that the dominant source reactor on March 20, 2011 temporally changed from Unit 3 to Unit 2.
Chino, Masamichi; Terada, Hiroaki; Nagai, Haruyasu; Katata, Genki; Mikami, Satoshi; Torii, Tatsuo; Saito, Kimiaki; Nishizawa, Yukiyasu
2016-08-22
The Fukushima Daiichi nuclear power reactor units that generated large amounts of airborne discharges during the period of March 12-21, 2011 were identified individually by analyzing the combination of measured (134)Cs/(137)Cs depositions on ground surfaces and atmospheric transport and deposition simulations. Because the values of (134)Cs/(137)Cs are different in reactor units owing to fuel burnup differences, the (134)Cs/(137)Cs ratio measured in the environment was used to determine which reactor unit ultimately contaminated a specific area. Atmospheric dispersion model simulations were used for predicting specific areas contaminated by each dominant release. Finally, by comparing the results from both sources, the specific reactor units that yielded the most dominant atmospheric release quantities could be determined. The major source reactor units were Unit 1 in the afternoon of March 12, 2011, Unit 2 during the period from the late night of March 14 to the morning of March 15, 2011. These results corresponded to those assumed in our previous source term estimation studies. Furthermore, new findings suggested that the major source reactors from the evening of March 15, 2011 were Units 2 and 3 and that the dominant source reactor on March 20, 2011 temporally changed from Unit 3 to Unit 2.
Chino, Masamichi; Terada, Hiroaki; Nagai, Haruyasu; Katata, Genki; Mikami, Satoshi; Torii, Tatsuo; Saito, Kimiaki; Nishizawa, Yukiyasu
2016-01-01
The Fukushima Daiichi nuclear power reactor units that generated large amounts of airborne discharges during the period of March 12–21, 2011 were identified individually by analyzing the combination of measured 134Cs/137Cs depositions on ground surfaces and atmospheric transport and deposition simulations. Because the values of 134Cs/137Cs are different in reactor units owing to fuel burnup differences, the 134Cs/137Cs ratio measured in the environment was used to determine which reactor unit ultimately contaminated a specific area. Atmospheric dispersion model simulations were used for predicting specific areas contaminated by each dominant release. Finally, by comparing the results from both sources, the specific reactor units that yielded the most dominant atmospheric release quantities could be determined. The major source reactor units were Unit 1 in the afternoon of March 12, 2011, Unit 2 during the period from the late night of March 14 to the morning of March 15, 2011. These results corresponded to those assumed in our previous source term estimation studies. Furthermore, new findings suggested that the major source reactors from the evening of March 15, 2011 were Units 2 and 3 and that the dominant source reactor on March 20, 2011 temporally changed from Unit 3 to Unit 2. PMID:27546490
NASA Astrophysics Data System (ADS)
Mønster, J.; Rella, C.; Jacobson, G. A.; He, Y.; Hoffnagle, J.; Scheutz, C.
2012-12-01
Nitrous oxide is a powerful greenhouse gas considered 298 times stronger than carbon dioxide on a hundred years term (Solomon et al. 2007). The increasing global concentration is of great concern and is receiving increasing attention in various scientific and industrial fields. Nitrous oxide is emitted from both natural and anthropogenic sources. Inventories of source specific fugitive nitrous oxide emissions are often estimated on the basis of modeling and mass balance. While these methods are well-developed, actual measurements for quantification of the emissions can be a useful tool for verifying the existing estimation methods as well as providing validation for initiatives targeted at lowering unwanted nitrous oxide emissions. One approach to performing such measurements is the tracer dilution method (Galle et al. 2001), in which a tracer gas is released at the source location at a known flow. The ratio of downwind concentrations of both the tracer gas and nitrous oxide gives the ratios of the emissions rates. This tracer dilution method can be done with both stationary and mobile measurements; in either case, real-time measurements of both tracer and analyte gas is required, which places high demands on the analytical detection method. To perform the nitrous oxide measurements, a novel, robust instrument capable of real-time nitrous oxide measurements has been developed, based on cavity ring-down spectroscopy and operating in the near-infrared spectral region. We present the results of the laboratory and field tests of this instrument in both California and Denmark. Furthermore, results are presented from measurements using the mobile plume method with a tracer gas (acetylene) to quantify the nitrous oxide and methane emissions from known sources such as waste water treatment plants and composting facilities. Nitrous oxide (blue) and methane (yellow) plumes downwind from a waste water treatment facility.
Coal burning leaves toxic heavy metal legacy in the Arctic
DOE Office of Scientific and Technical Information (OSTI.GOV)
McConnell, J.R.; Edwards, R.
2008-08-26
Toxic heavy metals emitted by industrial activities in the midlatitudes are transported through the atmosphere and deposited in the polar regions; bioconcentration and biomagnification in the food chain mean that even low levels of atmospheric deposition may threaten human health and Arctic ecosystems. Little is known about sources and long-term trends of most heavy metals before approximate to 1980, when modern measurements began, although heavy-metal pollution in the Arctic was widespread during recent decades. Lacking detailed, long-term measurements until now, ecologists, health researchers, and policy makers generally have assumed that contamination was highest during the 1960s and 1970s peak ofmore » industrial activity in North America and Europe. We present continuous 1772-2003 monthly and annually averaged deposition records for highly toxic thallium, cadmium, and lead from a Greenland ice core showing that atmospheric deposition was much higher than expected in the early 20th century, with tenfold increases from preindustrial levels by the early 1900s that were two to five times higher than during recent decades. Tracer measurements indicate that coal burning in North America and Europe was the likely source of these metals in the Arctic after 1860. Although these results show that heavy-metal pollution in the North Atlantic sector of the Arctic is substantially lower today than a century ago, contamination of other sectors may be increasing because of the rapid coal-driven growth of Asian economies.« less
NASA Astrophysics Data System (ADS)
Donne, Sarah; Bean, Christopher; Craig, David; Dias, Frederic; Christodoulides, Paul
2016-04-01
Microseisms are continuous seismic vibrations which propagate mainly as surface Rayleigh and Love waves. They are generated by the Earth's oceans and there are two main types; primary and secondary microseisms. Primary microseisms are generated through the interaction of travelling surface gravity ocean waves with the seafloor in shallow waters relative to the wavelength of the ocean wave. Secondary microseisms, on the other hand are generated when two opposing wave trains interact and a non-linear second order effect produces a pressure fluctuation which is depth independent. The conditions necessary to produce secondary microseisms are presented in Longuet-Higgins (1950) through the interaction of two travelling waves with the same wave period and which interact at an angle of 180 degrees. Equivalent surface pressure density (p2l) is modelled using the numerical ocean wave model Wavewatch III and this term is considered as the microseism source term. This work presents an investigation of the theoretical second order pressures generated through the interaction of travelling waves with varying wave amplitude, period and angle of incidence. Predicted seafloor pressures calculated off the Southwest coast of Ireland are compared with terrestrially recorded microseism records, measured seafloor pressures and oceanographic parameters. The work presented in this study suggests that a broad set of sea states can generate second order seafloor pressures that are consistent with seafloor pressure measurements. Local seismic arrays throughout Ireland allow us to investigate the temporal covariance of these seafloor pressures with microseism source locations.
Survey on the Performance of Source Localization Algorithms.
Fresno, José Manuel; Robles, Guillermo; Martínez-Tarifa, Juan Manuel; Stewart, Brian G
2017-11-18
The localization of emitters using an array of sensors or antennas is a prevalent issue approached in several applications. There exist different techniques for source localization, which can be classified into multilateration, received signal strength (RSS) and proximity methods. The performance of multilateration techniques relies on measured time variables: the time of flight (ToF) of the emission from the emitter to the sensor, the time differences of arrival (TDoA) of the emission between sensors and the pseudo-time of flight (pToF) of the emission to the sensors. The multilateration algorithms presented and compared in this paper can be classified as iterative and non-iterative methods. Both standard least squares (SLS) and hyperbolic least squares (HLS) are iterative and based on the Newton-Raphson technique to solve the non-linear equation system. The metaheuristic technique particle swarm optimization (PSO) used for source localisation is also studied. This optimization technique estimates the source position as the optimum of an objective function based on HLS and is also iterative in nature. Three non-iterative algorithms, namely the hyperbolic positioning algorithms (HPA), the maximum likelihood estimator (MLE) and Bancroft algorithm, are also presented. A non-iterative combined algorithm, MLE-HLS, based on MLE and HLS, is further proposed in this paper. The performance of all algorithms is analysed and compared in terms of accuracy in the localization of the position of the emitter and in terms of computational time. The analysis is also undertaken with three different sensor layouts since the positions of the sensors affect the localization; several source positions are also evaluated to make the comparison more robust. The analysis is carried out using theoretical time differences, as well as including errors due to the effect of digital sampling of the time variables. It is shown that the most balanced algorithm, yielding better results than the other algorithms in terms of accuracy and short computational time, is the combined MLE-HLS algorithm.
Survey on the Performance of Source Localization Algorithms
2017-01-01
The localization of emitters using an array of sensors or antennas is a prevalent issue approached in several applications. There exist different techniques for source localization, which can be classified into multilateration, received signal strength (RSS) and proximity methods. The performance of multilateration techniques relies on measured time variables: the time of flight (ToF) of the emission from the emitter to the sensor, the time differences of arrival (TDoA) of the emission between sensors and the pseudo-time of flight (pToF) of the emission to the sensors. The multilateration algorithms presented and compared in this paper can be classified as iterative and non-iterative methods. Both standard least squares (SLS) and hyperbolic least squares (HLS) are iterative and based on the Newton–Raphson technique to solve the non-linear equation system. The metaheuristic technique particle swarm optimization (PSO) used for source localisation is also studied. This optimization technique estimates the source position as the optimum of an objective function based on HLS and is also iterative in nature. Three non-iterative algorithms, namely the hyperbolic positioning algorithms (HPA), the maximum likelihood estimator (MLE) and Bancroft algorithm, are also presented. A non-iterative combined algorithm, MLE-HLS, based on MLE and HLS, is further proposed in this paper. The performance of all algorithms is analysed and compared in terms of accuracy in the localization of the position of the emitter and in terms of computational time. The analysis is also undertaken with three different sensor layouts since the positions of the sensors affect the localization; several source positions are also evaluated to make the comparison more robust. The analysis is carried out using theoretical time differences, as well as including errors due to the effect of digital sampling of the time variables. It is shown that the most balanced algorithm, yielding better results than the other algorithms in terms of accuracy and short computational time, is the combined MLE-HLS algorithm. PMID:29156565
NASA Astrophysics Data System (ADS)
Nara, H.; Tanimoto, H.; Mukai, H.; Nojiri, Y.; Tohjima, Y.; Machida, T.; Hashimoto, S.
2011-12-01
The National Institute for Environmental Studies (NIES) has been performing a long-term program for monitoring trace gases of atmospheric importance over the Pacific Ocean since 1995. The NIES Voluntary Observing Ships (NIES-VOS) program currently makes use of commercial cargo vessels because they operate regularly over fixed routes for long periods and sail over a wide area between various ports (e.g., between Japan and the United States, between Japan and Australia/New Zealand, and between Japan and southeast Asia). This program allows systematic and continuous measurements of non-CO2 greenhouse gases, providing long-term datasets for background air over the Pacific Ocean and regionally polluted air around east Asia. We observe both long-lived greenhouse gases (e.g., carbon dioxide) and short-lived air pollutants (e.g., tropospheric ozone, carbon monoxide) on a continuous basis. Flask samples are collected for later laboratory analysis of carbon dioxide, methane, nitrous oxide, and carbon monoxide by using gas chromatographic techniques. In addition, we recently installed cavity ringdown spectrometers for high-resolution measurement of methane and carbon dioxide to capture their highly variable features in regionally polluted air around southeast Asia (e.g., Hong Kong, Thailand, Singapore, Malaysia, Indonesia and Philippine), which is now thought to be a large source due to expanding socioeconomic activities as well as biomass burnings. Contrasting the Japan-Australia/New Zealand and Japan-southeast Asia cruises revealed regional characteristics of sources and sinks of these atmospherically important species, suggesting the existence of additional sources for methane, nitrous oxides, and carbon monoxide in this tropical Asian region.
NASA Astrophysics Data System (ADS)
Faes, Luca; Marinazzo, Daniele; Stramaglia, Sebastiano; Jurysta, Fabrice; Porta, Alberto; Giandomenico, Nollo
2016-05-01
This work introduces a framework to study the network formed by the autonomic component of heart rate variability (cardiac process η) and the amplitude of the different electroencephalographic waves (brain processes δ, θ, α, σ, β) during sleep. The framework exploits multivariate linear models to decompose the predictability of any given target process into measures of self-, causal and interaction predictability reflecting respectively the information retained in the process and related to its physiological complexity, the information transferred from the other source processes, and the information modified during the transfer according to redundant or synergistic interaction between the sources. The framework is here applied to the η, δ, θ, α, σ, β time series measured from the sleep recordings of eight severe sleep apnoea-hypopnoea syndrome (SAHS) patients studied before and after long-term treatment with continuous positive airway pressure (CPAP) therapy, and 14 healthy controls. Results show that the full and self-predictability of η, δ and θ decreased significantly in SAHS compared with controls, and were restored with CPAP for δ and θ but not for η. The causal predictability of η and δ occurred through significantly redundant source interaction during healthy sleep, which was lost in SAHS and recovered after CPAP. These results indicate that predictability analysis is a viable tool to assess the modifications of complexity and causality of the cerebral and cardiac processes induced by sleep disorders, and to monitor the restoration of the neuroautonomic control of these processes during long-term treatment.
Observation of seafloor crustal movement using the seafloor acoustic ranging on Kumano-nada
NASA Astrophysics Data System (ADS)
Osada, Y.; Kido, M.; Fujimoto, H.
2010-12-01
Along the Nankai Trough, where the Philippine Sea plate subducts under southeastern Japan with a convergence rate of about 65 mm/yr, large interplate thrust earthquakes of magnitude 8 class have occurred repeatedly with recurrence intervals of 100-200 years. About 60 years have passed since the last earthquakes happened in 1944 and 1946. Therefore it is important to monitor the tectonic activities in the Nankai Trough. Since most of the source region of the earthquakes is located beneath the ocean, an observation system is necessary in the offshore source region. We developed a seafloor acoustic ranging system to continuously monitor the seafloor crustal movement. We aim to monitor the activity in the splay faults in the rupture area of the Tonankai earthquake in the Nankai subduction zone. Slips along the active splay faults may be an important mechanism that the elastic strain caused by relative plate motion. We carried out two experiments, a short-term (one day) and a long-term (four month) experiments, to estimate the repeatability of acoustic measurements of this system. We deployed four PXPs (precision acoustic transponders) with about 600 m (M2-S1 baseline) and 920 m (M2-S2 base line) spacing in the long-term experiment. The standard deviation in acoustic measurements was about 1 cm on each baseline. In September 2008 we carried out an observation to monitor an active splay faults on Kumano-Nada prism slope. We deployed three PXPs with about 925 m (M1-S2 baseline) and 725 m (M1-S2 base line) spacing at the depth of some 2880 m. We recovered them in August 2010 to get data of acoustic measurements for 6 month and pressure measurements for 18 month. The round trip travel time shows a variation with peak-to-peak amplitude of about 1msec. We preliminarily collected the time series of round trip travel times using sound speed, which was estimated from measured temperature and pressure, and attitude data. We discuss the result of a variation of distance.
NASA Astrophysics Data System (ADS)
Russell, L. M.; Leaitch, W. R.; Liu, J.; Desiree, T. S.; Huang, L.; Sharma, S.; Chivulescu, A.; Veber, D.; Zhang, W.
2016-12-01
Long-term measurements of submicron aerosol particle chemical composition and size distributions are essential for evaluating whether global climate models correctly transport particles from lower latitudes to polar regions, especially in the winter months when satellite retrieval of aerosol properties is limited. In collaboration with ongoing measurements by the Dr. Neil Trivett Global Atmospheric Watch observatory at Alert, Nunavut (82.5°N; elevation 185 m-ASL), we measured the organic functional group composition of submicron aerosol particles sampled from the 10-m inlet from April 2012 to October 2014. The sampling site is approximately 10 km from the Alert station, and vehicle traffic is restricted except when filter sampling is stopped, making the impact of local emissions on submicron particle mass concentrations small. The organic functional group (OFG) composition is measured by Fourier Transform Infrared spectroscopy of samples collected on pre-loaded Teflon filters and stored and shipped frozen to La Jolla, California, for analysis. Samples were collected weekly to complement the twice hourly online measurements of non-refractory organic and inorganic composition by an Aerodyne ACSM. Organic components are shown to contribute a substantial fraction of the measured aerosol submicron mass year round. These measurements illustrate the seasonal contributions to the aerosol size distribution from OFG and illustrate the potential sources of the OFG at this remote site. The three largest OFG sources are transported fossil fuel combustion emissions from lower latitudes, sea spray and other marine particles, and episodic contributions from wildfires, volcanoes, and other high-latitude events. These sources are similar to those identified from earlier OFG measurements at Barrow, Alaska, and during the ICEALOT cruise in the Arctic Ocean.
Nonlinear derating of high-intensity focused ultrasound beams using Gaussian modal sums.
Dibaji, Seyed Ahmad Reza; Banerjee, Rupak K; Soneson, Joshua E; Myers, Matthew R
2013-11-01
A method is introduced for using measurements made in water of the nonlinear acoustic pressure field produced by a high-intensity focused ultrasound transducer to compute the acoustic pressure and temperature rise in a tissue medium. The acoustic pressure harmonics generated by nonlinear propagation are represented as a sum of modes having a Gaussian functional dependence in the radial direction. While the method is derived in the context of Gaussian beams, final results are applicable to general transducer profiles. The focal acoustic pressure is obtained by solving an evolution equation in the axial variable. The nonlinear term in the evolution equation for tissue is modeled using modal amplitudes measured in water and suitably reduced using a combination of "source derating" (experiments in water performed at a lower source acoustic pressure than in tissue) and "endpoint derating" (amplitudes reduced at the target location). Numerical experiments showed that, with proper combinations of source derating and endpoint derating, direct simulations of acoustic pressure and temperature in tissue could be reproduced by derating within 5% error. Advantages of the derating approach presented include applicability over a wide range of gains, ease of computation (a single numerical quadrature is required), and readily obtained temperature estimates from the water measurements.
Mercury evasion from a boreal peatland shortens the timeline for recovery from legacy pollution.
Osterwalder, Stefan; Bishop, Kevin; Alewell, Christine; Fritsche, Johannes; Laudon, Hjalmar; Åkerblom, Staffan; Nilsson, Mats B
2017-11-22
Peatlands are a major source of methylmercury that contaminates downstream aquatic food webs. The large store of mercury (Hg) in peatlands could be a source of Hg for over a century even if deposition is dramatically reduced. However, the reliability of Hg mass balances can be questioned due to missing long-term land-atmosphere flux measurements. We used a novel micrometeorological system for continuous measurement of Hg peatland-atmosphere exchange to derive the first annual Hg budget for a peatland. The evasion of Hg (9.4 µg m -2 yr -1 ) over the course of a year was seven times greater than stream Hg export, and over two times greater than wet bulk deposition to the boreal peatland. Measurements of dissolved gaseous Hg in the peat pore water also indicate Hg evasion. The net efflux may result from recent declines in atmospheric Hg concentrations that have turned the peatland from a net sink into a source of atmospheric Hg. This net Hg loss suggests that open boreal peatlands and downstream ecosystems can recover more rapidly from past atmospheric Hg deposition than previously assumed. This has important implications for future levels of methylmercury in boreal freshwater fish and the estimation of historical Hg accumulation rates from peat profiles.
Bosire Onyancha, Omwoyo
2008-05-01
As channels of communicating HIV/AIDS research information, serial publications and particularly journals are increasingly used in response to the pandemic. The last few decades have witnessed a proliferation of sources of HIV/AIDS-related information, bringing many challenges to collection-development librarians as well as to researchers. This study uses an informetric approach to examine the growth, productivity and scientific impact of these sources, during the period 1980 to 2005, and especially to measure performance in the publication and dissemination of HIV/AIDS research about or from eastern or southern Africa. Data were collected from MEDLINE, Science Citation Index (SCI), Social Sciences Citation Index (SSCI), and Ulrich's Periodical Directory. The analysis used Sitkis version 1.5, Microsoft Office Access, Microsoft Office Excel, Bibexcel, and Citespace version 2.0.1. The specific objectives were to identify the number of sources of HIV/AIDS-related information that have been published in the region, the coverage of these in key bibliographic databases, the most commonly used publication type for HIV/AIDS research, the countries in which the sources are published, the sources' productivity in terms of numbers of papers and citations, the most influential sources, the subject coverage of the sources, and the core sources of HIV/AIDS-information.
Efficient Development of High Fidelity Structured Volume Grids for Hypersonic Flow Simulations
NASA Technical Reports Server (NTRS)
Alter, Stephen J.
2003-01-01
A new technique for the control of grid line spacing and intersection angles of a structured volume grid, using elliptic partial differential equations (PDEs) is presented. Existing structured grid generation algorithms make use of source term hybridization to provide control of grid lines, imposing orthogonality implicitly at the boundary and explicitly on the interior of the domain. A bridging function between the two types of grid line control is typically used to blend the different orthogonality formulations. It is shown that utilizing such a bridging function with source term hybridization can result in the excessive use of computational resources and diminishes robustness. A new approach, Anisotropic Lagrange Based Trans-Finite Interpolation (ALBTFI), is offered as a replacement to source term hybridization. The ALBTFI technique captures the essence of the desired grid controls while improving the convergence rate of the elliptic PDEs when compared with source term hybridization. Grid generation on a blunt cone and a Shuttle Orbiter is used to demonstrate and assess the ALBTFI technique, which is shown to be as much as 50% faster, more robust, and produces higher quality grids than source term hybridization.
BWR ASSEMBLY SOURCE TERMS FOR WASTE PACKAGE DESIGN
DOE Office of Scientific and Technical Information (OSTI.GOV)
T.L. Lotz
1997-02-15
This analysis is prepared by the Mined Geologic Disposal System (MGDS) Waste Package Development Department (WPDD) to provide boiling water reactor (BWR) assembly radiation source term data for use during Waste Package (WP) design. The BWR assembly radiation source terms are to be used for evaluation of radiolysis effects at the WP surface, and for personnel shielding requirements during assembly or WP handling operations. The objectives of this evaluation are to generate BWR assembly radiation source terms that bound selected groupings of BWR assemblies, with regard to assembly average burnup and cooling time, which comprise the anticipated MGDS BWR commercialmore » spent nuclear fuel (SNF) waste stream. The source term data is to be provided in a form which can easily be utilized in subsequent shielding/radiation dose calculations. Since these calculations may also be used for Total System Performance Assessment (TSPA), with appropriate justification provided by TSPA, or radionuclide release rate analysis, the grams of each element and additional cooling times out to 25 years will also be calculated and the data included in the output files.« less
Optical performance of prototype horn-coupled TES bolometer arrays for SAFARI
NASA Astrophysics Data System (ADS)
Audley, Michael D.; de Lange, Gert; Gao, Jian-Rong; Khosropanah, Pourya; Hijmering, Richard; Ridder, Marcel L.
2016-07-01
The SAFARI Detector Test Facility is an ultra-low background optical testbed for characterizing ultra-sensitive prototype horn-coupled TES bolmeters for SAFARI, the grating spectrometer on board the proposed SPICA satellite. The testbed contains internal cold and hot black-body illuminators and a light-pipe for illumination with an external source. We have added reimaging optics to facilitate array optical measurements. The system is now being used for optical testing of prototype detector arrays read out with frequency-domain multiplexing. We present our latest optical measurements of prototype arrays and discuss these in terms of the instrument performance.
The effects of shared information on semantic calculations in the gene ontology.
Bible, Paul W; Sun, Hong-Wei; Morasso, Maria I; Loganantharaj, Rasiah; Wei, Lai
2017-01-01
The structured vocabulary that describes gene function, the gene ontology (GO), serves as a powerful tool in biological research. One application of GO in computational biology calculates semantic similarity between two concepts to make inferences about the functional similarity of genes. A class of term similarity algorithms explicitly calculates the shared information (SI) between concepts then substitutes this calculation into traditional term similarity measures such as Resnik, Lin, and Jiang-Conrath. Alternative SI approaches, when combined with ontology choice and term similarity type, lead to many gene-to-gene similarity measures. No thorough investigation has been made into the behavior, complexity, and performance of semantic methods derived from distinct SI approaches. We apply bootstrapping to compare the generalized performance of 57 gene-to-gene semantic measures across six benchmarks. Considering the number of measures, we additionally evaluate whether these methods can be leveraged through ensemble machine learning to improve prediction performance. Results showed that the choice of ontology type most strongly influenced performance across all evaluations. Combining measures into an ensemble classifier reduces cross-validation error beyond any individual measure for protein interaction prediction. This improvement resulted from information gained through the combination of ontology types as ensemble methods within each GO type offered no improvement. These results demonstrate that multiple SI measures can be leveraged for machine learning tasks such as automated gene function prediction by incorporating methods from across the ontologies. To facilitate future research in this area, we developed the GO Graph Tool Kit (GGTK), an open source C++ library with Python interface (github.com/paulbible/ggtk).
NASA Astrophysics Data System (ADS)
Giudici, Mauro; Casabianca, Davide; Comunian, Alessandro
2015-04-01
The basic classical inverse problem of groundwater hydrology aims at determining aquifer transmissivity (T ) from measurements of hydraulic head (h), estimates or measures of source terms and with the least possible knowledge on hydraulic transmissivity. The theory of inverse problems shows that this is an example of ill-posed problem, for which non-uniqueness and instability (or at least ill-conditioning) might preclude the computation of a physically acceptable solution. One of the methods to reduce the problems with non-uniqueness, ill-conditioning and instability is a tomographic approach, i.e., the use of data corresponding to independent flow situations. The latter might correspond to different hydraulic stimulations of the aquifer, i.e., to different pumping schedules and flux rates. Three inverse methods have been analyzed and tested to profit from the use of multiple sets of data: the Differential System Method (DSM), the Comparison Model Method (CMM) and the Double Constraint Method (DCM). DSM and CMM need h all over the domain and thus the first step for their application is the interpolation of measurements of h at sparse points. Moreover, they also need the knowledge of the source terms (aquifer recharge, well pumping rates) all over the aquifer. DSM is intrinsically based on the use of multiple data sets, which permit to write a first-order partial differential equation for T , whereas CMM and DCM were originally proposed to invert a single data set and have been extended to work with multiple data sets in this work. CMM and DCM are based on Darcy's law, which is used to update an initial guess of the T field with formulas based on a comparison of different hydraulic gradients. In particular, the CMM algorithm corrects the T estimate with ratio of the observed hydraulic gradient and that obtained with a comparison model which shares the same boundary conditions and source terms as the model to be calibrated, but a tentative T field. On the other hand the DCM algorithm applies the ratio of the hydraulic gradients obtained for two different forward models, one with the same boundary conditions and source terms as the model to be calibrated and the other one with prescribed head at the positions where in- or out-flow is known and h is measured. For DCM and CMM, multiple stimulation is used by updating the T field separately for each data set and then combining the resulting updated fields with different possible statistics (arithmetic, geometric or harmonic mean, median, least change, etc.). The three algorithms are tested and their characteristics and results are compared with a field data set, which was provided by prof. Fritz Stauffer (ETH) and corresponding to a pumping test in a thin alluvial aquifer in northern Switzerland. Three data sets are available and correspond to the undisturbed state, to the flow field created by a single pumping well and to the situation created by an 'hydraulic dipole', i.e., an extraction and an injection wells. These data sets permit to test the three inverse methods and the different options which can be chosen for their use.
Nuclear Spin Dependent Parity Violation in Diatomic Molecules
NASA Astrophysics Data System (ADS)
Altuntas, Emine; Cahn, Sidney; Demille, David; Kozlov, Mikhail
2016-05-01
Nuclear spin-dependent parity violation (NSD-PV) effects arise from exchange of the Z0 boson between electrons and the nucleus, and from interaction of electrons with the nuclear anapole moment, a parity-odd magnetic moment. The latter scales with nucleon number of the nucleus A as A 2 / 3 , whereas the Z0 coupling is independent of A. Thus the former is the dominant source of NSD-PV for nuclei with A >= 20. We study NSD-PV effects using diatomic molecules, where signals are dramatically amplified by bringing rotational levels of opposite parity close to degeneracy in a strong magnetic field. The NSD-PV interaction matrix element is measured using a Stark-interference technique. We present results that demonstrate statistical sensitivity to NSD-PV effects surpassing that of any previous atomic parity violation measurement, using the test system 138 Ba19 F. We report our progress on measuring and cancelling systematic effects due to combination of non-reversing stray E-fields, Enr with B-field inhomogeneities. Short-term prospects for measuring the nuclear anapole moment of 137 Ba19 F are discussed. In the long term, our technique is sufficiently general and sensitive to enable measurements across a broad range of nuclei.
Nuclear Spin Dependent Parity Violation in Diatomic Molecules
NASA Astrophysics Data System (ADS)
Altuntas, Emine; Cahn, Sidney; Demille, David
2016-09-01
Nuclear spin-dependent parity violation (NSD-PV) effects arise from exchange of the Z0 boson between electrons and the nucleus, and from interaction of electrons with the nuclear anapole moment, a parity-odd magnetic moment. The latter scales with nucleon number of the nucleus A as A2/3 , whereas the Z0 coupling is independent of A. Thus the former is the dominant source of NSD-PV for nuclei with A >= 20 . We study NSD-PV effects using diatomic molecules, where signals are dramatically amplified by bringing rotational levels of opposite parity close to degeneracy in a strong magnetic field. The NSD-PV interaction matrix element is measured using a Stark-interference technique. We present results that demonstrate statistical sensitivity to NSD-PV effects surpassing that of any previous atomic parity violation measurement, using the test system 138Ba19F. We report our progress on measuring and cancelling systematic effects due to combination of non-reversing stray E-fields, Enr with B-field inhomogeneities. Short-term prospects for measuring the nuclear anapole moment of 137Ba19F are discussed. In the long term, our technique is sufficiently general and sensitive to enable measurements across a broad range of nuclei.
NASA Astrophysics Data System (ADS)
Huang, Y. W.; Berman, E. S.; Owano, T. G.; Verfaillie, J. G.; Oikawa, P. Y.; Baldocchi, D. D.; Still, C. J.; Gardner, A.; Baer, D. S.; Rastogi, B.
2015-12-01
Stable CO2 isotopes provide information on biogeochemical processes that occur at the soil-plant-atmosphere interface. While δ13C measurement can provide information on the sources of the CO2, be it photosynthesis, natural gas combustion, other fossil fuel sources, landfills or other sources, δ18O, and δ17O are thought to be determined by the hydrological cycling of the CO2. Though researchers have called for analytical tools for CO2 isotope measurements that are reliable and field-deployable, developing such instrument remains a challenge. The carbon dioxide isotope analyzer developed by Los Gatos Research (LGR) uses LGR's patented Off-Axis ICOS (Integrated Cavity Output Spectroscopy) technology and incorporates proprietary internal thermal control for high sensitivity and optimal instrument stability. This new and improved analyzer measures CO2 concentration as well as δ13C, δ18O, and δ17O from CO2 at natural abundance (150-2500 ppm). The laboratory precision is ±200 ppb (1σ) in CO2 at 1 s, with a long-term (2 min) precision of ±20 ppb. The 1-second precision for both δ13C and δ18O is 0.7 ‰, and for δ17O is 1.8 ‰. The long-term (2 min) precision for both δ13C and δ18O is 0.08 ‰, and for δ17O is 0.18 ‰. The instrument has improved precision, stability and user interface over previous LGR CO2 isotope instruments and can be easily programmed for periodic referencing and sampling from different sources when coupled with LGR's multiport inlet unit (MIU). We have deployed two of these instruments at two different field sites, one at Twitchell Island in Sacramento County, CA to monitor the CO2 isotopic fluxes from an alfalfa field from 6/29/2015-7/13/2015, and the other at the Wind River Experimental Forest in Washington to monitor primarily the oxygen isotopes of CO2 within the canopy from 8/4/2015 through mid-November 2015. Methodology, laboratory development and testing and field performance are presented.
Technology integration performance assessment using lean principles in health care.
Rico, Florentino; Yalcin, Ali; Eikman, Edward A
2015-01-01
This study assesses the impact of an automated infusion system (AIS) integration at a positron emission tomography (PET) center based on "lean thinking" principles. The authors propose a systematic measurement system that evaluates improvement in terms of the "8 wastes." This adaptation to the health care context consisted of performance measurement before and after integration of AIS in terms of time, utilization of resources, amount of materials wasted/saved, system variability, distances traveled, and worker strain. The authors' observations indicate that AIS stands to be very effective in a busy PET department, such as the one in Moffitt Cancer Center, owing to its accuracy, pace, and reliability, especially after the necessary adjustments are made to reduce or eliminate the source of errors. This integration must be accompanied by a process reengineering exercise to realize the full potential of AIS in reducing waste and improving patient care and worker satisfaction. © The Author(s) 2014.
Accuracy of assessing the level of impulse sound from distant sources.
Wszołek, Tadeusz; Kłaczyński, Maciej
2007-01-01
Impulse sound events are characterised by ultra high pressures and low frequencies. Lower frequency sounds are generally less attenuated over a given distance in the atmosphere than higher frequencies. Thus, impulse sounds can be heard over greater distances and will be more affected by the environment. To calculate a long-term average immission level it is necessary to apply weighting factors like the probability of the occurrence of each weather condition during the relevant time period. This means that when measuring impulse noise at a long distance it is necessary to follow environmental parameters in many points along the way sound travels and also to have a database of sound transfer functions in the long term. The paper analyses the uncertainty of immission measurement results of impulse sound from cladding and destroying explosive materials. The influence of environmental conditions on the way sound travels is the focus of this paper.
Long and short term effects of plasma treatment on meristematic plant cells
NASA Astrophysics Data System (ADS)
Puač, N.; Živković, S.; Selaković, N.; Milutinović, M.; Boljević, J.; Malović, G.; Petrović, Z. Lj.
2014-05-01
In this paper, we will present results of plasma treatments of meristematic cells of Daucus carota. Plasma needle was used as an atmospheric pressure/gas composition source of non-equilibrium plasma in all treatments. Activity of antioxidant enzymes superoxide dismutase and catalase was measured immediately after plasma treatment and after two weeks following the treatment. Superoxide dismutase activity was increased in samples immediately after the plasma treatment. On the other hand, catalase activity was much higher in treated samples when measured two weeks after plasma treatment. These results show that there is a direct proof of the triggering of signal transduction in the cells by two reactive oxygen species H2O2 and O2-, causing enzyme activity and short and long term effects even during the growth of calli, where the information is passed to newborn cells over the period of two weeks.
Long-term monitoring of the Sedlec Ossuary - Analysis of hygrothermal conditions
NASA Astrophysics Data System (ADS)
Pavlík, Zbyšek; Balík, Lukáš; Maděra, Jiří; Černý, Robert
2016-07-01
The Sedlec Ossuary is one of the twelve UNESCO World Heritage Sites in the Czech Republic. Although the ossuary is listed among the most visited Czech tourist attractions, its technical state is almost critical and a radical renovation is necessary. On this account, hygrothermal performance of the ossuary is experimentally researched in the presented paper in order to get information on moisture sources and to get necessary data for optimized design of renovation treatments and reconstruction solutions that will allow preserve the historical significance of this attractive heritage site. Within the performed experimental analysis, the interior and exterior climatic conditions are monitored over an almost three year period together with relative humidity and temperature profiles measured in the most damage parts of the ossuary chapel. On the basis of measured data, the long-term hygrothermal state of the ossuary building is accessed and the periods of possible surface condensation are identified.
NASA Astrophysics Data System (ADS)
Zeb, Naila; Fahim Khokhar, Muhammad; Khan, Saud Ahmed; Noreen, Asma; Murtaza, Rabbia
2017-04-01
Air pollution is the expected key environmental issue of Pakistan as it is ranked among top polluted countries in the region. Ongoing rapid economic growth without any adequate measures is leading to worst air quality over time. The study aims to monitor long term atmospheric composition and association of trace gases over Pakistan. Tropospheric concentrations of CO, TOC, NO2 and HCHO derived from multiple satellite instruments are used for study from year 2005 to 2014. The study will provide first database for tropospheric trace gases over Pakistan. Spatio-temporal assessment identified hotspots and possible sources of trace gases over the Pakistan. High concentrations of trace gases are mainly observed over Punjab region, which may be attributed to its metropolitan importance. It is the major agricultural, industrialized and urbanized (nearly 60 % of the Pakistan's population) sector of the country. The expected sources are the agricultural fires, biomass/fossil fuel burning for heating purposes, urbanization, industrialization and meteorological variations. Seasonal variability is observed to explore seasonal patterns over the decade. Well defined seasonal cycles of trace gases are observed over the whole study period. The observed seasonal patterns also showed some noteworthy association among trace gases, which is further explored by different statistical tests. Seasonal Mann Kendall test is applied to test the significance of trend in series whereas correlation is carried out to measure the strength of association among trace gases. Strong correlation is observed for trace gases especially between CO and TOC. Partial Mann Kendall test is used to ideally identify the impact of each covariate on long term trend of CO and TOC by partialling out each correlating trace gas (covariate). It is observed that TOC, NO2 and HCHO has significant impact on long term trend of CO whereas, TOC critically depends on NO2 concentrations for long term increase over the region. Furthermore to explore causal relation, regression analysis is employed to estimate model for CO and TOC. This model numerically estimated the long term association of trace gases over the region.
Normal and anomalous diffusion in fluctuations of dust concentration nearby emission source
NASA Astrophysics Data System (ADS)
Szczurek, Andrzej; Maciejewska, Monika; Wyłomańska, Agnieszka; Sikora, Grzegorz; Balcerek, Michał; Teuerle, Marek
2018-02-01
Particulate matter (PM) is an important component of air. Nowadays, major attention is payed to fine dust. It has considerable environmental impact, including adverse effect on human health. One of important issues regarding PM is the temporal variation of its concentration. The variation contains information about factors influencing this quantity in time. The work focuses on the character of PM concentration dynamics indoors, in the vicinity of emission source. The objective was to recognize between the homogeneous or heterogeneous dynamics. The goal was achieved by detecting normal and anomalous diffusion in fluctuations of PM concentration. For this purpose we used anomalous diffusion exponent, β which was derived from Mean Square Displacement (MSD) analysis. The information about PM concentration dynamics may be used to design sampling strategy, which serves to attain representative information about PM behavior in time. The data analyzed in this work was collected from single-point PM concentration monitoring in the vicinity of seven emission sources in industrial environment. In majority of cases we observed heterogeneous character of PM concentration dynamics. It confirms the complexity of interactions between the emission sources and indoor environment. This result also votes against simplistic approach to PM concentration measurement indoors, namely their occasional character, short measurement periods and long term averaging.
Fingerprinting of bed sediment in the Tay Estuary, Scotland: an environmental magnetism approach
NASA Astrophysics Data System (ADS)
Jenkins, Pierre A.; Duck, Rob W.; Rowan, John S.; Walden, John
Sediment fingerprinting is commonly used for sediment provenance studies in lakes, rivers and reservoirs and on hillslopes and floodplains. This investigation explores the mixing of terrestrial and marine-derived sediment in the Tay Estuary, Scotland, using mineral magnetic attributes for fingerprinting. Samples representative of the estuary sediments and of four sources (end-members) were subjected to a suite of magnetic susceptibility and remanence measurements. Sediment samples from the beds of the Rivers Tay and Earn represented fluvial inputs while samples from the Angus and Fife coasts represented marine input. Multivariate discriminant and factor analysis showed that the sources could be separated on the basis of six magnetic parameters in a simple multivariate unmixing model to identify source contributions to estuarine bed sediments. Multi-domain magnetite signatures, characteristic of unweathered bedrock, dominate the magnetic measurements. Overall contributions of 3% from the River Earn, 17% from the River Tay, 29% from the Angus coast and 51% from the Fife coast source end-members, demonstrated the present-day regime of marine sediment derivation in the Tay Estuary. However, this conceals considerable spatial variability both along-estuary and in terms of sub-environments, with small-scale variations in sediment provenance reflecting local morphology, particularly areas of channel convergence.
Bauer, Timothy J
2013-06-15
The Jack Rabbit Test Program was sponsored in April and May 2010 by the Department of Homeland Security Transportation Security Administration to generate source data for large releases of chlorine and ammonia from transport tanks. In addition to a variety of data types measured at the release location, concentration versus time data was measured using sensors at distances up to 500 m from the tank. Release data were used to create accurate representations of the vapor flux versus time for the ten releases. This study was conducted to determine the importance of source terms and meteorological conditions in predicting downwind concentrations and the accuracy that can be obtained in those predictions. Each source representation was entered into an atmospheric transport and dispersion model using simplifying assumptions regarding the source characterization and meteorological conditions, and statistics for cloud duration and concentration at the sensor locations were calculated. A detailed characterization for one of the chlorine releases predicted 37% of concentration values within a factor of two, but cannot be considered representative of all the trials. Predictions of toxic effects at 200 m are relevant to incidents involving 1-ton chlorine tanks commonly used in parts of the United States and internationally. Published by Elsevier B.V.
NASA Astrophysics Data System (ADS)
Čufar, Aljaž; Batistoni, Paola; Conroy, Sean; Ghani, Zamir; Lengar, Igor; Milocco, Alberto; Packer, Lee; Pillon, Mario; Popovichev, Sergey; Snoj, Luka; JET Contributors
2017-03-01
At the Joint European Torus (JET) the ex-vessel fission chambers and in-vessel activation detectors are used as the neutron production rate and neutron yield monitors respectively. In order to ensure that these detectors produce accurate measurements they need to be experimentally calibrated. A new calibration of neutron detectors to 14 MeV neutrons, resulting from deuterium-tritium (DT) plasmas, is planned at JET using a compact accelerator based neutron generator (NG) in which a D/T beam impinges on a solid target containing T/D, producing neutrons by DT fusion reactions. This paper presents the analysis that was performed to model the neutron source characteristics in terms of energy spectrum, angle-energy distribution and the effect of the neutron generator geometry. Different codes capable of simulating the accelerator based DT neutron sources are compared and sensitivities to uncertainties in the generator's internal structure analysed. The analysis was performed to support preparation to the experimental measurements performed to characterize the NG as a calibration source. Further extensive neutronics analyses, performed with this model of the NG, will be needed to support the neutron calibration experiments and take into account various differences between the calibration experiment and experiments using the plasma as a source of neutrons.
Search for sterile neutrinos in the neutrino-4 experiment
NASA Astrophysics Data System (ADS)
Serebrov, A. P.; Ivochkin, V. G.; Samoilov, R. M.; Fomin, A. K.; Polyushkin, A. O.; Zinov'ev, V. G.; Neustroev, P. V.; Golovtsov, V. L.; Chernyi, A. V.; Zherebtsov, O. M.; Martem'yanov, V. P.; Tarasenkov, V. G.; Aleshin, V. I.; Petelin, A. L.; Izhutov, A. L.; Tuzov, A. A.; Sazontov, S. A.; Ryazanov, D. K.; Gromov, M. O.; Afanas'ev, V. V.; Zaitsev, M. E.; Chaikovskii, M. E.
2017-03-01
An experimental search for sterile neutrinos has been carried out at a neutrino facility based on the SM-3 nuclear reactor in Dimitrovgrad, Russia. The movable detector with passive shielding against the external radiation may be positioned at a distance varying between 6 and 12 m from the center of the reactor. The antineutrino flux has for the first time been measured using a movable detector placed close to the antineutrino source. The accuracy of the measurements is largely restricted by the cosmic background. The results of the measurements performed at small and large distances are analyzed in terms of the sterile-neutrino model parameters Δ m 14 2 and sin22θ14.
Excitation of Earth Rotation Variations "Observed" by Time-Variable Gravity
NASA Technical Reports Server (NTRS)
Chao, Ben F.; Cox, C. M.
2005-01-01
Time variable gravity measurements have been made over the past two decades using the space geodetic technique of satellite laser ranging, and more recently by the GRACE satellite mission with improved spatial resolutions. The degree-2 harmonic components of the time-variable gravity contain important information about the Earth s length-of-day and polar motion excitation functions, in a way independent to the traditional "direct" Earth rotation measurements made by, for example, the very-long-baseline interferometry and GPS. In particular, the (degree=2, order= 1) components give the mass term of the polar motion excitation; the (2,O) component, under certain mass conservation conditions, gives the mass term of the length-of-day excitation. Combining these with yet another independent source of angular momentum estimation calculated from global geophysical fluid models (for example the atmospheric angular momentum, in both mass and motion terms), in principle can lead to new insights into the dynamics, particularly the role or the lack thereof of the cores, in the excitation processes of the Earth rotation variations.
Time-frequency approach to underdetermined blind source separation.
Xie, Shengli; Yang, Liu; Yang, Jun-Mei; Zhou, Guoxu; Xiang, Yong
2012-02-01
This paper presents a new time-frequency (TF) underdetermined blind source separation approach based on Wigner-Ville distribution (WVD) and Khatri-Rao product to separate N non-stationary sources from M(M <; N) mixtures. First, an improved method is proposed for estimating the mixing matrix, where the negative value of the auto WVD of the sources is fully considered. Then after extracting all the auto-term TF points, the auto WVD value of the sources at every auto-term TF point can be found out exactly with the proposed approach no matter how many active sources there are as long as N ≤ 2M-1. Further discussion about the extraction of auto-term TF points is made and finally the numerical simulation results are presented to show the superiority of the proposed algorithm by comparing it with the existing ones.
NASA Astrophysics Data System (ADS)
Fawole, Olusegun G.; Cai, Xiaoming; Levine, James G.; Pinker, Rachel T.; MacKenzie, A. R.
2016-12-01
The West African region, with its peculiar climate and atmospheric dynamics, is a prominent source of aerosols. Reliable and long-term in situ measurements of aerosol properties are not readily available across the region. In this study, Version 2 Level 1.5 Aerosol Robotic Network (AERONET) data were used to study the absorption and size distribution properties of aerosols from dominant sources identified by trajectory analysis. The trajectory analysis was used to define four sources of aerosols over a 10 year period. Sorting the AERONET aerosol retrievals by these putative sources, the hypothesis that there exists an optically distinct gas flaring signal was tested. Dominance of each source cluster varies with season: desert-dust (DD) and biomass burning (BB) aerosols are dominant in months prior to the West African Monsoon (WAM); urban (UB) and gas flaring (GF) aerosol are dominant during the WAM months. BB aerosol, with single scattering albedo (SSA) at 675 nm value of 0.86 ± 0.03 and GF aerosol with SSA (675 nm) value of 0.9 ± 0.07, is the most absorbing of the aerosol categories. The range of Absorption Angstr&öm Exponent (AAE) for DD, BB, UB and GF classes are 1.99 ± 0.35, 1.45 ± 0.26, 1.21 ± 0.38 and 0.98 ± 0.25, respectively, indicating different aerosol composition for each source. The AAE (440-870 nm) and Angstr&öm Exponent (AE) (440-870 nm) relationships further show the spread and overlap of the variation of these optical and microphysical properties, presumably due in part to similarity in the sources of aerosols and in part, due to mixing of air parcels from different sources en route to the measurement site.
Long Range Transport was a Bigger NSS Source than DMS in the Remote Tropical MBL during PASE
NASA Astrophysics Data System (ADS)
Huebert, B. J.; Simpson, R. M.; Howell, S. G.; Blomquist, B.
2013-12-01
DMS was not the principal source of non-sea salt sulfate (NSS) mass in the remote marine boundary layer during the Pacific Atmospheric Sulfur Experiment (PASE), according to an Eulerian sulfur budget model based on chemical concentrations measured from the NCAR C-130 in the tropical Pacific. Each of our three (DMS, SO2, and NSS) self-consistent monthly- average budgets includes terms for surface exchange, entrainment, divergence, chemical formation, and chemical loss. The budget-derived DMS emission was (2.7 × 0.5 μmol m-2 d-1, our budget 'units'). SO2 sources include DMS + OH (1.4 × 0.4 units, assuming γ = 0.75) and entrainment from the free troposphere (FT) (0.8 × 0.2 units). Clouds were the most important chemical reactors for SO2 (-1.0 × 0.5 units). SO2 loss terms also include divergence (-0.9 × 0.3 units), dry deposition (-0.5 × 0.2 units), and OH + SO2 (-0.22 × 0.05 units). The total SO2 loss balanced the SO2 source. We found negligible NSS on particles from 2.6 μm to 10 μm diameter, the sea salt mass peak. Fine-particle NSS sources include in-cloud oxidation of SO2 by H2O2 (1.0 × 0.5 units), OH + SO2 (0.19 × 0.05 units), and entrainment (1.1 × 0.3 units in clean conditions; twice that when continental pollution is present). Only about 1/4 of emitted DMS becomes NSS. The NSS sources from entrainment and from DMS are similar in magnitude.
Laurent, Olivier; Hu, Jianlin; Li, Lianfa; Cockburn, Myles; Escobedo, Loraine; Kleeman, Michael J; Wu, Jun
2014-10-01
Low birth weight (LBW, <2500 g) has been associated with exposure to air pollution, but it is still unclear which sources or components of air pollution might be in play. The association between ultrafine particles and LBW has never been studied. To study the relationships between LBW in term born infants and exposure to particles by size fraction, source and chemical composition, and complementary components of air pollution in Los Angeles County (California, USA) over the period 2001-2008. Birth certificates (n=960,945) were geocoded to maternal residence. Primary particulate matter (PM) concentrations by source and composition were modeled. Measured fine PM, nitrogen dioxide and ozone concentrations were interpolated using empirical Bayesian kriging. Traffic indices were estimated. Associations between LBW and air pollution metrics were examined using generalized additive models, adjusting for maternal age, parity, race/ethnicity, education, neighborhood income, gestational age and infant sex. Increased LBW risks were associated with the mass of primary fine and ultrafine PM, with several major sources (especially gasoline, wood burning and commercial meat cooking) of primary PM, and chemical species in primary PM (elemental and organic carbon, potassium, iron, chromium, nickel, and titanium but not lead or arsenic). Increased LBW risks were also associated with total fine PM mass, nitrogen dioxide and local traffic indices (especially within 50 m from home), but not with ozone. Stronger associations were observed in infants born to women with low socioeconomic status, chronic hypertension, diabetes and a high body mass index. This study supports previously reported associations between traffic-related pollutants and LBW and suggests other pollution sources and components, including ultrafine particles, as possible risk factors. Copyright © 2014 Elsevier Inc. All rights reserved.