Endalamaw, Abraham; Bolton, W. Robert; Young-Robertson, Jessica M.; ...
2017-09-14
Modeling hydrological processes in the Alaskan sub-arctic is challenging because of the extreme spatial heterogeneity in soil properties and vegetation communities. Nevertheless, modeling and predicting hydrological processes is critical in this region due to its vulnerability to the effects of climate change. Coarse-spatial-resolution datasets used in land surface modeling pose a new challenge in simulating the spatially distributed and basin-integrated processes since these datasets do not adequately represent the small-scale hydrological, thermal, and ecological heterogeneity. The goal of this study is to improve the prediction capacity of mesoscale to large-scale hydrological models by introducing a small-scale parameterization scheme, which bettermore » represents the spatial heterogeneity of soil properties and vegetation cover in the Alaskan sub-arctic. The small-scale parameterization schemes are derived from observations and a sub-grid parameterization method in the two contrasting sub-basins of the Caribou Poker Creek Research Watershed (CPCRW) in Interior Alaska: one nearly permafrost-free (LowP) sub-basin and one permafrost-dominated (HighP) sub-basin. The sub-grid parameterization method used in the small-scale parameterization scheme is derived from the watershed topography. We found that observed soil thermal and hydraulic properties – including the distribution of permafrost and vegetation cover heterogeneity – are better represented in the sub-grid parameterization method than the coarse-resolution datasets. Parameters derived from the coarse-resolution datasets and from the sub-grid parameterization method are implemented into the variable infiltration capacity (VIC) mesoscale hydrological model to simulate runoff, evapotranspiration (ET), and soil moisture in the two sub-basins of the CPCRW. Simulated hydrographs based on the small-scale parameterization capture most of the peak and low flows, with similar accuracy in both sub-basins, compared to simulated hydrographs based on the coarse-resolution datasets. On average, the small-scale parameterization scheme improves the total runoff simulation by up to 50 % in the LowP sub-basin and by up to 10 % in the HighP sub-basin from the large-scale parameterization. This study shows that the proposed sub-grid parameterization method can be used to improve the performance of mesoscale hydrological models in the Alaskan sub-arctic watersheds.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Endalamaw, Abraham; Bolton, W. Robert; Young-Robertson, Jessica M.
Modeling hydrological processes in the Alaskan sub-arctic is challenging because of the extreme spatial heterogeneity in soil properties and vegetation communities. Nevertheless, modeling and predicting hydrological processes is critical in this region due to its vulnerability to the effects of climate change. Coarse-spatial-resolution datasets used in land surface modeling pose a new challenge in simulating the spatially distributed and basin-integrated processes since these datasets do not adequately represent the small-scale hydrological, thermal, and ecological heterogeneity. The goal of this study is to improve the prediction capacity of mesoscale to large-scale hydrological models by introducing a small-scale parameterization scheme, which bettermore » represents the spatial heterogeneity of soil properties and vegetation cover in the Alaskan sub-arctic. The small-scale parameterization schemes are derived from observations and a sub-grid parameterization method in the two contrasting sub-basins of the Caribou Poker Creek Research Watershed (CPCRW) in Interior Alaska: one nearly permafrost-free (LowP) sub-basin and one permafrost-dominated (HighP) sub-basin. The sub-grid parameterization method used in the small-scale parameterization scheme is derived from the watershed topography. We found that observed soil thermal and hydraulic properties – including the distribution of permafrost and vegetation cover heterogeneity – are better represented in the sub-grid parameterization method than the coarse-resolution datasets. Parameters derived from the coarse-resolution datasets and from the sub-grid parameterization method are implemented into the variable infiltration capacity (VIC) mesoscale hydrological model to simulate runoff, evapotranspiration (ET), and soil moisture in the two sub-basins of the CPCRW. Simulated hydrographs based on the small-scale parameterization capture most of the peak and low flows, with similar accuracy in both sub-basins, compared to simulated hydrographs based on the coarse-resolution datasets. On average, the small-scale parameterization scheme improves the total runoff simulation by up to 50 % in the LowP sub-basin and by up to 10 % in the HighP sub-basin from the large-scale parameterization. This study shows that the proposed sub-grid parameterization method can be used to improve the performance of mesoscale hydrological models in the Alaskan sub-arctic watersheds.« less
NASA Astrophysics Data System (ADS)
Pathiraja, S. D.; van Leeuwen, P. J.
2017-12-01
Model Uncertainty Quantification remains one of the central challenges of effective Data Assimilation (DA) in complex partially observed non-linear systems. Stochastic parameterization methods have been proposed in recent years as a means of capturing the uncertainty associated with unresolved sub-grid scale processes. Such approaches generally require some knowledge of the true sub-grid scale process or rely on full observations of the larger scale resolved process. We present a methodology for estimating the statistics of sub-grid scale processes using only partial observations of the resolved process. It finds model error realisations over a training period by minimizing their conditional variance, constrained by available observations. Special is that these realisations are binned conditioned on the previous model state during the minimization process, allowing for the recovery of complex error structures. The efficacy of the approach is demonstrated through numerical experiments on the multi-scale Lorenz 96' model. We consider different parameterizations of the model with both small and large time scale separations between slow and fast variables. Results are compared to two existing methods for accounting for model uncertainty in DA and shown to provide improved analyses and forecasts.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gustafson, William I.; Qian, Yun; Fast, Jerome D.
2011-07-13
Recent improvements to many global climate models include detailed, prognostic aerosol calculations intended to better reproduce the observed climate. However, the trace gas and aerosol fields are treated at the grid-cell scale with no attempt to account for sub-grid impacts on the aerosol fields. This paper begins to quantify the error introduced by the neglected sub-grid variability for the shortwave aerosol radiative forcing for a representative climate model grid spacing of 75 km. An analysis of the value added in downscaling aerosol fields is also presented to give context to the WRF-Chem simulations used for the sub-grid analysis. We foundmore » that 1) the impact of neglected sub-grid variability on the aerosol radiative forcing is strongest in regions of complex topography and complicated flow patterns, and 2) scale-induced differences in emissions contribute strongly to the impact of neglected sub-grid processes on the aerosol radiative forcing. The two of these effects together, when simulated at 75 km vs. 3 km in WRF-Chem, result in an average daytime mean bias of over 30% error in top-of-atmosphere shortwave aerosol radiative forcing for a large percentage of central Mexico during the MILAGRO field campaign.« less
Improvements in sub-grid, microphysics averages using quadrature based approaches
NASA Astrophysics Data System (ADS)
Chowdhary, K.; Debusschere, B.; Larson, V. E.
2013-12-01
Sub-grid variability in microphysical processes plays a critical role in atmospheric climate models. In order to account for this sub-grid variability, Larson and Schanen (2013) propose placing a probability density function on the sub-grid cloud microphysics quantities, e.g. autoconversion rate, essentially interpreting the cloud microphysics quantities as a random variable in each grid box. Random sampling techniques, e.g. Monte Carlo and Latin Hypercube, can be used to calculate statistics, e.g. averages, on the microphysics quantities, which then feed back into the model dynamics on the coarse scale. We propose an alternate approach using numerical quadrature methods based on deterministic sampling points to compute the statistical moments of microphysics quantities in each grid box. We have performed a preliminary test on the Kessler autoconversion formula, and, upon comparison with Latin Hypercube sampling, our approach shows an increased level of accuracy with a reduction in sample size by almost two orders of magnitude. Application to other microphysics processes is the subject of ongoing research.
NASA Astrophysics Data System (ADS)
Chen, Peng; Guo, Shilong; Li, Yanchao; Zhang, Yutao
2017-03-01
In this paper, an experimental and numerical investigation of premixed methane/air flame dynamics in a closed combustion vessel with a thin obstacle is described. In the experiment, high-speed video photography and a pressure transducer are used to study the flame shape changes and pressure dynamics. In the numerical simulation, four sub-grid scale viscosity models and three sub-grid scale combustion models are evaluated for their individual prediction compared with the experimental data. High-speed photographs show that the flame propagation process can be divided into five stages: spherical flame, finger-shaped flame, jet flame, mushroom-shaped flame and bidirectional propagation flame. Compared with the other sub-grid scale viscosity models and sub-grid scale combustion models, the dynamic Smagorinsky-Lilly model and the power-law flame wrinkling model are better able to predict the flame behaviour, respectively. Thus, coupling the dynamic Smagorinsky-Lilly model and the power-law flame wrinkling model, the numerical results demonstrate that flame shape change is a purely hydrodynamic phenomenon, and the mushroom-shaped flame and bidirectional propagation flame are the result of flame-vortex interaction. In addition, the transition from "corrugated flamelets" to "thin reaction zones" is observed in the simulation.
NASA Astrophysics Data System (ADS)
Montzka, Carsten; Herbst, Michael; Weihermüller, Lutz; Verhoef, Anne; Vereecken, Harry
2017-07-01
Agroecosystem models, regional and global climate models, and numerical weather prediction models require adequate parameterization of soil hydraulic properties. These properties are fundamental for describing and predicting water and energy exchange processes at the transition zone between solid earth and atmosphere, and regulate evapotranspiration, infiltration and runoff generation. Hydraulic parameters describing the soil water retention (WRC) and hydraulic conductivity (HCC) curves are typically derived from soil texture via pedotransfer functions (PTFs). Resampling of those parameters for specific model grids is typically performed by different aggregation approaches such a spatial averaging and the use of dominant textural properties or soil classes. These aggregation approaches introduce uncertainty, bias and parameter inconsistencies throughout spatial scales due to nonlinear relationships between hydraulic parameters and soil texture. Therefore, we present a method to scale hydraulic parameters to individual model grids and provide a global data set that overcomes the mentioned problems. The approach is based on Miller-Miller scaling in the relaxed form by Warrick, that fits the parameters of the WRC through all sub-grid WRCs to provide an effective parameterization for the grid cell at model resolution; at the same time it preserves the information of sub-grid variability of the water retention curve by deriving local scaling parameters. Based on the Mualem-van Genuchten approach we also derive the unsaturated hydraulic conductivity from the water retention functions, thereby assuming that the local parameters are also valid for this function. In addition, via the Warrick scaling parameter λ, information on global sub-grid scaling variance is given that enables modellers to improve dynamical downscaling of (regional) climate models or to perturb hydraulic parameters for model ensemble output generation. The present analysis is based on the ROSETTA PTF of Schaap et al. (2001) applied to the SoilGrids1km data set of Hengl et al. (2014). The example data set is provided at a global resolution of 0.25° at https://doi.org/10.1594/PANGAEA.870605.
Validation of Land-Surface Mosaic Heterogeneity in the GEOS DAS
NASA Technical Reports Server (NTRS)
Bosilovich, Michael G.; Molod, Andrea; Houser, Paul R.; Schubert, Siegfried
1999-01-01
The Mosaic Land-surface Model (LSM) has been included into the current GEOS Data Assimilation System (DAS). The LSM uses a more advanced representation of physical processes than previous versions of the GEOS DAS, including the representation of sub-grid heterogeneity of the land-surface through the Mosaic approach. As a first approximation, Mosaic assumes that all similar surface types within a grid-cell can be lumped together as a single'tile'. Within one GCM grid-cell, there might be 1 - 5 different tiles or surface types. All tiles are subjected to the grid-scale forcing (radiation, air temperature and specific humidity, and precipitation), and the sub-grid variability is a function of the tile characteristics. In this paper, we validate the LSM sub-grid scale variability (tiles) using a variety of surface observing stations from the Southern Great Plains (SGP) site of the Atmospheric Radiation Measurement (ARM) Program. One of the primary goals of SGP ARM is to study the variability of atmospheric radiation within a G,CM grid-cell. Enough surface data has been collected by ARM to extend this goal to sub-grid variability of the land-surface energy and water budgets. The time period of this study is the Summer of 1998 (June I - September 1). The ARM site data consists of surface meteorology, energy flux (eddy correlation and bowen ratio), soil water observations spread over an area similar to the size of a G-CM grid-cell. Various ARM stations are described as wheat and alfalfa crops, pasture and range land. The LSM tiles considered at the grid-space (2 x 2.5) nearest the ARM site include, grassland, deciduous forests, bare soil and dwarf trees. Surface energy and water balances for each tile type are compared with observations. Furthermore, we will discuss the land-surface sub-grid variability of both the ARM observations and the DAS.
Subgrid Modeling Geomorphological and Ecological Processes in Salt Marsh Evolution
NASA Astrophysics Data System (ADS)
Shi, F.; Kirby, J. T., Jr.; Wu, G.; Abdolali, A.; Deb, M.
2016-12-01
Numerical modeling a long-term evolution of salt marshes is challenging because it requires an extensive use of computational resources. Due to the presence of narrow tidal creeks, variations of salt marsh topography can be significant over spatial length scales on the order of a meter. With growing availability of high-resolution bathymetry measurements, like LiDAR-derived DEM data, it is increasingly desirable to run a high-resolution model in a large domain and for a long period of time to get trends of sedimentation patterns, morphological change and marsh evolution. However, high spatial-resolution poses a big challenge in both computational time and memory storage, when simulating a salt marsh with dimensions of up to O(100 km^2) with a small time step. In this study, we have developed a so-called Pre-storage, Sub-grid Model (PSM, Wu et al., 2015) for simulating flooding and draining processes in salt marshes. The simulation of Brokenbridge salt marsh, Delaware, shows that, with the combination of the sub-grid model and the pre-storage method, over 2 orders of magnitude computational speed-up can be achieved with minimal loss of model accuracy. We recently extended PSM to include a sediment transport component and models for biomass growth and sedimentation in the sub-grid model framework. The sediment transport model is formulated based on a newly derived sub-grid sediment concentration equation following Defina's (2000) area-averaging procedure. Suspended sediment transport is modeled by the advection-diffusion equation in the coarse grid level, but the local erosion and sedimentation rates are integrated over the sub-grid level. The morphological model is based on the existing morphological model in NearCoM (Shi et al., 2013), extended to include organic production from the biomass model. The vegetation biomass is predicted by a simple logistic equation model proposed by Marani et al. (2010). The biomass component is loosely coupled with hydrodynamic and sedimentation models owing to the different time scales of the physical and ecological processes. The coupled model is being applied to Delaware marsh evolution in response to rising sea level and changing sediment supplies.
ED(MF)n: Humidity-Convection Feedbacks in a Mass Flux Scheme Based on Resolved Size Densities
NASA Astrophysics Data System (ADS)
Neggers, R.
2014-12-01
Cumulus cloud populations remain at least partially unresolved in present-day numerical simulations of global weather and climate, and accordingly their impact on the larger-scale flow has to be represented through parameterization. Various methods have been developed over the years, ranging in complexity from the early bulk models relying on a single plume to more recent approaches that attempt to reconstruct the underlying probability density functions, such as statistical schemes and multiple plume approaches. Most of these "classic" methods capture key aspects of cumulus cloud populations, and have been successfully implemented in operational weather and climate models. However, the ever finer discretizations of operational circulation models, driven by advances in the computational efficiency of supercomputers, is creating new problems for existing sub-grid schemes. Ideally, a sub-grid scheme should automatically adapt its impact on the resolved scales to the dimension of the grid-box within which it is supposed to act. It can be argued that this is only possible when i) the scheme is aware of the range of scales of the processes it represents, and ii) it can distinguish between contributions as a function of size. How to conceptually represent this knowledge of scale in existing parameterization schemes remains an open question that is actively researched. This study considers a relatively new class of models for sub-grid transport in which ideas from the field of population dynamics are merged with the concept of multi plume modelling. More precisely, a multiple mass flux framework for moist convective transport is formulated in which the ensemble of plumes is created in "size-space". It is argued that thus resolving the underlying size-densities creates opportunities for introducing scale-awareness and scale-adaptivity in the scheme. The behavior of an implementation of this framework in the Eddy Diffusivity Mass Flux (EDMF) model, named ED(MF)n, is examined for a standard case of subtropical marine shallow cumulus. We ask if a system of multiple independently resolved plumes is able to automatically create the vertical profile of bulk (mass) flux at which the sub-grid scale transport balances the imposed larger-scale forcings in the cloud layer.
CO2 Flux Estimation Errors Associated with Moist Atmospheric Processes
NASA Technical Reports Server (NTRS)
Parazoo, N. C.; Denning, A. S.; Kawa, S. R.; Pawson, S.; Lokupitiya, R.
2012-01-01
Vertical transport by moist sub-grid scale processes such as deep convection is a well-known source of uncertainty in CO2 source/sink inversion. However, a dynamical link between vertical transport, satellite based retrievals of column mole fractions of CO2, and source/sink inversion has not yet been established. By using the same offline transport model with meteorological fields from slightly different data assimilation systems, we examine sensitivity of frontal CO2 transport and retrieved fluxes to different parameterizations of sub-grid vertical transport. We find that frontal transport feeds off background vertical CO2 gradients, which are modulated by sub-grid vertical transport. The implication for source/sink estimation is two-fold. First, CO2 variations contained in moist poleward moving air masses are systematically different from variations in dry equatorward moving air. Moist poleward transport is hidden from orbital sensors on satellites, causing a sampling bias, which leads directly to small but systematic flux retrieval errors in northern mid-latitudes. Second, differences in the representation of moist sub-grid vertical transport in GEOS-4 and GEOS-5 meteorological fields cause differences in vertical gradients of CO2, which leads to systematic differences in moist poleward and dry equatorward CO2 transport and therefore the fraction of CO2 variations hidden in moist air from satellites. As a result, sampling biases are amplified and regional scale flux errors enhanced, most notably in Europe (0.43+/-0.35 PgC /yr). These results, cast from the perspective of moist frontal transport processes, support previous arguments that the vertical gradient of CO2 is a major source of uncertainty in source/sink inversion.
Representation of sub-element scale variability in snow accumulation and ablation is increasingly recognized as important in distributed hydrologic modelling. Representing sub-grid scale variability may be accomplished through numerical integration of a nested grid or through a l...
Improving sub-grid scale accuracy of boundary features in regional finite-difference models
Panday, Sorab; Langevin, Christian D.
2012-01-01
As an alternative to grid refinement, the concept of a ghost node, which was developed for nested grid applications, has been extended towards improving sub-grid scale accuracy of flow to conduits, wells, rivers or other boundary features that interact with a finite-difference groundwater flow model. The formulation is presented for correcting the regular finite-difference groundwater flow equations for confined and unconfined cases, with or without Newton Raphson linearization of the nonlinearities, to include the Ghost Node Correction (GNC) for location displacement. The correction may be applied on the right-hand side vector for a symmetric finite-difference Picard implementation, or on the left-hand side matrix for an implicit but asymmetric implementation. The finite-difference matrix connectivity structure may be maintained for an implicit implementation by only selecting contributing nodes that are a part of the finite-difference connectivity. Proof of concept example problems are provided to demonstrate the improved accuracy that may be achieved through sub-grid scale corrections using the GNC schemes.
Decker, Jeremy D.; Hughes, J.D.
2013-01-01
Climate change and sea-level rise could cause substantial changes in urban runoff and flooding in low-lying coast landscapes. A major challenge for local government officials and decision makers is to translate the potential global effects of climate change into actionable and cost-effective adaptation and mitigation strategies at county and municipal scales. A MODFLOW process is used to represent sub-grid scale hydrology in urban settings to help address these issues. Coupled interception, surface water, depression, and unsaturated zone storage are represented. A two-dimensional diffusive wave approximation is used to represent overland flow. Three different options for representing infiltration and recharge are presented. Additional features include structure, barrier, and culvert flow between adjacent cells, specified stage boundaries, critical flow boundaries, source/sink surface-water terms, and the bi-directional runoff to MODFLOW Surface-Water Routing process. Some abilities of the Urban RunOff (URO) process are demonstrated with a synthetic problem using four land uses and varying cell coverages. Precipitation from a hypothetical storm was applied and cell by cell surface-water depth, groundwater level, infiltration rate, and groundwater recharge rate are shown. Results indicate the URO process has the ability to produce time-varying, water-content dependent infiltration and leakage, and successfully interacts with MODFLOW.
NASA Astrophysics Data System (ADS)
Nijzink, R. C.; Samaniego, L.; Mai, J.; Kumar, R.; Thober, S.; Zink, M.; Schäfer, D.; Savenije, H. H. G.; Hrachowitz, M.
2015-12-01
Heterogeneity of landscape features like terrain, soil, and vegetation properties affect the partitioning of water and energy. However, it remains unclear to which extent an explicit representation of this heterogeneity at the sub-grid scale of distributed hydrological models can improve the hydrological consistency and the robustness of such models. In this study, hydrological process complexity arising from sub-grid topography heterogeneity was incorporated in the distributed mesoscale Hydrologic Model (mHM). Seven study catchments across Europe were used to test whether (1) the incorporation of additional sub-grid variability on the basis of landscape-derived response units improves model internal dynamics, (2) the application of semi-quantitative, expert-knowledge based model constraints reduces model uncertainty; and (3) the combined use of sub-grid response units and model constraints improves the spatial transferability of the model. Unconstrained and constrained versions of both, the original mHM and mHMtopo, which allows for topography-based sub-grid heterogeneity, were calibrated for each catchment individually following a multi-objective calibration strategy. In addition, four of the study catchments were simultaneously calibrated and their feasible parameter sets were transferred to the remaining three receiver catchments. In a post-calibration evaluation procedure the probabilities of model and transferability improvement, when accounting for sub-grid variability and/or applying expert-knowledge based model constraints, were assessed on the basis of a set of hydrological signatures. In terms of the Euclidian distance to the optimal model, used as overall measure for model performance with respect to the individual signatures, the model improvement achieved by introducing sub-grid heterogeneity to mHM in mHMtopo was on average 13 %. The addition of semi-quantitative constraints to mHM and mHMtopo resulted in improvements of 13 and 19 % respectively, compared to the base case of the unconstrained mHM. Most significant improvements in signature representations were, in particular, achieved for low flow statistics. The application of prior semi-quantitative constraints further improved the partitioning between runoff and evaporative fluxes. Besides, it was shown that suitable semi-quantitative prior constraints in combination with the transfer function based regularization approach of mHM, can be beneficial for spatial model transferability as the Euclidian distances for the signatures improved on average by 2 %. The effect of semi-quantitative prior constraints combined with topography-guided sub-grid heterogeneity on transferability showed a more variable picture of improvements and deteriorations, but most improvements were observed for low flow statistics.
NASA Astrophysics Data System (ADS)
Nijzink, Remko C.; Samaniego, Luis; Mai, Juliane; Kumar, Rohini; Thober, Stephan; Zink, Matthias; Schäfer, David; Savenije, Hubert H. G.; Hrachowitz, Markus
2016-03-01
Heterogeneity of landscape features like terrain, soil, and vegetation properties affects the partitioning of water and energy. However, it remains unclear to what extent an explicit representation of this heterogeneity at the sub-grid scale of distributed hydrological models can improve the hydrological consistency and the robustness of such models. In this study, hydrological process complexity arising from sub-grid topography heterogeneity was incorporated into the distributed mesoscale Hydrologic Model (mHM). Seven study catchments across Europe were used to test whether (1) the incorporation of additional sub-grid variability on the basis of landscape-derived response units improves model internal dynamics, (2) the application of semi-quantitative, expert-knowledge-based model constraints reduces model uncertainty, and whether (3) the combined use of sub-grid response units and model constraints improves the spatial transferability of the model. Unconstrained and constrained versions of both the original mHM and mHMtopo, which allows for topography-based sub-grid heterogeneity, were calibrated for each catchment individually following a multi-objective calibration strategy. In addition, four of the study catchments were simultaneously calibrated and their feasible parameter sets were transferred to the remaining three receiver catchments. In a post-calibration evaluation procedure the probabilities of model and transferability improvement, when accounting for sub-grid variability and/or applying expert-knowledge-based model constraints, were assessed on the basis of a set of hydrological signatures. In terms of the Euclidian distance to the optimal model, used as an overall measure of model performance with respect to the individual signatures, the model improvement achieved by introducing sub-grid heterogeneity to mHM in mHMtopo was on average 13 %. The addition of semi-quantitative constraints to mHM and mHMtopo resulted in improvements of 13 and 19 %, respectively, compared to the base case of the unconstrained mHM. Most significant improvements in signature representations were, in particular, achieved for low flow statistics. The application of prior semi-quantitative constraints further improved the partitioning between runoff and evaporative fluxes. In addition, it was shown that suitable semi-quantitative prior constraints in combination with the transfer-function-based regularization approach of mHM can be beneficial for spatial model transferability as the Euclidian distances for the signatures improved on average by 2 %. The effect of semi-quantitative prior constraints combined with topography-guided sub-grid heterogeneity on transferability showed a more variable picture of improvements and deteriorations, but most improvements were observed for low flow statistics.
NASA Astrophysics Data System (ADS)
Gruber, S.; Fiddes, J.
2013-12-01
In mountainous topography, the difference in scale between atmospheric reanalyses (typically tens of kilometres) and relevant processes and phenomena near the Earth surface, such as permafrost or snow cover (meters to tens of meters) is most obvious. This contrast of scales is one of the major obstacles to using reanalysis data for the simulation of surface phenomena and to confronting reanalyses with independent observation. At the example of modelling permafrost in mountain areas (but simple to generalise to other phenomena and heterogeneous environments), we present and test methods against measurements for (A) scaling atmospheric data from the reanalysis to the ground level and (B) smart sampling of the heterogeneous landscape in order to set up a lumped model simulation that represents the high-resolution land surface. TopoSCALE (Part A, see http://dx.doi.org/10.5194/gmdd-6-3381-2013) is a scheme, which scales coarse-grid climate fields to fine-grid topography using pressure level data. In addition, it applies necessary topographic corrections e.g. those variables required for computation of radiation fields. This provides the necessary driving fields to the LSM. Tested against independent ground data, this scheme has been shown to improve the scaling and distribution of meteorological parameters in complex terrain, as compared to conventional methods, e.g. lapse rate based approaches. TopoSUB (Part B, see http://dx.doi.org/10.5194/gmd-5-1245-2012) is a surface pre-processor designed to sample a fine-grid domain (defined by a digital elevation model) along important topographical (or other) dimensions through a clustering scheme. This allows constructing a lumped model representing the main sources of fine-grid variability and applying a 1D LSM efficiently over large areas. Results can processed to derive (i) summary statistics at coarse-scale re-analysis grid resolution, (ii) high-resolution data fields spatialized to e.g., the fine-scale digital elevation model grid, or (iii) validation products for locations at which measurements exist, only. The ability of TopoSUB to approximate results simulated by a 2D distributed numerical LSM at a factor of ~10,000 less computations is demonstrated by comparison of 2D and lumped simulations. Successful application of the combined scheme in the European Alps is reported and based on its results, open issues for future research are outlined.
Connecting spatial and temporal scales of tropical precipitation in observations and the MetUM-GA6
NASA Astrophysics Data System (ADS)
Martin, Gill M.; Klingaman, Nicholas P.; Moise, Aurel F.
2017-01-01
This study analyses tropical rainfall variability (on a range of temporal and spatial scales) in a set of parallel Met Office Unified Model (MetUM) simulations at a range of horizontal resolutions, which are compared with two satellite-derived rainfall datasets. We focus on the shorter scales, i.e. from the native grid and time step of the model through sub-daily to seasonal, since previous studies have paid relatively little attention to sub-daily rainfall variability and how this feeds through to longer scales. We find that the behaviour of the deep convection parametrization in this model on the native grid and time step is largely independent of the grid-box size and time step length over which it operates. There is also little difference in the rainfall variability on larger/longer spatial/temporal scales. Tropical convection in the model on the native grid/time step is spatially and temporally intermittent, producing very large rainfall amounts interspersed with grid boxes/time steps of little or no rain. In contrast, switching off the deep convection parametrization, albeit at an unrealistic resolution for resolving tropical convection, results in very persistent (for limited periods), but very sporadic, rainfall. In both cases, spatial and temporal averaging smoothes out this intermittency. On the ˜ 100 km scale, for oceanic regions, the spectra of 3-hourly and daily mean rainfall in the configurations with parametrized convection agree fairly well with those from satellite-derived rainfall estimates, while at ˜ 10-day timescales the averages are overestimated, indicating a lack of intra-seasonal variability. Over tropical land the results are more varied, but the model often underestimates the daily mean rainfall (partly as a result of a poor diurnal cycle) but still lacks variability on intra-seasonal timescales. Ultimately, such work will shed light on how uncertainties in modelling small-/short-scale processes relate to uncertainty in climate change projections of rainfall distribution and variability, with a view to reducing such uncertainty through improved modelling of small-/short-scale processes.
Fast Grid Frequency Support from Distributed Inverter-Based Resources
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoke, Anderson F
This presentation summarizes power hardware-in-the-loop testing performed to evaluate the ability of distributed inverter-coupled generation to support grid frequency on the fastest time scales. The research found that distributed PV inverters and other DERs can effectively support the grid on sub-second time scales.
The general situation, (but exemplified in urban areas), where a significant degree of sub-grid variability (SGV) exists in grid models poses problems when comparing gridbased air quality modeling results with observations. Typically, grid models ignore or parameterize processes ...
M. M. Clark; T. H. Fletcher; R. R. Linn
2010-01-01
The chemical processes of gas phase combustion in wildland fires are complex and occur at length-scales that are not resolved in computational fluid dynamics (CFD) models of landscape-scale wildland fire. A new approach for modelling fire chemistry in HIGRAD/FIRETEC (a landscape-scale CFD wildfire model) applies a mixtureâ fraction model relying on thermodynamic...
Foust, Thomas D.; Ziegler, Jack L.; Pannala, Sreekanth; ...
2017-02-28
Here in this computational study, we model the mixing of biomass pyrolysis vapor with solid catalyst in circulating riser reactors with a focus on the determination of solid catalyst residence time distributions (RTDs). A comprehensive set of 2D and 3D simulations were conducted for a pilot-scale riser using the Eulerian-Eulerian two-fluid modeling framework with and without sub-grid-scale models for the gas-solids interaction. A validation test case was also simulated and compared to experiments, showing agreement in the pressure gradient and RTD mean and spread. For simulation cases, it was found that for accurate RTD prediction, the Johnson and Jackson partialmore » slip solids boundary condition was required for all models and a sub-grid model is useful so that ultra high resolutions grids that are very computationally intensive are not required. Finally, we discovered a 2/3 scaling relation for the RTD mean and spread when comparing resolved 2D simulations to validated unresolved 3D sub-grid-scale model simulations.« less
USING CMAQ FOR EXPOSURE MODELING AND CHARACTERIZING THE SUB-GRID VARIABILITY FOR EXPOSURE ESTIMATES
Atmospheric processes and the associated transport and dispersion of atmospheric pollutants are known to be highly variable in time and space. Current air quality models that characterize atmospheric chemistry effects, e.g. the Community Multi-scale Air Quality (CMAQ), provide vo...
NASA Astrophysics Data System (ADS)
Rouholahnejad, E.; Kirchner, J. W.
2016-12-01
Evapotranspiration (ET) is a key process in land-climate interactions and affects the dynamics of the atmosphere at local and regional scales. In estimating ET, most earth system models average over considerable sub-grid heterogeneity in land surface properties, precipitation (P), and potential evapotranspiration (PET). This spatial averaging could potentially bias ET estimates, due to the nonlinearities in the underlying relationships. In addition, most earth system models ignore lateral redistribution of water within and between grid cells, which could potentially alter both local and regional ET. Here we present a first attempt to quantify the effects of spatial heterogeneity and lateral redistribution on grid-cell-averaged ET as seen from the atmosphere over heterogeneous landscapes. Using a Budyko framework to express ET as a function of P and PET, we quantify how sub-grid heterogeneity affects average ET at the scale of typical earth system model grid cells. We show that averaging over sub-grid heterogeneity in P and PET, as typical earth system models do, leads to overestimates of average ET. We use a similar approach to quantify how lateral redistribution of water could affect average ET, as seen from the atmosphere. We show that where the aridity index P/PET increases with altitude, gravitationally driven lateral redistribution will increase average ET, implying that models that neglect lateral moisture redistribution will underestimate average ET. In contrast, where the aridity index P/PET decreases with altitude, gravitationally driven lateral redistribution will decrease average ET. This approach yields a simple conceptual framework and mathematical expressions for determining whether, and how much, spatial heterogeneity and lateral redistribution can affect regional ET fluxes as seen from the atmosphere. This analysis provides the basis for quantifying heterogeneity and redistribution effects on ET at regional and continental scales, which will be the focus of future work.
Simulating large-scale crop yield by using perturbed-parameter ensemble method
NASA Astrophysics Data System (ADS)
Iizumi, T.; Yokozawa, M.; Sakurai, G.; Nishimori, M.
2010-12-01
Toshichika Iizumi, Masayuki Yokozawa, Gen Sakurai, Motoki Nishimori Agro-Meteorology Division, National Institute for Agro-Environmental Sciences, Japan Abstract One of concerning issues of food security under changing climate is to predict the inter-annual variation of crop production induced by climate extremes and modulated climate. To secure food supply for growing world population, methodology that can accurately predict crop yield on a large scale is needed. However, for developing a process-based large-scale crop model with a scale of general circulation models (GCMs), 100 km in latitude and longitude, researchers encounter the difficulties in spatial heterogeneity of available information on crop production such as cultivated cultivars and management. This study proposed an ensemble-based simulation method that uses a process-based crop model and systematic parameter perturbation procedure, taking maize in U.S., China, and Brazil as examples. The crop model was developed modifying the fundamental structure of the Soil and Water Assessment Tool (SWAT) to incorporate the effect of heat stress on yield. We called the new model PRYSBI: the Process-based Regional-scale Yield Simulator with Bayesian Inference. The posterior probability density function (PDF) of 17 parameters, which represents the crop- and grid-specific features of the crop and its uncertainty under given data, was estimated by the Bayesian inversion analysis. We then take 1500 ensemble members of simulated yield values based on the parameter sets sampled from the posterior PDF to describe yearly changes of the yield, i.e. perturbed-parameter ensemble method. The ensemble median for 27 years (1980-2006) was compared with the data aggregated from the county yield. On a country scale, the ensemble median of the simulated yield showed a good correspondence with the reported yield: the Pearson’s correlation coefficient is over 0.6 for all countries. In contrast, on a grid scale, the correspondence is still high in most grids regardless of the countries. However, the model showed comparatively low reproducibility in the slope areas, such as around the Rocky Mountains in South Dakota, around the Great Xing'anling Mountains in Heilongjiang, and around the Brazilian Plateau. As there is a wide-ranging local climate conditions in the complex terrain, such as the slope of mountain, the GCM grid-scale weather inputs is likely one of major sources of error. The results of this study highlight the benefits of the perturbed-parameter ensemble method in simulating crop yield on a GCM grid scale: (1) the posterior PDF of parameter could quantify the uncertainty of parameter value of the crop model associated with the local crop production aspects; (2) the method can explicitly account for the uncertainty of parameter value in the crop model simulations; (3) the method achieve a Monte Carlo approximation of probability of sub-grid scale yield, accounting for the nonlinear response of crop yield to weather and management; (4) the method is therefore appropriate to aggregate the simulated sub-grid scale yields to a grid-scale yield and it may be a reason for high performance of the model in capturing inter-annual variation of yield.
Parameterization of Small-Scale Processes
1989-09-01
1989, Honolulu, Hawaii !7 COSATI CODES 18 SUBJECT TERMS (Continue on reverse if necessary and identify by block number) FELD GROUP SIJB- GROUP general...detailed sensitivit. studies to assess the dependence of results on the edd\\ viscosities and diffusivities by a direct comparison with certain observations...better sub-grid scale parameterization is to mount a concerted s .arch for model fits to observations. These would require exhaustive sensitivity studies
Shallow cumuli ensemble statistics for development of a stochastic parameterization
NASA Astrophysics Data System (ADS)
Sakradzija, Mirjana; Seifert, Axel; Heus, Thijs
2014-05-01
According to a conventional deterministic approach to the parameterization of moist convection in numerical atmospheric models, a given large scale forcing produces an unique response from the unresolved convective processes. This representation leaves out the small-scale variability of convection, as it is known from the empirical studies of deep and shallow convective cloud ensembles, there is a whole distribution of sub-grid states corresponding to the given large scale forcing. Moreover, this distribution gets broader with the increasing model resolution. This behavior is also consistent with our theoretical understanding of a coarse-grained nonlinear system. We propose an approach to represent the variability of the unresolved shallow-convective states, including the dependence of the sub-grid states distribution spread and shape on the model horizontal resolution. Starting from the Gibbs canonical ensemble theory, Craig and Cohen (2006) developed a theory for the fluctuations in a deep convective ensemble. The micro-states of a deep convective cloud ensemble are characterized by the cloud-base mass flux, which, according to the theory, is exponentially distributed (Boltzmann distribution). Following their work, we study the shallow cumulus ensemble statistics and the distribution of the cloud-base mass flux. We employ a Large-Eddy Simulation model (LES) and a cloud tracking algorithm, followed by a conditional sampling of clouds at the cloud base level, to retrieve the information about the individual cloud life cycles and the cloud ensemble as a whole. In the case of shallow cumulus cloud ensemble, the distribution of micro-states is a generalized exponential distribution. Based on the empirical and theoretical findings, a stochastic model has been developed to simulate the shallow convective cloud ensemble and to test the convective ensemble theory. Stochastic model simulates a compound random process, with the number of convective elements drawn from a Poisson distribution, and cloud properties sub-sampled from a generalized ensemble distribution. We study the role of the different cloud subtypes in a shallow convective ensemble and how the diverse cloud properties and cloud lifetimes affect the system macro-state. To what extent does the cloud-base mass flux distribution deviate from the simple Boltzmann distribution and how does it affect the results from the stochastic model? Is the memory, provided by the finite lifetime of individual clouds, of importance for the ensemble statistics? We also test for the minimal information given as an input to the stochastic model, able to reproduce the ensemble mean statistics and the variability in a convective ensemble. An important property of the resulting distribution of the sub-grid convective states is its scale-adaptivity - the smaller the grid-size, the broader the compound distribution of the sub-grid states.
NASA Technical Reports Server (NTRS)
Tao, Wei-Kuo; Moncrieff, Mitchell; Einaud, Franco (Technical Monitor)
2001-01-01
Numerical cloud models have been developed and applied extensively to study cloud-scale and mesoscale processes during the past four decades. The distinctive aspect of these cloud models is their ability to treat explicitly (or resolve) cloud-scale dynamics. This requires the cloud models to be formulated from the non-hydrostatic equations of motion that explicitly include the vertical acceleration terms since the vertical and horizontal scales of convection are similar. Such models are also necessary in order to allow gravity waves, such as those triggered by clouds, to be resolved explicitly. In contrast, the hydrostatic approximation, usually applied in global or regional models, does allow the presence of gravity waves. In addition, the availability of exponentially increasing computer capabilities has resulted in time integrations increasing from hours to days, domain grids boxes (points) increasing from less than 2000 to more than 2,500,000 grid points with 500 to 1000 m resolution, and 3-D models becoming increasingly prevalent. The cloud resolving model is now at a stage where it can provide reasonably accurate statistical information of the sub-grid, cloud-resolving processes poorly parameterized in climate models and numerical prediction models.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sarkar, Avik; Sun, Xin; Sundaresan, Sankaran
2014-04-23
The accuracy of coarse-grid multiphase CFD simulations of fluidized beds may be improved via the inclusion of filtered constitutive models. In our previous study (Sarkar et al., Chem. Eng. Sci., 104, 399-412), we developed such a set of filtered drag relationships for beds with immersed arrays of cooling tubes. Verification of these filtered drag models is addressed in this work. Predictions from coarse-grid simulations with the sub-grid filtered corrections are compared against accurate, highly-resolved simulations of full-scale turbulent and bubbling fluidized beds. The filtered drag models offer a computationally efficient yet accurate alternative for obtaining macroscopic predictions, but the spatialmore » resolution of meso-scale clustering heterogeneities is sacrificed.« less
Studies of Sub-Synchronous Oscillations in Large-Scale Wind Farm Integrated System
NASA Astrophysics Data System (ADS)
Yue, Liu; Hang, Mend
2018-01-01
With the rapid development and construction of large-scale wind farms and grid-connected operation, the series compensation wind power AC transmission is gradually becoming the main way of power usage and improvement of wind power availability and grid stability, but the integration of wind farm will change the SSO (Sub-Synchronous oscillation) damping characteristics of synchronous generator system. Regarding the above SSO problem caused by integration of large-scale wind farms, this paper focusing on doubly fed induction generator (DFIG) based wind farms, aim to summarize the SSO mechanism in large-scale wind power integrated system with series compensation, which can be classified as three types: sub-synchronous control interaction (SSCI), sub-synchronous torsional interaction (SSTI), sub-synchronous resonance (SSR). Then, SSO modelling and analysis methods are categorized and compared by its applicable areas. Furthermore, this paper summarizes the suppression measures of actual SSO projects based on different control objectives. Finally, the research prospect on this field is explored.
Are atmospheric updrafts a key to unlocking climate forcing and sensitivity?
Donner, Leo J.; O'Brien, Travis A.; Rieger, Daniel; ...
2016-10-20
Both climate forcing and climate sensitivity persist as stubborn uncertainties limiting the extent to which climate models can provide actionable scientific scenarios for climate change. A key, explicit control on cloud–aerosol interactions, the largest uncertainty in climate forcing, is the vertical velocity of cloud-scale updrafts. Model-based studies of climate sensitivity indicate that convective entrainment, which is closely related to updraft speeds, is an important control on climate sensitivity. Updraft vertical velocities also drive many physical processes essential to numerical weather prediction. Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climatemore » and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climate and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying vs in climate models may capture this behavior, but it has not been accounted for when parameterizing cloud and precipitation processes in current models. New observations of convective vertical velocities offer a potentially promising path toward developing process-level cloud models and parameterizations for climate and numerical weather prediction. Taking account of the scale dependence of resolved vertical velocities offers a path to matching cloud-scale physical processes and their driving dynamics more realistically, with a prospect of reduced uncertainty in both climate forcing and sensitivity.« less
Are atmospheric updrafts a key to unlocking climate forcing and sensitivity?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Donner, Leo J.; O'Brien, Travis A.; Rieger, Daniel
Both climate forcing and climate sensitivity persist as stubborn uncertainties limiting the extent to which climate models can provide actionable scientific scenarios for climate change. A key, explicit control on cloud–aerosol interactions, the largest uncertainty in climate forcing, is the vertical velocity of cloud-scale updrafts. Model-based studies of climate sensitivity indicate that convective entrainment, which is closely related to updraft speeds, is an important control on climate sensitivity. Updraft vertical velocities also drive many physical processes essential to numerical weather prediction. Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climatemore » and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climate and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying vs in climate models may capture this behavior, but it has not been accounted for when parameterizing cloud and precipitation processes in current models. New observations of convective vertical velocities offer a potentially promising path toward developing process-level cloud models and parameterizations for climate and numerical weather prediction. Taking account of the scale dependence of resolved vertical velocities offers a path to matching cloud-scale physical processes and their driving dynamics more realistically, with a prospect of reduced uncertainty in both climate forcing and sensitivity.« less
Evolution of aerosol downwind of a major highway
NASA Astrophysics Data System (ADS)
Liggio, J.; Staebler, R. M.; Brook, J.; Li, S.; Vlasenko, A. L.; Sjostedt, S. J.; Gordon, M.; Makar, P.; Mihele, C.; Evans, G. J.; Jeong, C.; Wentzell, J. J.; Lu, G.; Lee, P.
2010-12-01
Primary aerosol from traffic emissions can have a considerable impact local and regional scale air quality. In order to assess the effect of these emissions and of future emissions scenarios, air quality models are required which utilize emissions representative of real world conditions. Often, the emissions processing systems which provide emissions input for the air quality models rely on laboratory testing of individual vehicles under non-ambient conditions. However, on the sub-grid scale particle evolution may lead to changes in the primary emitted size distribution and gas-particle partitioning that are not properly considered when the emissions are ‘instantly mixed’ within the grid volume. The affect of this modeling convention on model results is not well understood. In particular, changes in organic gas/particle partitioning may result in particle evaporation or condensation onto pre-existing aerosol. The result is a change in the particle distribution and/or an increase in the organic mass available for subsequent gas-phase oxidation. These effects may be missing from air-quality models, and a careful analysis of field data is necessary to quantify their impact. A study of the sub-grid evolution of aerosols (FEVER; Fast Evolution of Vehicle Emissions from Roadways) was conducted in the Toronto area in the summer of 2010. The study included mobile measurements of particle size distributions with a Fast mobility particle sizer (FMPS), aerosol composition with an Aerodyne aerosol mass spectrometer (AMS), black carbon (SP2, PA, LII), VOCs (PTR-MS) and other trace gases. The mobile laboratory was used to measure the concentration gradient of the emissions at perpendicular distances from the highway as well as the physical and chemical evolution of the aerosol. Stationary sites at perpendicular distances and upwind from the highway also monitored the particle size distribution. In addition, sonic anemometers mounted on the mobile lab provided measurements of turbulent dispersion as a function of distance from the highway, and a traffic camera was used to determine traffic density, composition and speed. These measurements differ from previous studies in that turbulence is measured under realistic conditions and hence the relationship of the aerosol evolution to atmospheric stability and mixing will also be quantified. Preliminary results suggest that aerosol size and composition does change on the sub-grid scale, and sub-grid scale parameterizations of turbulence and particle chemistry should be included in models to accurately represent these effects.
The U.S. Environmental Protection Agency (U.S. EPA) is extending its Models-3/Community Multiscale Air Quality (CMAQ) Modeling System to provide detailed gridded air quality concentration fields and sub-grid variability characterization at neighborhood scales and in urban areas...
NASA Astrophysics Data System (ADS)
Suciu, L. G.; Griffin, R. J.; Masiello, C. A.
2017-12-01
Wildfires and prescribed burning are important sources of particulate and gaseous pyrogenic organic carbon (PyOC) emissions to the atmosphere. These emissions impact atmospheric chemistry, air quality and climate, but the spatial and temporal variabilities of these impacts are poorly understood, primarily because small and fresh fire plumes are not well predicted by three-dimensional Eulerian chemical transport models due to their coarser grid size. Generally, this results in underestimation of downwind deposition of PyOC, hydroxyl radical reactivity, secondary organic aerosol formation and ozone (O3) production. However, such models are very good for simulation of multiple atmospheric processes that could affect the lifetimes of PyOC emissions over large spatiotemporal scales. Finer resolution models, such as Lagrangian reactive plumes models (or plume-in-grid), could be used to trace fresh emissions at the sub-grid level of the Eulerian model. Moreover, Lagrangian plume models need background chemistry predicted by the Eulerian models to accurately simulate the interactions of the plume material with the background air during plume aging. Therefore, by coupling the two models, the physico-chemical evolution of the biomass burning plumes can be tracked from local to regional scales. In this study, we focus on the physico-chemical changes of PyOC emissions from sub-grid to grid levels using an existing chemical mechanism. We hypothesize that finer scale Lagrangian-Eulerian simulations of several prescribed burns in the U.S. will allow more accurate downwind predictions (validated by airborne observations from smoke plumes) of PyOC emissions (i.e., submicron particulate matter, organic aerosols, refractory black carbon) as well as O3 and other trace gases. Simulation results could be used to optimize the implementation of additional PyOC speciation in the existing chemical mechanism.
The goal of achieving verisimilitude of air quality simulations to observations is problematic. Chemical transport models such as the Community Multi-Scale Air Quality (CMAQ) modeling system produce volume averages of pollutant concentration fields. When grid sizes are such tha...
NASA Technical Reports Server (NTRS)
Avissar, Roni; Chen, Fei
1993-01-01
Generated by landscape discontinuities (e.g., sea breezes) mesoscale circulation processes are not represented in large-scale atmospheric models (e.g., general circulation models), which have an inappropiate grid-scale resolution. With the assumption that atmospheric variables can be separated into large scale, mesoscale, and turbulent scale, a set of prognostic equations applicable in large-scale atmospheric models for momentum, temperature, moisture, and any other gaseous or aerosol material, which includes both mesoscale and turbulent fluxes is developed. Prognostic equations are also developed for these mesoscale fluxes, which indicate a closure problem and, therefore, require a parameterization. For this purpose, the mean mesoscale kinetic energy (MKE) per unit of mass is used, defined as E-tilde = 0.5 (the mean value of u'(sub i exp 2), where u'(sub i) represents the three Cartesian components of a mesoscale circulation (the angle bracket symbol is the grid-scale, horizontal averaging operator in the large-scale model, and a tilde indicates a corresponding large-scale mean value). A prognostic equation is developed for E-tilde, and an analysis of the different terms of this equation indicates that the mesoscale vertical heat flux, the mesoscale pressure correlation, and the interaction between turbulence and mesoscale perturbations are the major terms that affect the time tendency of E-tilde. A-state-of-the-art mesoscale atmospheric model is used to investigate the relationship between MKE, landscape discontinuities (as characterized by the spatial distribution of heat fluxes at the earth's surface), and mesoscale sensible and latent heat fluxes in the atmosphere. MKE is compared with turbulence kinetic energy to illustrate the importance of mesoscale processes as compared to turbulent processes. This analysis emphasizes the potential use of MKE to bridge between landscape discontinuities and mesoscale fluxes and, therefore, to parameterize mesoscale fluxes generated by such subgrid-scale landscape discontinuities in large-scale atmospheric models.
NASA Astrophysics Data System (ADS)
Im, Eun-Soon; Coppola, Erika; Giorgi, Filippo
2010-05-01
Since anthropogenic climate change is a rather important factor for the future human life all over the planet and its effects are not globally uniform, climate information at regional or local scales become more and more important for an accurate assessment of the potential impact of climate change on societies and ecosystems. High resolution information with suitably fine-scale for resolving complex geographical features could be a critical factor for successful linkage between climate models and impact assessment studies. However, scale mismatch between them still remains major problem. One method for overcoming the resolution limitations of global climate models and for adding regional details to coarse-grid global projections is to use dynamical downscaling by means of a regional climate model. In this study, the ECHAM5/MPI-OM (1.875 degree) A1B scenario simulation has been dynamically downscaled by using two different approaches within the framework of RegCM3 modeling system. First, a mosaic-type parameterization of subgrid-scale topography and land use (Sub-BATS) is applied over the European Alpine region. The Sub-BATS system is composed of 15 km coarse-grid cell and 3 km sub-grid cell. Second, we developed the RegCM3 one-way double-nested system, with the mother domain encompassing the eastern regions of Asia at 60 km grid spacing and the nested domain covering the Korean Peninsula at 20 km grid spacing. By comparing the regional climate model output and the driving global model ECHAM5/MPI-OM output, it is possible to estimate the added value of physically-based dynamical downscaling when for example impact studies at hydrological scale are performed.
Sub-grid drag model for immersed vertical cylinders in fluidized beds
Verma, Vikrant; Li, Tingwen; Dietiker, Jean -Francois; ...
2017-01-03
Immersed vertical cylinders are often used as heat exchanger in gas-solid fluidized beds. Computational Fluid Dynamics (CFD) simulations are computationally expensive for large scale systems with bundles of cylinders. Therefore sub-grid models are required to facilitate simulations on a coarse grid, where internal cylinders are treated as a porous medium. The influence of cylinders on the gas-solid flow tends to enhance segregation and affect the gas-solid drag. A correction to gas-solid drag must be modeled using a suitable sub-grid constitutive relationship. In the past, Sarkar et al. have developed a sub-grid drag model for horizontal cylinder arrays based on 2Dmore » simulations. However, the effect of a vertical cylinder arrangement was not considered due to computational complexities. In this study, highly resolved 3D simulations with vertical cylinders were performed in small periodic domains. These simulations were filtered to construct a sub-grid drag model which can then be implemented in coarse-grid simulations. Gas-solid drag was filtered for different solids fractions and a significant reduction in drag was identified when compared with simulation without cylinders and simulation with horizontal cylinders. Slip velocities significantly increase when vertical cylinders are present. Lastly, vertical suspension drag due to vertical cylinders is insignificant however substantial horizontal suspension drag is observed which is consistent to the finding for horizontal cylinders.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Li; He, Ya-Ling; Kang, Qinjun
2013-12-15
A coupled (hybrid) simulation strategy spatially combining the finite volume method (FVM) and the lattice Boltzmann method (LBM), called CFVLBM, is developed to simulate coupled multi-scale multi-physicochemical processes. In the CFVLBM, computational domain of multi-scale problems is divided into two sub-domains, i.e., an open, free fluid region and a region filled with porous materials. The FVM and LBM are used for these two regions, respectively, with information exchanged at the interface between the two sub-domains. A general reconstruction operator (RO) is proposed to derive the distribution functions in the LBM from the corresponding macro scalar, the governing equation of whichmore » obeys the convection–diffusion equation. The CFVLBM and the RO are validated in several typical physicochemical problems and then are applied to simulate complex multi-scale coupled fluid flow, heat transfer, mass transport, and chemical reaction in a wall-coated micro reactor. The maximum ratio of the grid size between the FVM and LBM regions is explored and discussed. -- Highlights: •A coupled simulation strategy for simulating multi-scale phenomena is developed. •Finite volume method and lattice Boltzmann method are coupled. •A reconstruction operator is derived to transfer information at the sub-domains interface. •Coupled multi-scale multiple physicochemical processes in micro reactor are simulated. •Techniques to save computational resources and improve the efficiency are discussed.« less
NASA Astrophysics Data System (ADS)
Greer, A. T.; Woodson, C. B.
2016-02-01
Because of the complexity and extremely large size of marine ecosystems, research attention has a strong focus on modelling the system through space and time to elucidate processes driving ecosystem state. One of the major weaknesses of current modelling approaches is the reliance on a particular grid cell size (usually 10's of km in the horizontal & water column mean) to capture the relevant processes, even though empirical research has shown that marine systems are highly structured on fine scales, and this structure can persist over relatively long time scales (days to weeks). Fine-scale features can have a strong influence on the predator-prey interactions driving trophic transfer. Here we apply a statistic, the AB ratio, used to quantify increased predator production due to predator-prey overlap on fine scales in a manner that is computationally feasible for larger scale models. We calculated the AB ratio for predator-prey distributions throughout the scientific literature, as well as for data obtained with a towed plankton imaging system, demonstrating that averaging across a typical model grid cell neglects the fine-scale predator-prey overlap that is an essential component of ecosystem productivity. Organisms from a range of trophic levels and oceanographic regions tended to overlap with their prey both in the horizontal and vertical dimensions. When predator swimming over a diel cycle was incorporated, the amount of production indicated by the AB ratio increased substantially. For the plankton image data, the AB ratio was higher with increasing sampling resolution, especially when prey were highly aggregated. We recommend that ecosystem models incorporate more fine-scale information both to more accurately capture trophic transfer processes and to capitalize on the increasing sampling resolution and data volume from empirical studies.
NASA Astrophysics Data System (ADS)
Yue, Chao; Ciais, Philippe; Li, Wei
2018-02-01
Several modelling studies reported elevated carbon emissions from historical land use change (ELUC) by including bidirectional transitions on the sub-grid scale (termed gross land use change), dominated by shifting cultivation and other land turnover processes. However, most dynamic global vegetation models (DGVMs) that have implemented gross land use change either do not account for sub-grid secondary lands, or often have only one single secondary land tile over a model grid cell and thus cannot account for various rotation lengths in shifting cultivation and associated secondary forest age dynamics. Therefore, it remains uncertain how realistic the past ELUC estimations are and how estimated ELUC will differ between the two modelling approaches with and without multiple sub-grid secondary land cohorts - in particular secondary forest cohorts. Here we investigated historical ELUC over 1501-2005 by including sub-grid forest age dynamics in a DGVM. We run two simulations, one with no secondary forests (Sageless) and the other with sub-grid secondary forests of six age classes whose demography is driven by historical land use change (Sage). Estimated global ELUC for 1501-2005 is 176 Pg C in Sage compared to 197 Pg C in Sageless. The lower ELUC values in Sage arise mainly from shifting cultivation in the tropics under an assumed constant rotation length of 15 years, being 27 Pg C in Sage in contrast to 46 Pg C in Sageless. Estimated cumulative ELUC values from wood harvest in the Sage simulation (31 Pg C) are however slightly higher than Sageless (27 Pg C) when the model is forced by reconstructed harvested areas because secondary forests targeted in Sage for harvest priority are insufficient to meet the prescribed harvest area, leading to wood harvest being dominated by old primary forests. An alternative approach to quantify wood harvest ELUC, i.e. always harvesting the close-to-mature forests in both Sageless and Sage, yields similar values of 33 Pg C by both simulations. The lower ELUC from shifting cultivation in Sage simulations depends on the predefined forest clearing priority rules in the model and the assumed rotation length. A set of sensitivity model runs over Africa reveal that a longer rotation length over the historical period likely results in higher emissions. Our results highlight that although gross land use change as a former missing emission component is included by a growing number of DGVMs, its contribution to overall ELUC remains uncertain and tends to be overestimated when models ignore sub-grid secondary forests.
Using Unsupervised Learning to Unlock the Potential of Hydrologic Similarity
NASA Astrophysics Data System (ADS)
Chaney, N.; Newman, A. J.
2017-12-01
By clustering environmental data into representative hydrologic response units (HRUs), hydrologic similarity aims to harness the covariance between a system's physical environment and its hydrologic response to create reduced-order models. This is the primary approach through which sub-grid hydrologic processes are represented in large-scale models (e.g., Earth System Models). Although the possibilities of hydrologic similarity are extensive, its practical implementations have been limited to 1-d bins of oversimplistic metrics of hydrologic response (e.g., topographic index)—this is a missed opportunity. In this presentation we will show how unsupervised learning is unlocking the potential of hydrologic similarity; clustering methods enable generalized frameworks to effectively and efficiently harness the petabytes of global environmental data to robustly characterize sub-grid heterogeneity in large-scale models. To illustrate the potential that unsupervised learning has towards advancing hydrologic similarity, we introduce a hierarchical clustering algorithm (HCA) that clusters very high resolution (30-100 meters) elevation, soil, climate, and land cover data to assemble a domain's representative HRUs. These HRUs are then used to parameterize the sub-grid heterogeneity in land surface models; for this study we use the GFDL LM4 model—the land component of the GFDL Earth System Model. To explore HCA and its impacts on the hydrologic system we use a ¼ grid cell in southeastern California as a test site. HCA is used to construct an ensemble of 9 different HRU configurations—each configuration has a different number of HRUs; for each ensemble member LM4 is run between 2002 and 2014 with a 26 year spinup. The analysis of the ensemble of model simulations show that: 1) clustering the high-dimensional environmental data space leads to a robust representation of the role of the physical environment in the coupled water, energy, and carbon cycles at a relatively low number of HRUs; 2) the reduced-order model with around 300 HRUs effectively reproduces the fully distributed model simulation (30 meters) with less than 1/1000 of computational expense; 3) assigning each grid cell of the fully distributed grid to an HRU via HCA enables novel visualization methods for large-scale models—this has significant implications for how these models are applied and evaluated. We will conclude by outlining the potential that this work has within operational prediction systems including numerical weather prediction, Earth System models, and Early Warning systems.
THE RELATIONSHIP BETWEEN {nu}{sub max} AND AGE t FROM ZAMS TO RGB-TIP FOR LOW-MASS STARS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tang, Y. K.; Gai, N., E-mail: tyk450@163.com, E-mail: ning.gai@hotmail.com
2013-07-10
Stellar age is an important quantity in astrophysics, which is useful for many fields both in the universe and galaxies. It cannot be determined by direct measurements, but can only be estimated or inferred. We attempt to find a useful indicator of stellar age, which is accurate from the zero-age main sequence to the tip of red giant branch for low-mass stars. Using the Yale Rotation and Evolution Code (YREC), a grid of stellar models has been constructed. Meanwhile, the frequency of maximum oscillations' power {nu}{sub max} and the large frequency separation {Delta}{nu} are calculated using the scaling relations. Formore » the stars, the masses of which are from 0.8 M{sub Sun} to 2.8 M{sub Sun }, we can obtain the {nu}{sub max} and stellar age by combing the scaling relations with the four sets of grid models (YREC, Dotter et al., Marigo et al., and YY isochrones). We find that {nu}{sub max} is tightly correlated and decreases monotonically with the age of the star from the main sequence to the red giant evolutionary stages. Moreover, we find that the line shapes of the curves in the Age versus {nu}{sub max} diagram, which is plotted by the four sets of grid models, are consistent for red giants with masses from 1.1 M{sub Sun} to 2.8 M{sub Sun }. For red giants, the differences of correlation coefficients between Age and {nu}{sub max} for different grid models are minor and can be ignored. Interestingly, we find two peaks that correspond to the subgiants and bump of red giants in the Age versus {nu}{sub max} diagram. By general linear least-squares, we make the polynomial fitting and deduce the relationship between log(Age) and log({nu}{sub max}) in red giants' evolutionary state.« less
Methods and apparatus of analyzing electrical power grid data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hafen, Ryan P.; Critchlow, Terence J.; Gibson, Tara D.
Apparatus and methods of processing large-scale data regarding an electrical power grid are described. According to one aspect, a method of processing large-scale data regarding an electrical power grid includes accessing a large-scale data set comprising information regarding an electrical power grid; processing data of the large-scale data set to identify a filter which is configured to remove erroneous data from the large-scale data set; using the filter, removing erroneous data from the large-scale data set; and after the removing, processing data of the large-scale data set to identify an event detector which is configured to identify events of interestmore » in the large-scale data set.« less
NASA Astrophysics Data System (ADS)
Rouholahnejad, E.; Fan, Y.; Kirchner, J. W.; Miralles, D. G.
2017-12-01
Most Earth system models (ESM) average over considerable sub-grid heterogeneity in land surface properties, and overlook subsurface lateral flow. This could potentially bias evapotranspiration (ET) estimates and has implications for future temperature predictions, since overestimations in ET imply greater latent heat fluxes and potential underestimation of dry and warm conditions in the context of climate change. Here we quantify the bias in evaporation estimates that may arise from the fact that ESMs average over considerable heterogeneity in surface properties, and also neglect lateral transfer of water across the heterogeneous landscapes at global scale. We use a Budyko framework to express ET as a function of P and PET to derive simple sub-grid closure relations that quantify how spatial heterogeneity and lateral transfer could affect average ET as seen from the atmosphere. We show that averaging over sub-grid heterogeneity in P and PET, as typical Earth system models do, leads to overestimation of average ET. Our analysis at global scale shows that the effects of sub-grid heterogeneity will be most pronounced in steep mountainous areas where the topographic gradient is high and where P is inversely correlated with PET across the landscape. In addition, we use the Total Water Storage (TWS) anomaly estimates from the Gravity Recovery and Climate Experiment (GRACE) remote sensing product and assimilate it into the Global Land Evaporation Amsterdam Model (GLEAM) to correct for existing free drainage lower boundary condition in GLEAM and quantify whether, and how much, accounting for changes in terrestrial storage can improve the simulation of soil moisture and regional ET fluxes at global scale.
Sub-grid-scale description of turbulent magnetic reconnection in magnetohydrodynamics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Widmer, F., E-mail: widmer@mps.mpg.de; Institut für Astrophysik, Georg-August-Universität, Friedrich-Hund-Platz 1, 37077 Göttingen; Büchner, J.
Magnetic reconnection requires, at least locally, a non-ideal plasma response. In collisionless space and astrophysical plasmas, turbulence could transport energy from large to small scales where binary particle collisions are rare. We have investigated the influence of small scale magnetohydrodynamics (MHD) turbulence on the reconnection rate in the framework of a compressible MHD approach including sub-grid-scale (SGS) turbulence. For this sake, we considered Harris-type and force-free current sheets with finite guide magnetic fields directed out of the reconnection plane. The goal is to find out whether unresolved by conventional simulations MHD turbulence can enhance the reconnection process in high-Reynolds-number astrophysicalmore » plasmas. Together with the MHD equations, we solve evolution equations for the SGS energy and cross-helicity due to turbulence according to a Reynolds-averaged turbulence model. The SGS turbulence is self-generated and -sustained through the inhomogeneities of the mean fields. By this way, the feedback of the unresolved turbulence into the MHD reconnection process is taken into account. It is shown that the turbulence controls the regimes of reconnection by its characteristic timescale τ{sub t}. The dependence on resistivity was investigated for large-Reynolds-number plasmas for Harris-type as well as force-free current sheets with guide field. We found that magnetic reconnection depends on the relation between the molecular and apparent effective turbulent resistivity. We found that the turbulence timescale τ{sub t} decides whether fast reconnection takes place or whether the stored energy is just diffused away to small scale turbulence. If the amount of energy transferred from large to small scales is enhanced, fast reconnection can take place. Energy spectra allowed us to characterize the different regimes of reconnection. It was found that reconnection is even faster for larger Reynolds numbers controlled by the molecular resistivity η, as long as the initial level of turbulence is not too large. This implies that turbulence plays an important role to reach the limit of fast reconnection in large Reynolds number plasmas even for smaller amounts of turbulence.« less
Downscaling scheme to drive soil-vegetation-atmosphere transfer models
NASA Astrophysics Data System (ADS)
Schomburg, Annika; Venema, Victor; Lindau, Ralf; Ament, Felix; Simmer, Clemens
2010-05-01
The earth's surface is characterized by heterogeneity at a broad range of scales. Weather forecast models and climate models are not able to resolve this heterogeneity at the smaller scales. Many processes in the soil or at the surface, however, are highly nonlinear. This holds, for example, for evaporation processes, where stomata or aerodynamic resistances are nonlinear functions of the local micro-climate. Other examples are threshold dependent processes, e.g., the generation of runoff or the melting of snow. It has been shown that using averaged parameters in the computation of these processes leads to errors and especially biases, due to the involved nonlinearities. Thus it is necessary to account for the sub-grid scale surface heterogeneities in atmospheric modeling. One approach to take the variability of the earth's surface into account is the mosaic approach. Here the soil-vegetation-atmosphere transfer (SVAT) model is run on an explicit higher resolution than the atmospheric part of a coupled model, which is feasible due to generally lower computational costs of a SVAT model compared to the atmospheric part. The question arises how to deal with the scale differences at the interface between the two resolutions. Usually the assumption of a homogeneous forcing for all sub-pixels is made. However, over a heterogeneous surface, usually the boundary layer is also heterogeneous. Thus, by assuming a constant atmospheric forcing again biases in the turbulent heat fluxes may occur due to neglected atmospheric forcing variability. Therefore we have developed and tested a downscaling scheme to disaggregate the atmospheric variables of the lower atmosphere that are used as input to force a SVAT model. Our downscaling scheme consists of three steps: 1) a bi-quadratic spline interpolation of the coarse-resolution field; 2) a "deterministic" part, where relationships between surface and near-surface variables are exploited; and 3) a noise-generation step, in which the still missing, not explained, variance is added as noise. The scheme has been developed and tested based on high-resolution (400 m) model output of the weather forecast (and regional climate) COSMO model. Downscaling steps 1 and 2 reduce the error made by the homogeneous assumption considerably, whereas the third step leads to close agreement of the sub-grid scale variance with the reference. This is, however, achieved at the cost of higher root mean square errors. Thus, before applying the downscaling system to atmospheric data a decision should be made whether the lowest possible errors (apply only downscaling step 1 and 2) or a most realistic sub-grid scale variability (apply also step 3) is desired. This downscaling scheme is currently being implemented into the COSMO model, where it will be used in combination with the mosaic approach. However, this downscaling scheme can also be applied to drive stand-alone SVAT models or hydrological models, which usually also need high-resolution atmospheric forcing data.
Domain decomposition by the advancing-partition method for parallel unstructured grid generation
NASA Technical Reports Server (NTRS)
Banihashemi, legal representative, Soheila (Inventor); Pirzadeh, Shahyar Z. (Inventor)
2012-01-01
In a method for domain decomposition for generating unstructured grids, a surface mesh is generated for a spatial domain. A location of a partition plane dividing the domain into two sections is determined. Triangular faces on the surface mesh that intersect the partition plane are identified. A partition grid of tetrahedral cells, dividing the domain into two sub-domains, is generated using a marching process in which a front comprises only faces of new cells which intersect the partition plane. The partition grid is generated until no active faces remain on the front. Triangular faces on each side of the partition plane are collected into two separate subsets. Each subset of triangular faces is renumbered locally and a local/global mapping is created for each sub-domain. A volume grid is generated for each sub-domain. The partition grid and volume grids are then merged using the local-global mapping.
NASA Astrophysics Data System (ADS)
Qu, Yue; Slootsky, Michael; Forrest, Stephen
2015-10-01
We demonstrate a method for extracting waveguided light trapped in the organic and indium tin oxide layers of bottom emission organic light emitting devices (OLEDs) using a patterned planar grid layer (sub-anode grid) between the anode and the substrate. The scattering layer consists of two transparent materials with different refractive indices on a period sufficiently large to avoid diffraction and other unwanted wavelength-dependent effects. The position of the sub-anode grid outside of the OLED active region allows complete freedom in varying its dimensions and materials from which it is made without impacting the electrical characteristics of the device itself. Full wave electromagnetic simulation is used to study the efficiency dependence on refractive indices and geometric parameters of the grid. We show the fabrication process and characterization of OLEDs with two different grids: a buried sub-anode grid consisting of two dielectric materials, and an air sub-anode grid consisting of a dielectric material and gridline voids. Using a sub-anode grid, substrate plus air modes quantum efficiency of an OLED is enhanced from (33+/-2)% to (40+/-2)%, resulting in an increase in external quantum efficiency from (14+/-1)% to (18+/-1)%, with identical electrical characteristics to that of a conventional device. By varying the thickness of the electron transport layer (ETL) of sub-anode grid OLEDs, we find that all power launched into the waveguide modes is scattered into substrate. We also demonstrate a sub-anode grid combined with a thick ETL significantly reduces surface plasmon polaritons, and results in an increase in substrate plus air modes by a >50% compared with a conventional OLED. The wavelength, viewing angle and molecular orientational independence provided by this approach make this an attractive and general solution to the problem of extracting waveguided light and reducing plasmon losses in OLEDs.
An Experimental Investigation of Unsteady Surface Pressure on an Airfoil in Turbulence
NASA Technical Reports Server (NTRS)
Mish, Patrick F.; Devenport, William J.
2003-01-01
Measurements of fluctuating surface pressure were made on a NACA 0015 airfoil immersed in grid generated turbulence. The airfoil model has a 2 ft chord and spans the 6 ft Virginia Tech Stability Wind Tunnel test section. Two grids were used to investigate the effects of turbulence length scale on the surface pressure response. A large grid which produced turbulence with an integral scale 13% of the chord and a smaller grid which produced turbulence with an integral scale 1.3% of the chord. Measurements were performed at angles of attack, alpha from 0 to 20 . An array of microphones mounted subsurface was used to measure the unsteady surface pressure. The goal of this measurement was to characterize the effects of angle of attack on the inviscid response. Lift spectra calculated from pressure measurements at each angle of attack revealed two distinct interaction regions; for omega(sub r) = omega b / U(sub infinity) is less than 10 a reduction in unsteady lift of up to 7 decibels (dB) occurs while an increase occurs for omega(sub r) is greater than 10 as the angle of attack is increased. The reduction in unsteady lift at low omega(sub r) with increasing angle of attack is a result that has never before been shown either experimentally or theoretically. The source of the reduction in lift spectral level appears to be closely related to the distortion of inflow turbulence based on analysis of surface pressure spanwise correlation length scales. Furthermore, while the distortion of the inflow appears to be critical in this experiment, this effect does not seem to be significant in larger integral scale (relative to the chord) flows based on the previous experimental work of McKeough suggesting the airfoils size relative to the inflow integral scale is critical in defining how the airfoil will respond under variation of angle of attack. A prediction scheme is developed that correctly accounts for the effects of distortion when the inflow integral scale is small relative to the airfoil chord. This scheme utilizes Rapid Distortion Theory to account for the distortion of the inflow with the distortion field modeled using a circular cylinder.
NASA Astrophysics Data System (ADS)
Yue, Chao; Ciais, Philippe; Luyssaert, Sebastiaan; Li, Wei; McGrath, Matthew J.; Chang, Jinfeng; Peng, Shushi
2018-01-01
Land use change (LUC) is among the main anthropogenic disturbances in the global carbon cycle. Here we present the model developments in a global dynamic vegetation model ORCHIDEE-MICT v8.4.2 for a more realistic representation of LUC processes. First, we included gross land use change (primarily shifting cultivation) and forest wood harvest in addition to net land use change. Second, we included sub-grid evenly aged land cohorts to represent secondary forests and to keep track of the transient stage of agricultural lands since LUC. Combination of these two features allows the simulation of shifting cultivation with a rotation length involving mainly secondary forests instead of primary ones. Furthermore, a set of decision rules regarding the land cohorts to be targeted in different LUC processes have been implemented. Idealized site-scale simulation has been performed for miombo woodlands in southern Africa assuming an annual land turnover rate of 5 % grid cell area between forest and cropland. The result shows that the model can correctly represent forest recovery and cohort aging arising from agricultural abandonment. Such a land turnover process, even though without a net change in land cover, yields carbon emissions largely due to the imbalance between the fast release from forest clearing and the slow uptake from agricultural abandonment. The simulation with sub-grid land cohorts gives lower emissions than without, mainly because the cleared secondary forests have a lower biomass carbon stock than the mature forests that are otherwise cleared when sub-grid land cohorts are not considered. Over the region of southern Africa, the model is able to account for changes in different forest cohort areas along with the historical changes in different LUC activities, including regrowth of old forests when LUC area decreases. Our developments provide possibilities to account for continental or global forest demographic change resulting from past anthropogenic and natural disturbances.
NASA Astrophysics Data System (ADS)
Eveleth, R.; Cassar, N.; Doney, S. C.; Munro, D. R.; Sweeney, C.
2017-05-01
Using simultaneous sub-kilometer resolution underway measurements of surface O2/Ar, total O2 and pCO2 from annual austral summer surveys in 2012, 2013 and 2014, we explore the impacts of biological and physical processes on the O2 and pCO2 system spatial and interannual variability at the Western Antarctic Peninsula (WAP). In the WAP, mean O2/Ar supersaturation was (7.6±9.1)% and mean pCO2 supersaturation was (-28±22)%. We see substantial spatial variability in O2 and pCO2 including sub-mesoscale/mesoscale variability with decorrelation length scales of 4.5 km, consistent with the regional Rossby radius. This variability is embedded within onshore-offshore gradients. O2 in the LTER grid region is driven primarily by biological processes as seen by the median ratio of the magnitude of biological oxygen (O2/Ar) to physical oxygen (Ar) supersaturation anomalies (%) relative to atmospheric equilibrium (2.6), however physical processes have a more pronounced influence in the southern onshore region of the grid where we see active sea-ice melting. Total O2 measurements should be interpreted with caution in regions of significant sea-ice formation and melt and glacial meltwater input. pCO2 undersaturation predominantly reflects biological processes in the LTER grid. In contrast we compare these results to the Drake Passage where gas supersaturations vary by smaller magnitudes and decorrelate at length scales of 12 km, in line with latitudinal changes in the regional Rossby radius. Here biological processes induce smaller O2/Ar supersaturations (mean (0.14±1.3)%) and pCO2 undersaturations (mean (-2.8±3.9)%) than in the WAP, and pressure changes, bubble and gas exchange fluxes drive stable Ar supersaturations.
NASA Astrophysics Data System (ADS)
Matsui, H.; Buffett, B. A.
2017-12-01
The flow in the Earth's outer core is expected to have vast length scale from the geometry of the outer core to the thickness of the boundary layer. Because of the limitation of the spatial resolution in the numerical simulations, sub-grid scale (SGS) modeling is required to model the effects of the unresolved field on the large-scale fields. We model the effects of sub-grid scale flow and magnetic field using a dynamic scale similarity model. Four terms are introduced for the momentum flux, heat flux, Lorentz force and magnetic induction. The model was previously used in the convection-driven dynamo in a rotating plane layer and spherical shell using the Finite Element Methods. In the present study, we perform large eddy simulations (LES) using the dynamic scale similarity model. The scale similarity model is implement in Calypso, which is a numerical dynamo model using spherical harmonics expansion. To obtain the SGS terms, the spatial filtering in the horizontal directions is done by taking the convolution of a Gaussian filter expressed in terms of a spherical harmonic expansion, following Jekeli (1981). A Gaussian field is also applied in the radial direction. To verify the present model, we perform a fully resolved direct numerical simulation (DNS) with the truncation of the spherical harmonics L = 255 as a reference. And, we perform unresolved DNS and LES with SGS model on coarser resolution (L= 127, 84, and 63) using the same control parameter as the resolved DNS. We will discuss the verification results by comparison among these simulations and role of small scale fields to large scale fields through the role of the SGS terms in LES.
A Virtual Study of Grid Resolution on Experiments of a Highly-Resolved Turbulent Plume
NASA Astrophysics Data System (ADS)
Maisto, Pietro M. F.; Marshall, Andre W.; Gollner, Michael J.; Fire Protection Engineering Department Collaboration
2017-11-01
An accurate representation of sub-grid scale turbulent mixing is critical for modeling fire plumes and smoke transport. In this study, PLIF and PIV diagnostics are used with the saltwater modeling technique to provide highly-resolved instantaneous field measurements in unconfined turbulent plumes useful for statistical analysis, physical insight, and model validation. The effect of resolution was investigated employing a virtual interrogation window (of varying size) applied to the high-resolution field measurements. Motivated by LES low-pass filtering concepts, the high-resolution experimental data in this study can be analyzed within the interrogation windows (i.e. statistics at the sub-grid scale) and on interrogation windows (i.e. statistics at the resolved scale). A dimensionless resolution threshold (L/D*) criterion was determined to achieve converged statistics on the filtered measurements. Such a criterion was then used to establish the relative importance between large and small-scale turbulence phenomena while investigating specific scales for the turbulent flow. First order data sets start to collapse at a resolution of 0.3D*, while for second and higher order statistical moments the interrogation window size drops down to 0.2D*.
NASA Technical Reports Server (NTRS)
Zhou, YE; Vahala, George
1993-01-01
The advection of a passive scalar by incompressible turbulence is considered using recursive renormalization group procedures in the differential sub grid shell thickness limit. It is shown explicitly that the higher order nonlinearities induced by the recursive renormalization group procedure preserve Galilean invariance. Differential equations, valid for the entire resolvable wave number k range, are determined for the eddy viscosity and eddy diffusivity coefficients, and it is shown that higher order nonlinearities do not contribute as k goes to 0, but have an essential role as k goes to k(sub c) the cutoff wave number separating the resolvable scales from the sub grid scales. The recursive renormalization transport coefficients and the associated eddy Prandtl number are in good agreement with the k-dependent transport coefficients derived from closure theories and experiments.
Zhang, Yimei; Li, Shuai; Wang, Fei; Chen, Zhuang; Chen, Jie; Wang, Liqun
2018-09-01
Toxicity of heavy metals from industrialization poses critical concern, and analysis of sources associated with potential human health risks is of unique significance. Assessing human health risk of pollution sources (factored health risk) concurrently in the whole and the sub region can provide more instructive information to protect specific potential victims. In this research, we establish a new expression model of human health risk based on quantitative analysis of sources contribution in different spatial scales. The larger scale grids and their spatial codes are used to initially identify the level of pollution risk, the type of pollution source and the sensitive population at high risk. The smaller scale grids and their spatial codes are used to identify the contribution of various sources of pollution to each sub region (larger grid) and to assess the health risks posed by each source for each sub region. The results of case study show that, for children (sensitive populations, taking school and residential area as major region of activity), the major pollution source is from the abandoned lead-acid battery plant (ALP), traffic emission and agricultural activity. The new models and results of this research present effective spatial information and useful model for quantifying the hazards of source categories and human health a t complex industrial system in the future. Copyright © 2018 Elsevier Ltd. All rights reserved.
Black hole feeding and feedback: the physics inside the `sub-grid'
NASA Astrophysics Data System (ADS)
Negri, A.; Volonteri, M.
2017-05-01
Black holes (BHs) are believed to be a key ingredient of galaxy formation. However, the galaxy-BH interplay is challenging to study due to the large dynamical range and complex physics involved. As a consequence, hydrodynamical cosmological simulations normally adopt sub-grid models to track the unresolved physical processes, in particular BH accretion; usually the spatial scale where the BH dominates the hydrodynamical processes (the Bondi radius) is unresolved, and an approximate Bondi-Hoyle accretion rate is used to estimate the growth of the BH. By comparing hydrodynamical simulations at different resolutions (300, 30, 3 pc) using a Bondi-Hoyle approximation to sub-parsec runs with non-parametrized accretion, our aim is to probe how well an approximated Bondi accretion is able to capture the BH accretion physics and the subsequent feedback on the galaxy. We analyse an isolated galaxy simulation that includes cooling, star formation, Type Ia and Type II supernovae, BH accretion and active galactic nuclei feedback (radiation pressure, Compton heating/cooling) where mass, momentum and energy are deposited in the interstellar medium through conical winds. We find that on average the approximated Bondi formalism can lead to both over- and underestimations of the BH growth, depending on resolution and on how the variables entering into the Bondi-Hoyle formalism are calculated.
The influence of sub-grid scale motions on particle collision in homogeneous isotropic turbulence
NASA Astrophysics Data System (ADS)
Xiong, Yan; Li, Jing; Liu, Zhaohui; Zheng, Chuguang
2018-02-01
The absence of sub-grid scale (SGS) motions leads to severe errors in particle pair dynamics, which represents a great challenge to the large eddy simulation of particle-laden turbulent flow. In order to address this issue, data from direct numerical simulation (DNS) of homogenous isotropic turbulence coupled with Lagrangian particle tracking are used as a benchmark to evaluate the corresponding results of filtered DNS (FDNS). It is found that the filtering process in FDNS will lead to a non-monotonic variation of the particle collision statistics, including radial distribution function, radial relative velocity, and the collision kernel. The peak of radial distribution function shifts to the large-inertia region due to the lack of SGS motions, and the analysis of the local flowstructure characteristic variable at particle position indicates that the most effective interaction scale between particles and fluid eddies is increased in FDNS. Moreover, this scale shifting has an obvious effect on the odd-order moments of the probability density function of radial relative velocity, i.e. the skewness, which exhibits a strong correlation to the variance of radial distribution function in FDNS. As a whole, the radial distribution function, together with radial relative velocity, can compensate the SGS effects for the collision kernel in FDNS when the Stokes number based on the Kolmogorov time scale is greater than 3.0. However, it still leaves considerable errors for { St}_k <3.0.
Quantifying the impact of sub-grid surface wind variability on sea salt and dust emissions in CAM5
NASA Astrophysics Data System (ADS)
Zhang, Kai; Zhao, Chun; Wan, Hui; Qian, Yun; Easter, Richard C.; Ghan, Steven J.; Sakaguchi, Koichi; Liu, Xiaohong
2016-02-01
This paper evaluates the impact of sub-grid variability of surface wind on sea salt and dust emissions in the Community Atmosphere Model version 5 (CAM5). The basic strategy is to calculate emission fluxes multiple times, using different wind speed samples of a Weibull probability distribution derived from model-predicted grid-box mean quantities. In order to derive the Weibull distribution, the sub-grid standard deviation of surface wind speed is estimated by taking into account four mechanisms: turbulence under neutral and stable conditions, dry convective eddies, moist convective eddies over the ocean, and air motions induced by mesoscale systems and fine-scale topography over land. The contributions of turbulence and dry convective eddy are parameterized using schemes from the literature. Wind variabilities caused by moist convective eddies and fine-scale topography are estimated using empirical relationships derived from an operational weather analysis data set at 15 km resolution. The estimated sub-grid standard deviations of surface wind speed agree well with reference results derived from 1 year of global weather analysis at 15 km resolution and from two regional model simulations with 3 km grid spacing.The wind-distribution-based emission calculations are implemented in CAM5. In terms of computational cost, the increase in total simulation time turns out to be less than 3 %. Simulations at 2° resolution indicate that sub-grid wind variability has relatively small impacts (about 7 % increase) on the global annual mean emission of sea salt aerosols, but considerable influence on the emission of dust. Among the considered mechanisms, dry convective eddies and mesoscale flows associated with topography are major causes of dust emission enhancement. With all the four mechanisms included and without additional adjustment of uncertain parameters in the model, the simulated global and annual mean dust emission increase by about 50 % compared to the default model. By tuning the globally constant dust emission scale factor, the global annual mean dust emission, aerosol optical depth, and top-of-atmosphere radiative fluxes can be adjusted to the level of the default model, but the frequency distribution of dust emission changes, with more contribution from weaker wind events and less contribution from stronger wind events. In Africa and Asia, the overall frequencies of occurrence of dust emissions increase, and the seasonal variations are enhanced, while the geographical patterns of the emission frequency show little change.
Quantifying the impact of sub-grid surface wind variability on sea salt and dust emissions in CAM5
Zhang, Kai; Zhao, Chun; Wan, Hui; ...
2016-02-12
This paper evaluates the impact of sub-grid variability of surface wind on sea salt and dust emissions in the Community Atmosphere Model version 5 (CAM5). The basic strategy is to calculate emission fluxes multiple times, using different wind speed samples of a Weibull probability distribution derived from model-predicted grid-box mean quantities. In order to derive the Weibull distribution, the sub-grid standard deviation of surface wind speed is estimated by taking into account four mechanisms: turbulence under neutral and stable conditions, dry convective eddies, moist convective eddies over the ocean, and air motions induced by mesoscale systems and fine-scale topography overmore » land. The contributions of turbulence and dry convective eddy are parameterized using schemes from the literature. Wind variabilities caused by moist convective eddies and fine-scale topography are estimated using empirical relationships derived from an operational weather analysis data set at 15 km resolution. The estimated sub-grid standard deviations of surface wind speed agree well with reference results derived from 1 year of global weather analysis at 15 km resolution and from two regional model simulations with 3 km grid spacing.The wind-distribution-based emission calculations are implemented in CAM5. In terms of computational cost, the increase in total simulation time turns out to be less than 3 %. Simulations at 2° resolution indicate that sub-grid wind variability has relatively small impacts (about 7 % increase) on the global annual mean emission of sea salt aerosols, but considerable influence on the emission of dust. Among the considered mechanisms, dry convective eddies and mesoscale flows associated with topography are major causes of dust emission enhancement. With all the four mechanisms included and without additional adjustment of uncertain parameters in the model, the simulated global and annual mean dust emission increase by about 50 % compared to the default model. By tuning the globally constant dust emission scale factor, the global annual mean dust emission, aerosol optical depth, and top-of-atmosphere radiative fluxes can be adjusted to the level of the default model, but the frequency distribution of dust emission changes, with more contribution from weaker wind events and less contribution from stronger wind events. Lastly, in Africa and Asia, the overall frequencies of occurrence of dust emissions increase, and the seasonal variations are enhanced, while the geographical patterns of the emission frequency show little change.« less
Quantifying the impact of sub-grid surface wind variability on sea salt and dust emissions in CAM5
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhang, Kai; Zhao, Chun; Wan, Hui
This paper evaluates the impact of sub-grid variability of surface wind on sea salt and dust emissions in the Community Atmosphere Model version 5 (CAM5). The basic strategy is to calculate emission fluxes multiple times, using different wind speed samples of a Weibull probability distribution derived from model-predicted grid-box mean quantities. In order to derive the Weibull distribution, the sub-grid standard deviation of surface wind speed is estimated by taking into account four mechanisms: turbulence under neutral and stable conditions, dry convective eddies, moist convective eddies over the ocean, and air motions induced by mesoscale systems and fine-scale topography overmore » land. The contributions of turbulence and dry convective eddy are parameterized using schemes from the literature. Wind variabilities caused by moist convective eddies and fine-scale topography are estimated using empirical relationships derived from an operational weather analysis data set at 15 km resolution. The estimated sub-grid standard deviations of surface wind speed agree well with reference results derived from 1 year of global weather analysis at 15 km resolution and from two regional model simulations with 3 km grid spacing.The wind-distribution-based emission calculations are implemented in CAM5. In terms of computational cost, the increase in total simulation time turns out to be less than 3 %. Simulations at 2° resolution indicate that sub-grid wind variability has relatively small impacts (about 7 % increase) on the global annual mean emission of sea salt aerosols, but considerable influence on the emission of dust. Among the considered mechanisms, dry convective eddies and mesoscale flows associated with topography are major causes of dust emission enhancement. With all the four mechanisms included and without additional adjustment of uncertain parameters in the model, the simulated global and annual mean dust emission increase by about 50 % compared to the default model. By tuning the globally constant dust emission scale factor, the global annual mean dust emission, aerosol optical depth, and top-of-atmosphere radiative fluxes can be adjusted to the level of the default model, but the frequency distribution of dust emission changes, with more contribution from weaker wind events and less contribution from stronger wind events. Lastly, in Africa and Asia, the overall frequencies of occurrence of dust emissions increase, and the seasonal variations are enhanced, while the geographical patterns of the emission frequency show little change.« less
NASA Astrophysics Data System (ADS)
Li, Y.; McDougall, T. J.
2016-02-01
Coarse resolution ocean models lack knowledge of spatial correlations between variables on scales smaller than the grid scale. Some researchers have shown that these spatial correlations play a role in the poleward heat flux. In order to evaluate the poleward transport induced by the spatial correlations at a fixed horizontal position, an equation is obtained to calculate the approximate transport from velocity gradients. The equation involves two terms that can be added to the quasi-Stokes streamfunction (based on temporal correlations) to incorporate the contribution of spatial correlations. Moreover, these new terms do not need to be parameterized and is ready to be evaluated by using model data directly. In this study, data from a high resolution ocean model have been used to estimate the accuracy of this HRM approach for improving the horizontal property fluxes in coarse-resolution ocean models. A coarse grid is formed by sub-sampling and box-car averaging the fine grid scale. The transport calculated on the coarse grid is then compared to the transport on original high resolution grid scale accumulated over a corresponding number of grid boxes. The preliminary results have shown that the estimate on coarse resolution grids roughly match the corresponding transports on high resolution grids.
Tropical precipitation extremes: Response to SST-induced warming in aquaplanet simulations
NASA Astrophysics Data System (ADS)
Bhattacharya, Ritthik; Bordoni, Simona; Teixeira, João.
2017-04-01
Scaling of tropical precipitation extremes in response to warming is studied in aquaplanet experiments using the global Weather Research and Forecasting (WRF) model. We show how the scaling of precipitation extremes is highly sensitive to spatial and temporal averaging: while instantaneous grid point extreme precipitation scales more strongly than the percentage increase (˜7% K-1) predicted by the Clausius-Clapeyron (CC) relationship, extremes for zonally and temporally averaged precipitation follow a slight sub-CC scaling, in agreement with results from Climate Model Intercomparison Project (CMIP) models. The scaling depends crucially on the employed convection parameterization. This is particularly true when grid point instantaneous extremes are considered. These results highlight how understanding the response of precipitation extremes to warming requires consideration of dynamic changes in addition to the thermodynamic response. Changes in grid-scale precipitation, unlike those in convective-scale precipitation, scale linearly with the resolved flow. Hence, dynamic changes include changes in both large-scale and convective-scale motions.
Lai, Canhai; Xu, Zhijie; Li, Tingwen; ...
2017-08-05
In virtual design and scale up of pilot-scale carbon capture systems, the coupled reactive multiphase flow problem must be solved to predict the adsorber's performance and capture efficiency under various operation conditions. This paper focuses on the detailed computational fluid dynamics (CFD) modeling of a pilot-scale fluidized bed adsorber equipped with vertical cooling tubes. Multiphase Flow with Interphase eXchanges (MFiX), an open-source multiphase flow CFD solver, is used for the simulations with custom code to simulate the chemical reactions and filtered sub-grid models to capture the effect of the unresolved details in the coarser mesh for simulations with reasonable accuracymore » and manageable computational effort. Previously developed filtered models for horizontal cylinder drag, heat transfer, and reaction kinetics have been modified to derive the 2D filtered models representing vertical cylinders in the coarse-grid CFD simulations. The effects of the heat exchanger configurations (i.e., horizontal or vertical tubes) on the adsorber's hydrodynamics and CO 2 capture performance are then examined. A one-dimensional three-region process model is briefly introduced for comparison purpose. The CFD model matches reasonably well with the process model while provides additional information about the flow field that is not available with the process model.« less
NCAR global model topography generation software for unstructured grids
NASA Astrophysics Data System (ADS)
Lauritzen, P. H.; Bacmeister, J. T.; Callaghan, P. F.; Taylor, M. A.
2015-06-01
It is the purpose of this paper to document the NCAR global model topography generation software for unstructured grids. Given a model grid, the software computes the fraction of the grid box covered by land, the gridbox mean elevation, and associated sub-grid scale variances commonly used for gravity wave and turbulent mountain stress parameterizations. The software supports regular latitude-longitude grids as well as unstructured grids; e.g. icosahedral, Voronoi, cubed-sphere and variable resolution grids. As an example application and in the spirit of documenting model development, exploratory simulations illustrating the impacts of topographic smoothing with the NCAR-DOE CESM (Community Earth System Model) CAM5.2-SE (Community Atmosphere Model version 5.2 - Spectral Elements dynamical core) are shown.
Sub-grid drag models for horizontal cylinder arrays immersed in gas-particle multiphase flows
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sarkar, Avik; Sun, Xin; Sundaresan, Sankaran
2013-09-08
Immersed cylindrical tube arrays often are used as heat exchangers in gas-particle fluidized beds. In multiphase computational fluid dynamics (CFD) simulations of large fluidized beds, explicit resolution of small cylinders is computationally infeasible. Instead, the cylinder array may be viewed as an effective porous medium in coarse-grid simulations. The cylinders' influence on the suspension as a whole, manifested as an effective drag force, and on the relative motion between gas and particles, manifested as a correction to the gas-particle drag, must be modeled via suitable sub-grid constitutive relationships. In this work, highly resolved unit-cell simulations of flow around an arraymore » of horizontal cylinders, arranged in a staggered configuration, are filtered to construct sub-grid, or `filtered', drag models, which can be implemented in coarse-grid simulations. The force on the suspension exerted by the cylinders is comprised of, as expected, a buoyancy contribution, and a kinetic component analogous to fluid drag on a single cylinder. Furthermore, the introduction of tubes also is found to enhance segregation at the scale of the cylinder size, which, in turn, leads to a reduction in the filtered gas-particle drag.« less
A grid of MHD models for stellar mass loss and spin-down rates of solar analogs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cohen, O.; Drake, J. J.
2014-03-01
Stellar winds are believed to be the dominant factor in the spin-down of stars over time. However, stellar winds of solar analogs are poorly constrained due to observational challenges. In this paper, we present a grid of magnetohydrodynamic models to study and quantify the values of stellar mass loss and angular momentum loss rates as a function of the stellar rotation period, magnetic dipole component, and coronal base density. We derive simple scaling laws for the loss rates as a function of these parameters, and constrain the possible mass loss rate of stars with thermally driven winds. Despite the successmore » of our scaling law in matching the results of the model, we find a deviation between the 'solar dipole' case and a real case based on solar observations that overestimates the actual solar mass loss rate by a factor of three. This implies that the model for stellar fields might require a further investigation with additional complexity. Mass loss rates in general are largely controlled by the magnetic field strength, with the wind density varying in proportion to the confining magnetic pressure B {sup 2}. We also find that the mass loss rates obtained using our grid models drop much faster with the increase in rotation period than scaling laws derived using observed stellar activity. For main-sequence solar-like stars, our scaling law for angular momentum loss versus poloidal magnetic field strength retrieves the well-known Skumanich decline of angular velocity with time, Ω{sub *}∝t {sup –1/2}, if the large-scale poloidal magnetic field scales with rotation rate as B{sub p}∝Ω{sub ⋆}{sup 2}.« less
The role of optimality in characterizing CO2 seepage from geological carbon sequestration sites
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cortis, Andrea; Oldenburg, Curtis M.; Benson, Sally M.
Storage of large amounts of carbon dioxide (CO{sub 2}) in deep geological formations for greenhouse gas mitigation is gaining momentum and moving from its conceptual and testing stages towards widespread application. In this work we explore various optimization strategies for characterizing surface leakage (seepage) using near-surface measurement approaches such as accumulation chambers and eddy covariance towers. Seepage characterization objectives and limitations need to be defined carefully from the outset especially in light of large natural background variations that can mask seepage. The cost and sensitivity of seepage detection are related to four critical length scales pertaining to the size ofmore » the: (1) region that needs to be monitored; (2) footprint of the measurement approach, and (3) main seepage zone; and (4) region in which concentrations or fluxes are influenced by seepage. Seepage characterization objectives may include one or all of the tasks of detecting, locating, and quantifying seepage. Each of these tasks has its own optimal strategy. Detecting and locating seepage in a region in which there is no expected or preferred location for seepage nor existing evidence for seepage requires monitoring on a fixed grid, e.g., using eddy covariance towers. The fixed-grid approaches needed to detect seepage are expected to require large numbers of eddy covariance towers for large-scale geologic CO{sub 2} storage. Once seepage has been detected and roughly located, seepage zones and features can be optimally pinpointed through a dynamic search strategy, e.g., employing accumulation chambers and/or soil-gas sampling. Quantification of seepage rates can be done through measurements on a localized fixed grid once the seepage is pinpointed. Background measurements are essential for seepage detection in natural ecosystems. Artificial neural networks are considered as regression models useful for distinguishing natural system behavior from anomalous behavior suggestive of CO{sub 2} seepage without need for detailed understanding of natural system processes. Because of the local extrema in CO{sub 2} fluxes and concentrations in natural systems, simple steepest-descent algorithms are not effective and evolutionary computation algorithms are proposed as a paradigm for dynamic monitoring networks to pinpoint CO{sub 2} seepage areas.« less
Use of upscaled elevation and surface roughness data in two-dimensional surface water models
Hughes, J.D.; Decker, J.D.; Langevin, C.D.
2011-01-01
In this paper, we present an approach that uses a combination of cell-block- and cell-face-averaging of high-resolution cell elevation and roughness data to upscale hydraulic parameters and accurately simulate surface water flow in relatively low-resolution numerical models. The method developed allows channelized features that preferentially connect large-scale grid cells at cell interfaces to be represented in models where these features are significantly smaller than the selected grid size. The developed upscaling approach has been implemented in a two-dimensional finite difference model that solves a diffusive wave approximation of the depth-integrated shallow surface water equations using preconditioned Newton–Krylov methods. Computational results are presented to show the effectiveness of the mixed cell-block and cell-face averaging upscaling approach in maintaining model accuracy, reducing model run-times, and how decreased grid resolution affects errors. Application examples demonstrate that sub-grid roughness coefficient variations have a larger effect on simulated error than sub-grid elevation variations.
This presentation explains the importance of the fine-scale features for air toxics exposure modeling. The paper presents a new approach to combine local-scale and regional model results for the National Air Toxic Assessment. The technique has been evaluated with a chemical tra...
Uncertainty quantification in LES of channel flow
Safta, Cosmin; Blaylock, Myra; Templeton, Jeremy; ...
2016-07-12
Here, in this paper, we present a Bayesian framework for estimating joint densities for large eddy simulation (LES) sub-grid scale model parameters based on canonical forced isotropic turbulence direct numerical simulation (DNS) data. The framework accounts for noise in the independent variables, and we present alternative formulations for accounting for discrepancies between model and data. To generate probability densities for flow characteristics, posterior densities for sub-grid scale model parameters are propagated forward through LES of channel flow and compared with DNS data. Synthesis of the calibration and prediction results demonstrates that model parameters have an explicit filter width dependence andmore » are highly correlated. Discrepancies between DNS and calibrated LES results point to additional model form inadequacies that need to be accounted for.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brinkman, Gregory
2015-09-01
The Renewable Electricity Futures Study (RE Futures)--an analysis of the costs and grid impacts of integrating large amounts of renewable electricity generation into the U.S. power system--examined renewable energy resources, technical issues regarding the integration of these resources into the grid, and the costs associated with high renewable penetration scenarios. These scenarios included up to 90% of annual generation from renewable sources, although most of the analysis was focused on 80% penetration scenarios. Hourly production cost modeling was performed to understand the operational impacts of high penetrations. One of the conclusions of RE Futures was that further work was necessarymore » to understand whether the operation of the system was possible at sub-hourly time scales and during transient events. This study aimed to address part of this by modeling the operation of the power system at sub-hourly time scales using newer methodologies and updated data sets for transmission and generation infrastructure. The goal of this work was to perform a detailed, sub-hourly analysis of very high penetration scenarios for a single interconnection (the Western Interconnection). It focused on operational impacts, and it helps verify that the operational results from the capacity expansion models are useful. The primary conclusion of this study is that sub-hourly operation of the grid is possible with renewable generation levels between 80% and 90%.« less
NASA Astrophysics Data System (ADS)
Kumar, R.; Samaniego, L. E.; Livneh, B.
2013-12-01
Knowledge of soil hydraulic properties such as porosity and saturated hydraulic conductivity is required to accurately model the dynamics of near-surface hydrological processes (e.g. evapotranspiration and root-zone soil moisture dynamics) and provide reliable estimates of regional water and energy budgets. Soil hydraulic properties are commonly derived from pedo-transfer functions using soil textural information recorded during surveys, such as the fractions of sand and clay, bulk density, and organic matter content. Typically large scale land-surface models are parameterized using a relatively coarse soil map with little or no information on parametric sub-grid variability. In this study we analyze the impact of sub-grid soil variability on simulated hydrological fluxes over the Mississippi River Basin (≈3,240,000 km2) at multiple spatio-temporal resolutions. A set of numerical experiments were conducted with the distributed mesoscale hydrologic model (mHM) using two soil datasets: (a) the Digital General Soil Map of the United States or STATSGO2 (1:250 000) and (b) the recently collated Harmonized World Soil Database based on the FAO-UNESCO Soil Map of the World (1:5 000 000). mHM was parameterized with the multi-scale regionalization technique that derives distributed soil hydraulic properties via pedo-transfer functions and regional coefficients. Within the experimental framework, the 3-hourly model simulations were conducted at four spatial resolutions ranging from 0.125° to 1°, using meteorological datasets from the NLDAS-2 project for the time period 1980-2012. Preliminary results indicate that the model was able to capture observed streamflow behavior reasonably well with both soil datasets, in the major sub-basins (i.e. the Missouri, the Upper Mississippi, the Ohio, the Red, and the Arkansas). However, the spatio-temporal patterns of simulated water fluxes and states (e.g. soil moisture, evapotranspiration) from both simulations, showed marked differences; particularly at a shorter time scale (hours to days) in regions with coarse texture sandy soils. Furthermore, the partitioning of total runoff into near-surface interflows and baseflow components was also significantly different between the two simulations. Simulations with the coarser soil map produced comparatively higher baseflows. At longer time scales (months to seasons) where climatic factors plays a major role, the integrated fluxes and states from both sets of model simulations match fairly closely, despite the apparent discrepancy in the partitioning of total runoff.
Characterization of Slosh Damping for Ortho-Grid and Iso-Grid Internal Tank Structures
NASA Technical Reports Server (NTRS)
Westra, Douglas G.; Sansone, Marco D.; Eberhart, Chad J.; West, Jeffrey S.
2016-01-01
Grid stiffened tank structures such as Ortho-Grid and Iso-Grid are widely used in cryogenic tanks for providing stiffening to the tank while reducing mass, compared to tank walls of constant cross-section. If the structure is internal to the tank, it will positively affect the fluid dynamic behavior of the liquid propellant, in regard to fluid slosh damping. As NASA and commercial companies endeavor to explore the solar system, vehicles will by necessity become more mass efficient, and design margin will be reduced where possible. Therefore, if the damping characteristics of the Ortho-Grid and Iso-Grid structure is understood, their positive damping effect can be taken into account in the systems design process. Historically, damping by internal structures has been characterized by rules of thumb and for Ortho-Grid, empirical design tools intended for slosh baffles of much larger cross-section have been used. There is little or no information available to characterize the slosh behavior of Iso-Grid internal structure. Therefore, to take advantage of these structures for their positive damping effects, there is much need for obtaining additional data and tools to characterize them. Recently, the NASA Marshall Space Flight Center conducted both sub-scale testing and computational fluid dynamics (CFD) simulations of slosh damping for Ortho-Grid and Iso-Grid tanks for cylindrical tanks containing water. Enhanced grid meshing techniques were applied to the geometrically detailed and complex Ortho-Grid and Iso-Grid structures. The Loci-STREAM CFD program with the Volume of Fluid Method module for tracking and locating the water-air fluid interface was used to conduct the simulations. The CFD simulations were validated with the test data and new empirical models for predicting damping and frequency of Ortho-Grid and Iso-Grid structures were generated.
GRID-BASED EXPLORATION OF COSMOLOGICAL PARAMETER SPACE WITH SNAKE
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mikkelsen, K.; Næss, S. K.; Eriksen, H. K., E-mail: kristin.mikkelsen@astro.uio.no
2013-11-10
We present a fully parallelized grid-based parameter estimation algorithm for investigating multidimensional likelihoods called Snake, and apply it to cosmological parameter estimation. The basic idea is to map out the likelihood grid-cell by grid-cell according to decreasing likelihood, and stop when a certain threshold has been reached. This approach improves vastly on the 'curse of dimensionality' problem plaguing standard grid-based parameter estimation simply by disregarding grid cells with negligible likelihood. The main advantages of this method compared to standard Metropolis-Hastings Markov Chain Monte Carlo methods include (1) trivial extraction of arbitrary conditional distributions; (2) direct access to Bayesian evidences; (3)more » better sampling of the tails of the distribution; and (4) nearly perfect parallelization scaling. The main disadvantage is, as in the case of brute-force grid-based evaluation, a dependency on the number of parameters, N{sub par}. One of the main goals of the present paper is to determine how large N{sub par} can be, while still maintaining reasonable computational efficiency; we find that N{sub par} = 12 is well within the capabilities of the method. The performance of the code is tested by comparing cosmological parameters estimated using Snake and the WMAP-7 data with those obtained using CosmoMC, the current standard code in the field. We find fully consistent results, with similar computational expenses, but shorter wall time due to the perfect parallelization scheme.« less
On the Subgrid-Scale Modeling of Compressible Turbulence
NASA Technical Reports Server (NTRS)
Squires, Kyle; Zeman, Otto
1990-01-01
A new sub-grid scale model is presented for the large-eddy simulation of compressible turbulence. In the proposed model, compressibility contributions have been incorporated in the sub-grid scale eddy viscosity which, in the incompressible limit, reduce to a form originally proposed by Smagorinsky (1963). The model has been tested against a simple extension of the traditional Smagorinsky eddy viscosity model using simulations of decaying, compressible homogeneous turbulence. Simulation results show that the proposed model provides greater dissipation of the compressive modes of the resolved-scale velocity field than does the Smagorinsky eddy viscosity model. For an initial r.m.s. turbulence Mach number of 1.0, simulations performed using the Smagorinsky model become physically unrealizable (i.e., negative energies) because of the inability of the model to sufficiently dissipate fluctuations due to resolved scale velocity dilations. The proposed model is able to provide the necessary dissipation of this energy and maintain the realizability of the flow. Following Zeman (1990), turbulent shocklets are considered to dissipate energy independent of the Kolmogorov energy cascade. A possible parameterization of dissipation by turbulent shocklets for Large-Eddy Simulation is also presented.
Unstructured grid modelling of offshore wind farm impacts on seasonally stratified shelf seas
NASA Astrophysics Data System (ADS)
Cazenave, Pierre William; Torres, Ricardo; Allen, J. Icarus
2016-06-01
Shelf seas comprise approximately 7% of the world's oceans and host enormous economic activity. Development of energy installations (e.g. Offshore Wind Farms (OWFs), tidal turbines) in response to increased demand for renewable energy requires a careful analysis of potential impacts. Recent remote sensing observations have identified kilometre-scale impacts from OWFs. Existing modelling evaluating monopile impacts has fallen into two camps: small-scale models with individually resolved turbines looking at local effects; and large-scale analyses but with sub-grid scale turbine parameterisations. This work straddles both scales through a 3D unstructured grid model (FVCOM): wind turbine monopiles in the eastern Irish Sea are explicitly described in the grid whilst the overall grid domain covers the south-western UK shelf. Localised regions of decreased velocity extend up to 250 times the monopile diameter away from the monopile. Shelf-wide, the amplitude of the M2 tidal constituent increases by up to 7%. The turbines enhance localised vertical mixing which decreases seasonal stratification. The spatial extent of this extends well beyond the turbines into the surrounding seas. With significant expansion of OWFs on continental shelves, this work highlights the importance of how OWFs may impact coastal (e.g. increased flooding risk) and offshore (e.g. stratification and nutrient cycling) areas.
GEWEX Cloud Systems Study (GCSS)
NASA Technical Reports Server (NTRS)
Moncrieff, Mitch
1993-01-01
The Global Energy and Water Cycle Experiment (GEWEX) Cloud Systems Study (GCSS) program seeks to improve the physical understanding of sub-grid scale cloud processes and their representation in parameterization schemes. By improving the description and understanding of key cloud system processes, GCSS aims to develop the necessary parameterizations in climate and numerical weather prediction (NWP) models. GCSS will address these issues mainly through the development and use of cloud-resolving or cumulus ensemble models to generate realizations of a set of archetypal cloud systems. The focus of GCSS is on mesoscale cloud systems, including precipitating convectively-driven cloud systems like MCS's and boundary layer clouds, rather than individual clouds, and on their large-scale effects. Some of the key scientific issues confronting GCSS that particularly relate to research activities in the central U.S. are presented.
Atmospheric Boundary Layer Modeling for Combined Meteorology and Air Quality Systems
Atmospheric Eulerian grid models for mesoscale and larger applications require sub-grid models for turbulent vertical exchange processes, particularly within the Planetary Boundary Layer (PSL). In combined meteorology and air quality modeling systems consistent PSL modeling of wi...
CFD Script for Rapid TPS Damage Assessment
NASA Technical Reports Server (NTRS)
McCloud, Peter
2013-01-01
This grid generation script creates unstructured CFD grids for rapid thermal protection system (TPS) damage aeroheating assessments. The existing manual solution is cumbersome, open to errors, and slow. The invention takes a large-scale geometry grid and its large-scale CFD solution, and creates a unstructured patch grid that models the TPS damage. The flow field boundary condition for the patch grid is then interpolated from the large-scale CFD solution. It speeds up the generation of CFD grids and solutions in the modeling of TPS damages and their aeroheating assessment. This process was successfully utilized during STS-134.
NASA Astrophysics Data System (ADS)
Singh, Gurjeet; Panda, Rabindra K.; Mohanty, Binayak P.; Jana, Raghavendra B.
2016-05-01
Strategic ground-based sampling of soil moisture across multiple scales is necessary to validate remotely sensed quantities such as NASA's Soil Moisture Active Passive (SMAP) product. In the present study, in-situ soil moisture data were collected at two nested scale extents (0.5 km and 3 km) to understand the trend of soil moisture variability across these scales. This ground-based soil moisture sampling was conducted in the 500 km2 Rana watershed situated in eastern India. The study area is characterized as sub-humid, sub-tropical climate with average annual rainfall of about 1456 mm. Three 3x3 km square grids were sampled intensively once a day at 49 locations each, at a spacing of 0.5 km. These intensive sampling locations were selected on the basis of different topography, soil properties and vegetation characteristics. In addition, measurements were also made at 9 locations around each intensive sampling grid at 3 km spacing to cover a 9x9 km square grid. Intensive fine scale soil moisture sampling as well as coarser scale samplings were made using both impedance probes and gravimetric analyses in the study watershed. The ground-based soil moisture samplings were conducted during the day, concurrent with the SMAP descending overpass. Analysis of soil moisture spatial variability in terms of areal mean soil moisture and the statistics of higher-order moments, i.e., the standard deviation, and the coefficient of variation are presented. Results showed that the standard deviation and coefficient of variation of measured soil moisture decreased with extent scale by increasing mean soil moisture.
NASA Astrophysics Data System (ADS)
Tan, Z.; Schneider, T.; Teixeira, J.; Lam, R.; Pressel, K. G.
2014-12-01
Sub-grid scale (SGS) closures in current climate models are usually decomposed into several largely independent parameterization schemes for different cloud and convective processes, such as boundary layer turbulence, shallow convection, and deep convection. These separate parameterizations usually do not converge as the resolution is increased or as physical limits are taken. This makes it difficult to represent the interactions and smooth transition among different cloud and convective regimes. Here we present an eddy-diffusivity mass-flux (EDMF) closure that represents all sub-grid scale turbulent, convective, and cloud processes in a unified parameterization scheme. The buoyant updrafts and precipitative downdrafts are parameterized with a prognostic multiple-plume mass-flux (MF) scheme. The prognostic term for the mass flux is kept so that the life cycles of convective plumes are better represented. The interaction between updrafts and downdrafts are parameterized with the buoyancy-sorting model. The turbulent mixing outside plumes is represented by eddy diffusion, in which eddy diffusivity (ED) is determined from a turbulent kinetic energy (TKE) calculated from a TKE balance that couples the environment with updrafts and downdrafts. Similarly, tracer variances are decomposed consistently between updrafts, downdrafts and the environment. The closure is internally coupled with a probabilistic cloud scheme and a simple precipitation scheme. We have also developed a relatively simple two-stream radiative scheme that includes the longwave (LW) and shortwave (SW) effects of clouds, and the LW effect of water vapor. We have tested this closure in a single-column model for various regimes spanning stratocumulus, shallow cumulus, and deep convection. The model is also run towards statistical equilibrium with climatologically relevant large-scale forcings. These model tests are validated against large-eddy simulation (LES) with the same forcings. The comparison of results verifies the capacity of this closure to realistically represent different cloud and convective processes. Implementation of the closure in an idealized GCM allows us to study cloud feedbacks to climate change and to study the interactions between clouds, convections, and the large-scale circulation.
A LES-Langevin model for turbulence
NASA Astrophysics Data System (ADS)
Dolganov, Rostislav; Dubrulle, Bérengère; Laval, Jean-Philippe
2006-11-01
The rationale for Large Eddy Simulation is rooted in our inability to handle all degrees of freedom (N˜10^16 for Re˜10^7). ``Deterministic'' models based on eddy-viscosity seek to reproduce the intensification of the energy transport. However, they fail to reproduce backward energy transfer (backscatter) from small to large scale, which is an essentiel feature of the turbulence near wall or in boundary layer. To capture this backscatter, ``stochastic'' strategies have been developed. In the present talk, we shall discuss such a strategy, based on a Rapid Distorsion Theory (RDT). Specifically, we first divide the small scale contribution to the Reynolds Stress Tensor in two parts: a turbulent viscosity and the pseudo-Lamb vector, representing the nonlinear cross terms of resolved and sub-grid scales. We then estimate the dynamics of small-scale motion by the RDT applied to Navier-Stockes equation. We use this to model the cross term evolution by a Langevin equation, in which the random force is provided by sub-grid pressure terms. Our LES model is thus made of a truncated Navier-Stockes equation including the turbulent force and a generalized Langevin equation for the latter, integrated on a twice-finer grid. The backscatter is automatically included in our stochastic model of the pseudo-Lamb vector. We apply this model to the case of homogeneous isotropic turbulence and turbulent channel flow.
Harnessing Big Data to Represent 30-meter Spatial Heterogeneity in Earth System Models
NASA Astrophysics Data System (ADS)
Chaney, N.; Shevliakova, E.; Malyshev, S.; Van Huijgevoort, M.; Milly, C.; Sulman, B. N.
2016-12-01
Terrestrial land surface processes play a critical role in the Earth system; they have a profound impact on the global climate, food and energy production, freshwater resources, and biodiversity. One of the most fascinating yet challenging aspects of characterizing terrestrial ecosystems is their field-scale (˜30 m) spatial heterogeneity. It has been observed repeatedly that the water, energy, and biogeochemical cycles at multiple temporal and spatial scales have deep ties to an ecosystem's spatial structure. Current Earth system models largely disregard this important relationship leading to an inadequate representation of ecosystem dynamics. In this presentation, we will show how existing global environmental datasets can be harnessed to explicitly represent field-scale spatial heterogeneity in Earth system models. For each macroscale grid cell, these environmental data are clustered according to their field-scale soil and topographic attributes to define unique sub-grid tiles. The state-of-the-art Geophysical Fluid Dynamics Laboratory (GFDL) land model is then used to simulate these tiles and their spatial interactions via the exchange of water, energy, and nutrients along explicit topographic gradients. Using historical simulations over the contiguous United States, we will show how a robust representation of field-scale spatial heterogeneity impacts modeled ecosystem dynamics including the water, energy, and biogeochemical cycles as well as vegetation composition and distribution.
A satellite simulator for TRMM PR applied to climate model simulations
NASA Astrophysics Data System (ADS)
Spangehl, T.; Schroeder, M.; Bodas-Salcedo, A.; Hollmann, R.; Riley Dellaripa, E. M.; Schumacher, C.
2017-12-01
Climate model simulations have to be compared against observation based datasets in order to assess their skill in representing precipitation characteristics. Here we use a satellite simulator for TRMM PR in order to evaluate simulations performed with MPI-ESM (Earth system model of the Max Planck Institute for Meteorology in Hamburg, Germany) performed within the MiKlip project (https://www.fona-miklip.de/, funded by Federal Ministry of Education and Research in Germany). While classical evaluation methods focus on geophysical parameters such as precipitation amounts, the application of the satellite simulator enables an evaluation in the instrument's parameter space thereby reducing uncertainties on the reference side. The CFMIP Observation Simulator Package (COSP) provides a framework for the application of satellite simulators to climate model simulations. The approach requires the introduction of sub-grid cloud and precipitation variability. Radar reflectivities are obtained by applying Mie theory, with the microphysical assumptions being chosen to match the atmosphere component of MPI-ESM (ECHAM6). The results are found to be sensitive to the methods used to distribute the convective precipitation over the sub-grid boxes. Simple parameterization methods are used to introduce sub-grid variability of convective clouds and precipitation. In order to constrain uncertainties a comprehensive comparison with sub-grid scale convective precipitation variability which is deduced from TRMM PR observations is carried out.
Structure and modeling of turbulence
DOE Office of Scientific and Technical Information (OSTI.GOV)
Novikov, E.A.
The {open_quotes}vortex strings{close_quotes} scale l{sub s} {approximately} LRe{sup -3/10} (L-external scale, Re - Reynolds number) is suggested as a grid scale for the large-eddy simulation. Various aspects of the structure of turbulence and subgrid modeling are described in terms of conditional averaging, Markov processes with dependent increments and infinitely divisible distributions. The major request from the energy, naval, aerospace and environmental engineering communities to the theory of turbulence is to reduce the enormous number of degrees of freedom in turbulent flows to a level manageable by computer simulations. The vast majority of these degrees of freedom is in the small-scalemore » motion. The study of the structure of turbulence provides a basis for subgrid-scale (SGS) models, which are necessary for the large-eddy simulations (LES).« less
Simplified galaxy formation with mesh-less hydrodynamics
NASA Astrophysics Data System (ADS)
Lupi, Alessandro; Volonteri, Marta; Silk, Joseph
2017-09-01
Numerical simulations have become a necessary tool to describe the complex interactions among the different processes involved in galaxy formation and evolution, unfeasible via an analytic approach. The last decade has seen a great effort by the scientific community in improving the sub-grid physics modelling and the numerical techniques used to make numerical simulations more predictive. Although the recently publicly available code gizmo has proven to be successful in reproducing galaxy properties when coupled with the model of the MUFASA simulations and the more sophisticated prescriptions of the Feedback In Realistic Environment (FIRE) set-up, it has not been tested yet using delayed cooling supernova feedback, which still represent a reasonable approach for large cosmological simulations, for which detailed sub-grid models are prohibitive. In order to limit the computational cost and to be able to resolve the disc structure in the galaxies we perform a suite of zoom-in cosmological simulations with rather low resolution centred around a sub-L* galaxy with a halo mass of 3 × 1011 M⊙ at z = 0, to investigate the ability of this simple model, coupled with the new hydrodynamic method of gizmo, to reproduce observed galaxy scaling relations (stellar to halo mass, stellar and baryonic Tully-Fisher, stellar mass-metallicity and mass-size). We find that the results are in good agreement with the main scaling relations, except for the total stellar mass, larger than that predicted by the abundance matching technique, and the effective sizes for the most massive galaxies in the sample, which are too small.
Evapotranspiration and cloud variability at regional sub-grid scales
NASA Astrophysics Data System (ADS)
Vila-Guerau de Arellano, Jordi; Sikma, Martin; Pedruzo-Bagazgoitia, Xabier; van Heerwaarden, Chiel; Hartogensis, Oscar; Ouwersloot, Huug
2017-04-01
In regional and global models uncertainties arise due to our incomplete understanding of the coupling between biochemical and physical processes. Representing their impact depends on our ability to calculate these processes using physically sound parameterizations, since they are unresolved at scales smaller than the grid size. More specifically over land, the coupling between evapotranspiration, turbulent transport of heat and moisture, and clouds lacks a combined representation to take these sub-grid scales interactions into account. Our approach is based on understanding how radiation, surface exchange, turbulent transport and moist convection are interacting from the leaf- to the cloud scale. We therefore place special emphasis on plant stomatal aperture as the main regulator of CO2-assimilation and water transpiration, a key source of moisture source to the atmosphere. Plant functionality is critically modulated by interactions with atmospheric conditions occurring at very short spatiotemporal scales such as cloud radiation perturbations or water vapour turbulent fluctuations. By explicitly resolving these processes, the LES (large-eddy simulation) technique is enabling us to characterize and better understand the interactions between canopies and the local atmosphere. This includes the adaption time of vegetation to rapid changes in atmospheric conditions driven by turbulence or the presence of cumulus clouds. Our LES experiments are based on explicitly coupling the diurnal atmospheric dynamics to a plant physiology model. Our general hypothesis is that different partitioning of direct and diffuse radiation leads to different responses of the vegetation. As a result there are changes in the water use efficiencies and shifts in the partitioning of sensible and latent heat fluxes under the presence of clouds. Our presentation is as follows. First, we discuss the ability of LES to reproduce the surface energy balance including photosynthesis and CO2 soil respiration coupled to the dynamics of a convective boundary layer. LES results are compared with a complete set of surface and upper-air meteorological and carbon-dioxide observations gathered during a representative day at the 213-meter meteorological tall tower at Cabauw. Second, we perform systematic numerical experiments under a wide range of background wind conditions and stomatal aperture response time. Our analysis unravel how thin clouds, characterized by lower values of the cloud optical depth, have a different impact on evapotranspiration compared to thick clouds due to differences in the partitioning between direct and diffuse radiation at canopy level. Related to this detailed simulation, we discuss how new instrumental techniques, e.g. scintillometery, enable us to obtain new observational insight of the coupling between clouds and vegetation. We will close the presentation with open questions regarding the need to include parameterizations for these interactions at short spatiotemporal scales in regional or climate models.
NASA Technical Reports Server (NTRS)
Xu, Kuan-Man; Cheng, Anning
2007-01-01
The effects of subgrid-scale condensation and transport become more important as the grid spacings increase from those typically used in large-eddy simulation (LES) to those typically used in cloud-resolving models (CRMs). Incorporation of these effects can be achieved by a joint probability density function approach that utilizes higher-order moments of thermodynamic and dynamic variables. This study examines how well shallow cumulus and stratocumulus clouds are simulated by two versions of a CRM that is implemented with low-order and third-order turbulence closures (LOC and TOC) when a typical CRM horizontal resolution is used and what roles the subgrid-scale and resolved-scale processes play as the horizontal grid spacing of the CRM becomes finer. Cumulus clouds were mostly produced through subgrid-scale transport processes while stratocumulus clouds were produced through both subgrid-scale and resolved-scale processes in the TOC version of the CRM when a typical CRM grid spacing is used. The LOC version of the CRM relied upon resolved-scale circulations to produce both cumulus and stratocumulus clouds, due to small subgrid-scale transports. The mean profiles of thermodynamic variables, cloud fraction and liquid water content exhibit significant differences between the two versions of the CRM, with the TOC results agreeing better with the LES than the LOC results. The characteristics, temporal evolution and mean profiles of shallow cumulus and stratocumulus clouds are weakly dependent upon the horizontal grid spacing used in the TOC CRM. However, the ratio of the subgrid-scale to resolved-scale fluxes becomes smaller as the horizontal grid spacing decreases. The subcloud-layer fluxes are mostly due to the resolved scales when a grid spacing less than or equal to 1 km is used. The overall results of the TOC simulations suggest that a 1-km grid spacing is a good choice for CRM simulation of shallow cumulus and stratocumulus.
NASA Astrophysics Data System (ADS)
Haverd, V.; Smith, B.; Nieradzik, L. P.; Briggs, P. R.
2014-02-01
Poorly constrained rates of biomass turnover are a key limitation of Earth system models (ESM). In light of this, we recently proposed a new approach encoded in a model called Populations-Order-Physiology (POP), for the simulation of woody ecosystem stand dynamics, demography and disturbance-mediated heterogeneity. POP is suitable for continental to global applications and designed for coupling to the terrestrial ecosystem component of any ESM. POP bridges the gap between first generation Dynamic Vegetation Models (DVMs) with simple large-area parameterisations of woody biomass (typically used in current ESMs) and complex second generation DVMs, that explicitly simulate demographic processes and landscape heterogeneity of forests. The key simplification in the POP approach, compared with second-generation DVMs, is to compute physiological processes such as assimilation at grid-scale (with CABLE or a similar land surface model), but to partition the grid-scale biomass increment among age classes defined at sub grid-scale, each subject to its own dynamics. POP was successfully demonstrated along a savanna transect in northern Australia, replicating the effects of strong rainfall and fire disturbance gradients on observed stand productivity and structure. Here, we extend the application of POP to a range of forest types around the globe, employing paired observations of stem biomass and density from forest inventory data to calibrate model parameters governing stand demography and biomass evolution. The calibrated POP model is then coupled to the CABLE land surface model and the combined model (CABLE-POP) is evaluated against leaf-stem allometry observations from forest stands ranging in age from 3 to 200 yr. Results indicate that simulated biomass pools conform well with observed allometry. We conclude that POP represents a preferable alternative to large-area parameterisations of woody biomass turnover, typically used in current ESMs.
Evaluation of CASL boiling model for DNB performance in full scale 5x5 fuel bundle with spacer grids
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Seung Jun
As one of main tasks for FY17 CASL-THM activity, Evaluation study on applicability of the CASL baseline boiling model for 5x5 DNB application is conducted and the predictive capability of the DNB analysis is reported here. While the baseline CASL-boiling model (GEN- 1A) approach has been successfully implemented and validated with a single pipe application in the previous year’s task, the extended DNB validation for realistic sub-channels with detailed spacer grid configurations are tasked in FY17. The focus area of the current study is to demonstrate the robustness and feasibility of the CASL baseline boiling model for DNB performance inmore » a full 5x5 fuel bundle application. A quantitative evaluation of the DNB predictive capability is performed by comparing with corresponding experimental measurements (i.e. reference for the model validation). The reference data are provided from the Westinghouse Electricity Company (WEC). Two different grid configurations tested here include Non-Mixing Vane Grid (NMVG), and Mixing Vane Grid (MVG). Thorough validation studies with two sub-channel configurations are performed at a wide range of realistic PWR operational conditions.« less
Scales of variability of black carbon plumes and their dependence on resolution of ECHAM6-HAM
NASA Astrophysics Data System (ADS)
Weigum, Natalie; Stier, Philip; Schutgens, Nick; Kipling, Zak
2015-04-01
Prediction of the aerosol effect on climate depends on the ability of three-dimensional numerical models to accurately estimate aerosol properties. However, a limitation of traditional grid-based models is their inability to resolve variability on scales smaller than a grid box. Past research has shown that significant aerosol variability exists on scales smaller than these grid-boxes, which can lead to discrepancies between observations and aerosol models. The aim of this study is to understand how a global climate model's (GCM) inability to resolve sub-grid scale variability affects simulations of important aerosol features. This problem is addressed by comparing observed black carbon (BC) plume scales from the HIPPO aircraft campaign to those simulated by ECHAM-HAM GCM, and testing how model resolution affects these scales. This study additionally investigates how model resolution affects BC variability in remote and near-source regions. These issues are examined using three different approaches: comparison of observed and simulated along-flight-track plume scales, two-dimensional autocorrelation analysis, and 3-dimensional plume analysis. We find that the degree to which GCMs resolve variability can have a significant impact on the scales of BC plumes, and it is important for models to capture the scales of aerosol plume structures, which account for a large degree of aerosol variability. In this presentation, we will provide further results from the three analysis techniques along with a summary of the implication of these results on future aerosol model development.
Validation of the RegCM4-Subgrid module for the high resolution climate simulation over Korea
NASA Astrophysics Data System (ADS)
Lee, C.; Im, E.; Chang, K.; Choi, Y.
2010-12-01
Given the discernable evidences of climate changes due to human activity, there is a growing demand for the reliable climate change scenario in response to future emission forcing. One of the most significant impacts of climate changes can be that on the hydrological process. Changes in the seasonality and the low and high rainfall extremes can influence the water balance of river basin, with several consequences for societies and ecosystems. In fact, recent studies have reported that East Asia including the Korean peninsula is regarded to be a highly vulnerability region under global warming, especially for water resources. As an attempt to accurately assess the impact of climate change over Korea, we developed the dynamical downscaling system using the RegCM4 with a mosaic-type parameterization of subgrid-scale topography and land use (Sub-BATS). The Sub-BATS system is composed of 20 km coarse-grid cell and 4 km sub-grid cell. Before a full climate change simulation is carried out, we performed the simulation spanning the 19-year periods (1989-2007) with the lateral boundary fields obtained from the ERA-Interim reanalysis. The Korean peninsula is characterized by narrow mountain systems surrounded by ocean, and covered by a relatively dense observational network (approximate 400 stations), which provides an excellent dataset to validate a finescale downscaled results over the region. The evaluation of simulated surface variables (e.g. temperature, precipitation, snow, runoff) shows the usefulness of the RegCM4-Subgrid module as a tool to produce fine scale climate information of surface processes for coupling with hydrological model over the Korean peninsula Acknowledgements This work was supported by the Korea Science and Engineering Foundation (KOSEF) grant funded by the Korea government(MEST) (No. 2009-0085533), and by the "Advanced research on industrial meteorology" and " Development of meteorological resources for green growth." of National Institute of Meteorological Research (NIMR), funded by the Korea Meteorological Administration(KMA).
The integrated process rates (IPR) estimated by the Eta-CMAQ model at grid cells along the trajectory of the air mass transport path were analyzed to quantitatively investigate the relative importance of physical and chemical processes for O3 formation and evolution ov...
Using a Virtual Experiment to Analyze Infiltration Process from Point to Grid-cell Size Scale
NASA Astrophysics Data System (ADS)
Barrios, M. I.
2013-12-01
The hydrological science requires the emergence of a consistent theoretical corpus driving the relationships between dominant physical processes at different spatial and temporal scales. However, the strong spatial heterogeneities and non-linearities of these processes make difficult the development of multiscale conceptualizations. Therefore, scaling understanding is a key issue to advance this science. This work is focused on the use of virtual experiments to address the scaling of vertical infiltration from a physically based model at point scale to a simplified physically meaningful modeling approach at grid-cell scale. Numerical simulations have the advantage of deal with a wide range of boundary and initial conditions against field experimentation. The aim of the work was to show the utility of numerical simulations to discover relationships between the hydrological parameters at both scales, and to use this synthetic experience as a media to teach the complex nature of this hydrological process. The Green-Ampt model was used to represent vertical infiltration at point scale; and a conceptual storage model was employed to simulate the infiltration process at the grid-cell scale. Lognormal and beta probability distribution functions were assumed to represent the heterogeneity of soil hydraulic parameters at point scale. The linkages between point scale parameters and the grid-cell scale parameters were established by inverse simulations based on the mass balance equation and the averaging of the flow at the point scale. Results have shown numerical stability issues for particular conditions and have revealed the complex nature of the non-linear relationships between models' parameters at both scales and indicate that the parameterization of point scale processes at the coarser scale is governed by the amplification of non-linear effects. The findings of these simulations have been used by the students to identify potential research questions on scale issues. Moreover, the implementation of this virtual lab improved the ability to understand the rationale of these process and how to transfer the mathematical models to computational representations.
NASA Astrophysics Data System (ADS)
Perez, Adrianna; Moreno, Jorge; Naiman, Jill; Ramirez-Ruiz, Enrico; Hopkins, Philip F.
2017-01-01
In this work, we analyze the environments surrounding star clusters of simulated merging galaxies. Our framework employs Feedback In Realistic Environments (FIRE) model (Hopkins et al., 2014). The FIRE project is a high resolution cosmological simulation that resolves star forming regions and incorporates stellar feedback in a physically realistic way. The project focuses on analyzing the properties of the star clusters formed in merging galaxies. The locations of these star clusters are identified with astrodendro.py, a publicly available dendrogram algorithm. Once star cluster properties are extracted, they will be used to create a sub-grid (smaller than the resolution scale of FIRE) of gas confinement in these clusters. Then, we can examine how the star clusters interact with these available gas reservoirs (either by accreting this mass or blowing it out via feedback), which will determine many properties of the cluster (star formation history, compact object accretion, etc). These simulations will further our understanding of star formation within stellar clusters during galaxy evolution. In the future, we aim to enhance sub-grid prescriptions for feedback specific to processes within star clusters; such as, interaction with stellar winds and gas accretion onto black holes and neutron stars.
The self-organization of grid cells in 3D
Stella, Federico; Treves, Alessandro
2015-01-01
Do we expect periodic grid cells to emerge in bats, or perhaps dolphins, exploring a three-dimensional environment? How long will it take? Our self-organizing model, based on ring-rate adaptation, points at a complex answer. The mathematical analysis leads to asymptotic states resembling face centered cubic (FCC) and hexagonal close packed (HCP) crystal structures, which are calculated to be very close to each other in terms of cost function. The simulation of the full model, however, shows that the approach to such asymptotic states involves several sub-processes over distinct time scales. The smoothing of the initially irregular multiple fields of individual units and their arrangement into hexagonal grids over certain best planes are observed to occur relatively quickly, even in large 3D volumes. The correct mutual orientation of the planes, though, and the coordinated arrangement of different units, take a longer time, with the network showing no sign of convergence towards either a pure FCC or HCP ordering. DOI: http://dx.doi.org/10.7554/eLife.05913.001 PMID:25821989
Effect of particle size distribution on the hydrodynamics of dense CFB risers
NASA Astrophysics Data System (ADS)
Bakshi, Akhilesh; Khanna, Samir; Venuturumilli, Raj; Altantzis, Christos; Ghoniem, Ahmed
2015-11-01
Circulating Fluidized Beds (CFB) are favorable in the energy and chemical industries, due to their high efficiency. While accurate hydrodynamic modeling is essential for optimizing performance, most CFB riser simulations are performed assuming equally-sized solid particles, owing to limited computational resources. Even though this approach yields reasonable predictions, it neglects commonly observed experimental findings suggesting the strong effect of particle size distribution (psd) on the hydrodynamics and chemical conversion. Thus, this study is focused on the inclusion of discrete particle sizes to represent the psd and its effect on fluidization via 2D numerical simulations. The particle sizes and corresponding mass fluxes are obtained using experimental data in dense CFB riser while the modeling framework is described in Bakshi et al 2015. Simulations are conducted at two scales: (a) fine grid to resolve heterogeneous structures and (b) coarse grid using EMMS sub-grid modifications. Using suitable metrics which capture bed dynamics, this study provides insights into segregation and mixing of particles as well as highlights need for improved sub-grid models.
Feng, Sha; Vogelmann, Andrew M.; Li, Zhijin; ...
2015-01-20
Fine-resolution three-dimensional fields have been produced using the Community Gridpoint Statistical Interpolation (GSI) data assimilation system for the U.S. Department of Energy’s Atmospheric Radiation Measurement Program (ARM) Southern Great Plains region. The GSI system is implemented in a multi-scale data assimilation framework using the Weather Research and Forecasting model at a cloud-resolving resolution of 2 km. From the fine-resolution three-dimensional fields, large-scale forcing is derived explicitly at grid-scale resolution; a subgrid-scale dynamic component is derived separately, representing subgrid-scale horizontal dynamic processes. Analyses show that the subgrid-scale dynamic component is often a major component over the large-scale forcing for grid scalesmore » larger than 200 km. The single-column model (SCM) of the Community Atmospheric Model version 5 (CAM5) is used to examine the impact of the grid-scale and subgrid-scale dynamic components on simulated precipitation and cloud fields associated with a mesoscale convective system. It is found that grid-scale size impacts simulated precipitation, resulting in an overestimation for grid scales of about 200 km but an underestimation for smaller grids. The subgrid-scale dynamic component has an appreciable impact on the simulations, suggesting that grid-scale and subgrid-scale dynamic components should be considered in the interpretation of SCM simulations.« less
Distributed intrusion detection system based on grid security model
NASA Astrophysics Data System (ADS)
Su, Jie; Liu, Yahui
2008-03-01
Grid computing has developed rapidly with the development of network technology and it can solve the problem of large-scale complex computing by sharing large-scale computing resource. In grid environment, we can realize a distributed and load balance intrusion detection system. This paper first discusses the security mechanism in grid computing and the function of PKI/CA in the grid security system, then gives the application of grid computing character in the distributed intrusion detection system (IDS) based on Artificial Immune System. Finally, it gives a distributed intrusion detection system based on grid security system that can reduce the processing delay and assure the detection rates.
NASA Technical Reports Server (NTRS)
Sellers, Piers
2012-01-01
Soil wetness typically shows great spatial variability over the length scales of general circulation model (GCM) grid areas (approx 100 km ), and the functions relating evapotranspiration and photosynthetic rate to local-scale (approx 1 m) soil wetness are highly non-linear. Soil respiration is also highly dependent on very small-scale variations in soil wetness. We therefore expect significant inaccuracies whenever we insert a single grid area-average soil wetness value into a function to calculate any of these rates for the grid area. For the particular case of evapotranspiration., this method - use of a grid-averaged soil wetness value - can also provoke severe oscillations in the evapotranspiration rate and soil wetness under some conditions. A method is presented whereby the probability distribution timction(pdf) for soil wetness within a grid area is represented by binning. and numerical integration of the binned pdf is performed to provide a spatially-integrated wetness stress term for the whole grid area, which then permits calculation of grid area fluxes in a single operation. The method is very accurate when 10 or more bins are used, can deal realistically with spatially variable precipitation, conserves moisture exactly and allows for precise modification of the soil wetness pdf after every time step. The method could also be applied to other ecological problems where small-scale processes must be area-integrated, or upscaled, to estimate fluxes over large areas, for example in treatments of the terrestrial carbon budget or trace gas generation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhang, Xuesong
2012-12-17
Precipitation is an important input variable for hydrologic and ecological modeling and analysis. Next Generation Radar (NEXRAD) can provide precipitation products that cover most of the continental United States with a high resolution display of approximately 4 × 4 km2. Two major issues concerning the applications of NEXRAD data are (1) lack of a NEXRAD geo-processing and geo-referencing program and (2) bias correction of NEXRAD estimates. In this chapter, a geographic information system (GIS) based software that can automatically support processing of NEXRAD data for hydrologic and ecological models is presented. Some geostatistical approaches to calibrating NEXRAD data using rainmore » gauge data are introduced, and two case studies on evaluating accuracy of NEXRAD Multisensor Precipitation Estimator (MPE) and calibrating MPE with rain-gauge data are presented. The first case study examines the performance of MPE in mountainous region versus south plains and cold season versus warm season, as well as the effect of sub-grid variability and temporal scale on NEXRAD performance. From the results of the first case study, performance of MPE was found to be influenced by complex terrain, frozen precipitation, sub-grid variability, and temporal scale. Overall, the assessment of MPE indicates the importance of removing bias of the MPE precipitation product before its application, especially in the complex mountainous region. The second case study examines the performance of three MPE calibration methods using rain gauge observations in the Little River Experimental Watershed in Georgia. The comparison results show that no one method can perform better than the others in terms of all evaluation coefficients and for all time steps. For practical estimation of precipitation distribution, implementation of multiple methods to predict spatial precipitation is suggested.« less
NASA Astrophysics Data System (ADS)
Zeng, Jicai; Zha, Yuanyuan; Zhang, Yonggen; Shi, Liangsheng; Zhu, Yan; Yang, Jinzhong
2017-11-01
Multi-scale modeling of the localized groundwater flow problems in a large-scale aquifer has been extensively investigated under the context of cost-benefit controversy. An alternative is to couple the parent and child models with different spatial and temporal scales, which may result in non-trivial sub-model errors in the local areas of interest. Basically, such errors in the child models originate from the deficiency in the coupling methods, as well as from the inadequacy in the spatial and temporal discretizations of the parent and child models. In this study, we investigate the sub-model errors within a generalized one-way coupling scheme given its numerical stability and efficiency, which enables more flexibility in choosing sub-models. To couple the models at different scales, the head solution at parent scale is delivered downward onto the child boundary nodes by means of the spatial and temporal head interpolation approaches. The efficiency of the coupling model is improved either by refining the grid or time step size in the parent and child models, or by carefully locating the sub-model boundary nodes. The temporal truncation errors in the sub-models can be significantly reduced by the adaptive local time-stepping scheme. The generalized one-way coupling scheme is promising to handle the multi-scale groundwater flow problems with complex stresses and heterogeneity.
Inference of turbulence parameters from a ROMS simulation using the k-ε closure scheme
NASA Astrophysics Data System (ADS)
Thyng, Kristen M.; Riley, James J.; Thomson, Jim
2013-12-01
Comparisons between high resolution turbulence data from Admiralty Inlet, WA (USA), and a 65-meter horizontal grid resolution simulation using the hydrostatic ocean modelling code, Regional Ocean Modeling System (ROMS), show that the model's k-ε turbulence closure scheme performs reasonably well. Turbulent dissipation rates and Reynolds stresses agree within a factor of two, on average. Turbulent kinetic energy (TKE) also agrees within a factor of two, but only for motions within the observed inertial sub-range of frequencies (i.e., classic approximately isotropic turbulence). TKE spectra from the observations indicate that there is significant energy at lower frequencies than the inertial sub-range; these scales are not captured by the model closure scheme nor the model grid resolution. To account for scales not present in the model, the inertial sub-range is extrapolated to lower frequencies and then integrated to obtain an inferred, diagnostic total TKE, with improved agreement with the observed total TKE. The realistic behavior of the dissipation rate and Reynolds stress, combined with the adjusted total TKE, imply that ROMS simulations can be used to understand and predict spatial and temporal variations in turbulence. The results are suggested for application to siting tidal current turbines.
Evaluation of decadal hindcasts using satellite simulators
NASA Astrophysics Data System (ADS)
Spangehl, Thomas; Mazurkiewicz, Alex; Schröder, Marc
2013-04-01
The evaluation of dynamical ensemble forecast systems requires a solid validation of basic processes such as the global atmospheric water and energy cycle. The value of any validation approach strongly depends on the quality of the observational data records used. Current approaches utilize in situ measurements, remote sensing data and reanalyses. Related data records are subject to a number of uncertainties and limitations such as representativeness, spatial and temporal resolution and homogeneity. However, recently several climate data records with known and sufficient quality became available. In particular, the satellite data records offer the opportunity to obtain reference information on global scales including the oceans. Here we consider the simulation of satellite radiances from the climate model output enabling an evaluation in the instrument's parameter space to avoid uncertainties stemming from the application of retrieval schemes in order to minimise uncertainties on the reference side. Utilizing the CFMIP Observation Simulator Package (COSP) we develop satellite simulators for the Tropical Rainfall Measuring Mission precipitation radar (TRMM PR) and the Infrared Atmospheric Sounding Interferometer (IASI). The simulators are applied within the MiKlip project funded by BMBF (German Federal Ministry of Education and Research) to evaluate decadal climate predictions performed with the MPI-ESM developed at the Max Planck Institute for Meteorology. While TRMM PR enables the evaluation of the vertical structure of precipitation over tropical and sub-tropical areas, IASI is used to support the global evaluation of clouds and radiation. In a first step the reliability of the developed simulators needs to be explored. The simulation of radiances in the instrument space requires the generation of sub-grid scale variability from the climate model output. Furthermore, assumptions are made to simulate radiances such as, for example, the distribution of different hydrometeor types. Therefore, testing is performed to determine the extent to which the quality of the simulator results depends on the applied methods used to generate sub-grid variability (e.g. sub-grid resolution). Moreover, the sensitivity of results to the choice of different distributions of hydrometeors is explored. The model evaluation is carried out in a statistical manner using histograms of radar reflectivities (TRMM PR) and brightness temperatures (IASI). Finally, methods to deduce data suitable for probabilistic evaluation of decadal hindcasts such as simple indices are discussed.
Influence of lubrication forces in direct numerical simulations of particle-laden flows
NASA Astrophysics Data System (ADS)
Maitri, Rohit; Peters, Frank; Padding, Johan; Kuipers, Hans
2016-11-01
Accurate numerical representation of particle-laden flows is important for fundamental understanding and optimizing the complex processes such as proppant transport in fracking. Liquid-solid flows are fundamentally different from gas-solid flows because of lower density ratios (solid to fluid) and non-negligible lubrication forces. In this interface resolved model, fluid-solid coupling is achieved by incorporating the no-slip boundary condition implicitly at particle's surfaces by means of an efficient second order ghost-cell immersed boundary method. A fixed Eulerian grid is used for solving the Navier-Stokes equations and the particle-particle interactions are implemented using the soft sphere collision and sub-grid scale lubrication model. Due to the range of influence of lubrication force on a smaller scale than the grid size, it is important to implement the lubrication model accurately. In this work, different implementations of the lubrication model on particle dynamics are studied for various flow conditions. The effect of a particle surface roughness on lubrication force and the particle transport is also investigated. This study is aimed at developing a validated methodology to incorporate lubrication models in direct numerical simulation of particle laden flows. This research is supported from Grant 13CSER014 of the Foundation for Fundamental Research on Matter (FOM), which is part of the Netherlands Organisation for Scientific Research (NWO).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Allen, Melissa R.; Aziz, H. M. Abdul; Coletti, Mark A.
Changing human activity within a geographical location may have significant influence on the global climate, but that activity must be parameterized in such a way as to allow these high-resolution sub-grid processes to affect global climate within that modeling framework. Additionally, we must have tools that provide decision support and inform local and regional policies regarding mitigation of and adaptation to climate change. The development of next-generation earth system models, that can produce actionable results with minimum uncertainties, depends on understanding global climate change and human activity interactions at policy implementation scales. Unfortunately, at best we currently have only limitedmore » schemes for relating high-resolution sectoral emissions to real-time weather, ultimately to become part of larger regions and well-mixed atmosphere. Moreover, even our understanding of meteorological processes at these scales is imperfect. This workshop addresses these shortcomings by providing a forum for discussion of what we know about these processes, what we can model, where we have gaps in these areas and how we can rise to the challenge to fill these gaps.« less
Assessment of zero-equation SGS models for simulating indoor environment
NASA Astrophysics Data System (ADS)
Taghinia, Javad; Rahman, Md Mizanur; Tse, Tim K. T.
2016-12-01
The understanding of air-flow in enclosed spaces plays a key role to designing ventilation systems and indoor environment. The computational fluid dynamics aspects dictate that the large eddy simulation (LES) offers a subtle means to analyze complex flows with recirculation and streamline curvature effects, providing more robust and accurate details than those of Reynolds-averaged Navier-Stokes simulations. This work assesses the performance of two zero-equation sub-grid scale models: the Rahman-Agarwal-Siikonen-Taghinia (RAST) model with a single grid-filter and the dynamic Smagorinsky model with grid-filter and test-filter scales. This in turn allows a cross-comparison of the effect of two different LES methods in simulating indoor air-flows with forced and mixed (natural + forced) convection. A better performance against experiments is indicated with the RAST model in wall-bounded non-equilibrium indoor air-flows; this is due to its sensitivity toward both the shear and vorticity parameters.
Sub-kilometer Numerical Weather Prediction in complex urban areas
NASA Astrophysics Data System (ADS)
Leroyer, S.; Bélair, S.; Husain, S.; Vionnet, V.
2013-12-01
A Sub-kilometer atmospheric modeling system with grid-spacings of 2.5 km, 1 km and 250 m and including urban processes is currently being developed at the Meteorological Service of Canada (MSC) in order to provide more accurate weather forecasts at the city scale. Atmospheric lateral boundary conditions are provided with the 15-km Canadian Regional Deterministic Prediction System (RDPS). Surface physical processes are represented with the Town Energy Balance (TEB) model for the built-up covers and with the Interactions between the Surface, Biosphere, and Atmosphere (ISBA) land surface model for the natural covers. In this study, several research experiments over large metropolitan areas and using observational networks at the urban scale are presented, with a special emphasis on the representation of local atmospheric circulations and their impact on extreme weather forecasting. First, numerical simulations are performed over the Vancouver metropolitan area during a summertime Intense Observing Period (IOP of 14-15 August 2008) of the Environmental Prediction in Canadian Cities (EPiCC) observational network. The influence of the horizontal resolution on the fine-scale representation of the sea-breeze development over the city is highlighted (Leroyer et al., 2013). Then severe storms cases occurring in summertime within the Greater Toronto Area (GTA) are simulated. In view of supporting the 2015 PanAmerican and Para-Pan games to be hold in GTA, a dense observational network has been recently deployed over this region to support model evaluations at the urban and meso scales. In particular, simulations are conducted for the case of 8 July 2013 when exceptional rainfalls were recorded. Leroyer, S., S. Bélair, J. Mailhot, S.Z. Husain, 2013: Sub-kilometer Numerical Weather Prediction in an Urban Coastal Area: A case study over the Vancouver Metropolitan Area, submitted to Journal of Applied Meteorology and Climatology.
Simulation of the planetary boundary layer with the UCLA general circulation model
NASA Technical Reports Server (NTRS)
Suarez, M. J.; Arakawa, A.; Randall, D. A.
1981-01-01
A planetary boundary layer (PBL) model is presented which employs a mixed layer entrainment formulation to describe the mass exchange between the mixed layer with the upper, laminar atmosphere. A modified coordinate system couples the mixed layer model with large scale and sub-grid scale processes of a general circulation model. The vertical coordinate is configured as a sigma coordinate with the lower boundary, the top of the PBL, and the prescribed pressure level near the tropopause expressed as coordinate surfaces. The entrainment mass flux is parameterized by assuming the dissipation rate of turbulent kinetic energy to be proportional to the positive part of the generation by convection or mechanical production. The results of a simulation of July are presented for the entire globe.
On the use of MODIS and TRMM products to simulate hydrological processes in the La Plata Basin
NASA Astrophysics Data System (ADS)
Saavedra Valeriano, O. C.; Koike, T.; Berbery, E. H.
2009-12-01
La Plata basin is targeted to establish a distributed water-energy balance model using NASA and JAXA satellite products to estimate fluxes like the river discharge at sub-basin scales. The coupled model is called the Water and Energy Budget-based Distributed Hydrological Model (WEB-DHM), already tested with success in the Little Washita basin, Oklahoma, and the upper Tone River in Japan. The model demonstrated the ability to reproduce point-scale energy fluxes, CO2 flux, and river discharges. Moreover, the model showed the ability to predict the basin-scale surface soil moisture evolution in a spatially distributed fashion. In the context of the La Plata Basin, the first step was to set-up the water balance component of the distributed hydrological model of the entire basin using available global geographical data sets. The geomorphology of the basin was extracted using 1-km DEM resolution (obtained from EROS, Hydro 1K). The total delineated watershed reached 3.246 millions km2, identifying 145 sub-basins with a computing grid of 10-km. The distribution of land cover, land surface temperature, LAI and FPAR were obtained from MODIS products. In a first instance, the model was forced by gridded rainfall from the Climate Prediction Center (derived from available rain gauges) and satellite precipitation from TRMM 3B42 (NASA & JAXA). The simulated river discharge using both sources of data was compared and the overall low flow and normal peaks were identified. It was found that the extreme peaks tend to be overestimated when using TRMM 3B42. However, TRMM data allows tracking rainfall patterns which might be missed by the sparse distribution of rain gauges over some areas of the basin.
NARSTO critical review of photochemical models and modeling
NASA Astrophysics Data System (ADS)
Russell, Armistead; Dennis, Robin
Photochemical air quality models play a central role in both schentific investigation of how pollutants evlove in the atmosphere as well as developing policies to manage air quality. In the past 30 years, these models have evolved from rather crude representations of the physics and chemistry impacting trace species to their current state: comprehensive, but not complete. The evolution has included advancements in not only the level of process descriptions, but also the computational implementation, including numerical methods. As part of the NARSTO Critical Reviews, this article discusses the current strengths and weaknesses of air quality models and the modeling process. Current Eulerian models are found to represent well the primary processes impacting the evolution of trace species in most cases though some exceptions may exist. For example, sub-grid-scale processes, such as concentrated power plant plumes, are treated only approximately. It is not apparent how much such approximations affect their results and the polices based upon those results. A significant weakness has been in how investigators have addressed, and communicated, such uncertainties. Studies find that major uncertainties are due to model inputs, e.g., emissions and meteorology, more so than the model itself. One of the primary weakness identified is in the modeling process, not the models. Evaluation has been limited both due to data constraints. Seldom is there ample observational data to conduct a detailed model intercomparison using consistent data (e.g., the same emissions and meteorology). Further model advancement, and development of greater confidence in the use of models, is hampered by the lack of thorough evaluation and intercomparisons. Model advances are seen in the use of new tools for extending the interpretation of model results, e.g., process and sensitivity analysis, modeling systems to facilitate their use, and extension of model capabilities, e.g., aerosol dynamics capabilities and sub-grid-scale representations. Another possible direction that is the development and widespread use of a community model acting as a platform for multiple groups and agencies to collaborate and progress more rapidly.
NASA Astrophysics Data System (ADS)
Long, D.; Scanlon, B. R.; Longuevergne, L.; Chen, X.
2015-12-01
Increasing interest in use of GRACE satellites and a variety of new products to monitor changes in total water storage (TWS) underscores the need to assess the reliability of output from different products. The objective of this study was to assess skills and uncertainties of different approaches for processing GRACE data to restore signal losses caused by spatial filtering based on analysis of 1°×1° grid scale data and basin scale data in 60 river basins globally. Results indicate that scaling factors from six land surface models (LSMs), including four models from GLDAS-1 (Noah 2.7, Mosaic, VIC, and CLM 2.0), CLM 4.0, and WGHM, are similar over most humid, sub-humid, and high-latitude regions but can differ by up to 100% over arid and semi-arid basins and areas with intensive irrigation. Large differences in TWS anomalies from three processing approaches (scaling factor, additive, and multiplicative corrections) were found in arid and semi-arid regions, areas with intensive irrigation, and relatively small basins (e.g., ≤ 200,000 km2). Furthermore, TWS anomaly products from gridded data with CLM4.0 scaling factors and the additive correction approach more closely agree with WGHM output than the multiplicative correction approach. Estimation of groundwater storage changes using GRACE satellites requires caution in selecting an appropriate approach for restoring TWS changes. A priori ground-based data used in forward modeling can provide a powerful tool for explaining the distribution of signal gains or losses caused by low-pass filtering in specific regions of interest and should be very useful for more reliable estimation of groundwater storage changes using GRACE satellites.
A machine learning approach for efficient uncertainty quantification using multiscale methods
NASA Astrophysics Data System (ADS)
Chan, Shing; Elsheikh, Ahmed H.
2018-02-01
Several multiscale methods account for sub-grid scale features using coarse scale basis functions. For example, in the Multiscale Finite Volume method the coarse scale basis functions are obtained by solving a set of local problems over dual-grid cells. We introduce a data-driven approach for the estimation of these coarse scale basis functions. Specifically, we employ a neural network predictor fitted using a set of solution samples from which it learns to generate subsequent basis functions at a lower computational cost than solving the local problems. The computational advantage of this approach is realized for uncertainty quantification tasks where a large number of realizations has to be evaluated. We attribute the ability to learn these basis functions to the modularity of the local problems and the redundancy of the permeability patches between samples. The proposed method is evaluated on elliptic problems yielding very promising results.
Aspects on HTS applications in confined power grids
NASA Astrophysics Data System (ADS)
Arndt, T.; Grundmann, J.; Kuhnert, A.; Kummeth, P.; Nick, W.; Oomen, M.; Schacherer, C.; Schmidt, W.
2014-12-01
In an increasing number of electric power grids the share of distributed energy generation is also increasing. The grids have to cope with a considerable change of power flow, which has an impact on the optimum topology of the grids and sub-grids (high-voltage, medium-voltage and low-voltage sub-grids) and the size of quasi-autonomous grid sections. Furthermore the stability of grids is influenced by its size. Thus special benefits of HTS applications in the power grid might become most visible in confined power grids.
NASA Astrophysics Data System (ADS)
Cortinez, J. M.; Valocchi, A. J.; Herrera, P. A.
2013-12-01
Because of the finite size of numerical grids, it is very difficult to correctly account for processes that occur at different spatial scales to accurately simulate the migration of conservative and reactive compounds dissolved in groundwater. In one hand, transport processes in heterogeneous porous media are controlled by local-scale dispersion associated to transport processes at the pore-scale. On the other hand, variations of velocity at the continuum- or Darcy-scale produce spreading of the contaminant plume, which is referred to as macro-dispersion. Furthermore, under some conditions both effects interact, so that spreading may enhance the action of local-scale dispersion resulting in higher mixing, dilution and reaction rates. Traditionally, transport processes at different spatial scales have been included in numerical simulations by using a single dispersion coefficient. This approach implicitly assumes that the separate effects of local-dispersion and macro-dispersion can be added and represented by a unique effective dispersion coefficient. Moreover, the selection of the effective dispersion coefficient for numerical simulations usually do not consider the filtering effect of the grid size over the small-scale flow features. We have developed a multi-scale Lagragian numerical method that allows using two different dispersion coefficients to represent local- and macro-scale dispersion. This technique considers fluid particles that carry solute mass and whose locations evolve according to a deterministic component given by the grid-scale velocity and a stochastic component that corresponds to a block-effective macro-dispersion coefficient. Mass transfer between particles due to local-scale dispersion is approximated by a meshless method. We use our model to test under which transport conditions the combined effect of local- and macro-dispersion are additive and can be represented by a single effective dispersion coefficient. We also demonstrate that for the situations where both processes are additive, an effective grid-dependent dispersion coefficient can be derived based on the concept of block-effective dispersion. We show that the proposed effective dispersion coefficient is able to reproduce dilution, mixing and reaction rates for a wide range of transport conditions similar to the ones found in many practical applications.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Riebel, D.; Meixner, M.; Srinivasan, S.
We present results from the first application of the Grid of Red Supergiant and Asymptotic Giant Branch ModelS (GRAMS) model grid to the entire evolved stellar population of the Large Magellanic Cloud (LMC). GRAMS is a pre-computed grid of 80,843 radiative transfer models of evolved stars and circumstellar dust shells composed of either silicate or carbonaceous dust. We fit GRAMS models to {approx}30,000 asymptotic giant branch (AGB) and red supergiant (RSG) stars in the LMC, using 12 bands of photometry from the optical to the mid-infrared. Our published data set consists of thousands of evolved stars with individually determined evolutionarymore » parameters such as luminosity and mass-loss rate. The GRAMS grid has a greater than 80% accuracy rate discriminating between oxygen- and carbon-rich chemistry. The global dust injection rate to the interstellar medium (ISM) of the LMC from RSGs and AGB stars is on the order of 2.1 Multiplication-Sign 10{sup -5} M{sub Sun} yr{sup -1}, equivalent to a total mass injection rate (including the gas) into the ISM of {approx}6 Multiplication-Sign 10{sup -3} M{sub Sun} yr{sup -1}. Carbon stars inject two and a half times as much dust into the ISM as do O-rich AGB stars, but the same amount of mass. We determine a bolometric correction factor for C-rich AGB stars in the K{sub s} band as a function of J - K{sub s} color, BC{sub K{sub s}}= -0.40(J-K{sub s}){sup 2} + 1.83(J-K{sub s}) + 1.29. We determine several IR color proxies for the dust mass-loss rate (M-dot{sub d}) from C-rich AGB stars, such as log M-dot{sub d} = (-18.90/((K{sub s}-[8.0])+3.37) - 5.93. We find that a larger fraction of AGB stars exhibiting the 'long-secondary period' phenomenon are more O-rich than stars dominated by radial pulsations, and AGB stars without detectable mass loss do not appear on either the first-overtone or fundamental-mode pulsation sequences.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sarkar, Avik; Milioli, Fernando E.; Ozarkar, Shailesh
2016-10-01
The accuracy of fluidized-bed CFD predictions using the two-fluid model can be improved significantly, even when using coarse grids, by replacing the microscopic kinetic-theory-based closures with coarse-grained constitutive models. These coarse-grained constitutive relationships, called filtered models, account for the unresolved gas-particle structures (clusters and bubbles) via sub-grid corrections. Following the previous 2-D approaches of Igci et al. [AIChE J., 54(6), 1431-1448, 2008] and Milioli et al. [AIChE J., 59(9), 3265-3275, 2013], new filtered models are constructed from highly-resolved 3-D simulations of gas-particle flows. Although qualitatively similar to the older 2-D models, the new 3-D relationships exhibit noticeable quantitative and functionalmore » differences. In particular, the filtered stresses are strongly dependent on the gas-particle slip velocity. Closures for the filtered inter-phase drag, gas- and solids-phase pressures and viscosities are reported. A new model for solids stress anisotropy is also presented. These new filtered 3-D constitutive relationships are better suited to practical coarse-grid 3-D simulations of large, commercial-scale devices.« less
Multi-dimensional upwinding-based implicit LES for the vorticity transport equations
NASA Astrophysics Data System (ADS)
Foti, Daniel; Duraisamy, Karthik
2017-11-01
Complex turbulent flows such as rotorcraft and wind turbine wakes are characterized by the presence of strong coherent structures that can be compactly described by vorticity variables. The vorticity-velocity formulation of the incompressible Navier-Stokes equations is employed to increase numerical efficiency. Compared to the traditional velocity-pressure formulation, high order numerical methods and sub-grid scale models for the vorticity transport equation (VTE) have not been fully investigated. Consistent treatment of the convection and stretching terms also needs to be addressed. Our belief is that, by carefully designing sharp gradient-capturing numerical schemes, coherent structures can be more efficiently captured using the vorticity-velocity formulation. In this work, a multidimensional upwind approach for the VTE is developed using the generalized Riemann problem-based scheme devised by Parish et al. (Computers & Fluids, 2016). The algorithm obtains high resolution by augmenting the upwind fluxes with transverse and normal direction corrections. The approach is investigated with several canonical vortex-dominated flows including isolated and interacting vortices and turbulent flows. The capability of the technique to represent sub-grid scale effects is also assessed. Navy contract titled ``Turbulence Modelling Across Disparate Length Scales for Naval Computational Fluid Dynamics Applications,'' through Continuum Dynamics, Inc.
Wafer-scale aluminum nano-plasmonics
NASA Astrophysics Data System (ADS)
George, Matthew C.; Nielson, Stew; Petrova, Rumyana; Frasier, James; Gardner, Eric
2014-09-01
The design, characterization, and optical modeling of aluminum nano-hole arrays are discussed for potential applications in surface plasmon resonance (SPR) sensing, surface-enhanced Raman scattering (SERS), and surface-enhanced fluorescence spectroscopy (SEFS). In addition, recently-commercialized work on narrow-band, cloaked wire grid polarizers composed of nano-stacked metal and dielectric layers patterned over 200 mm diameter wafers for projection display applications is reviewed. The stacked sub-wavelength nanowire grid results in a narrow-band reduction in reflectance by 1-2 orders of magnitude, which can be tuned throughout the visible spectrum for stray light control.
NASA Astrophysics Data System (ADS)
Sivia, S. G.; Gheusi, F.; Mari, C.; Di Muro, A.
2015-05-01
In mesoscale models (resolution ~ 1 km) used for regional dispersion of pollution plumes the volcanic heat sources and emissions of gases and aerosols, as well as the induced atmospheric convective motions, are all sub-grid-scale processes (mostly true for weak effusive eruptions) which need to be parameterised. We propose a modified formulation of the EDMF scheme (eddy diffusivity/mass flux) proposed by Pergaud et al. (2009) which is based on a single sub-grid updraft model. It is used to represent volcano induced updrafts tested for a case study of the January 2010 summit eruption of Piton de la Fournaise (PdF) volcano. The validation of this modified formulation using a reference large eddy simulation (LES) focuses on the ability of the model to transport tracer concentrations up to 1-2 km above the ground in the lower troposphere as is the case of majority of PdF eruptions. The modelled volcanic plume agrees reasonably with the profiles of SO2 (sulfur dioxide) tracer concentrations and specific humidity found from the reference LES. Sensitivity tests performed for the modified formulation of the EDMF scheme emphasise the sensitivity of the parameterisation to ambient fresh air entrainment at the plume base.
Modeling water exchange and contaminant transport through a Baltic coastal region.
Engqvist, Anders; Döös, Kristofer; Andrejev, Oleg
2006-12-01
The water exchange of the Baltic coastal zone is characterized by its seasonally varying regimes. In the safety assessment of a potential repository for spent nuclear fuel, it is important to assess the consequences of a hypothetical leak of radionuclides through the seabed into a waterborne transport phase. In particular, estimates of the associated residence times in the near-shore coastal zone are of interest. There are several methods to quantify such measures, of which three are presented here. Using the coastal location of Forsmark (Sweden) as an example, methods based on passive tracers, particle trajectories, and the average age distribution of exogeneous water parcels are compared for a representative one-year cycle. Tracer-based methods can simulate diffusivity more realistically than the other methods. Trajectory-based methods can handle Lagrangian dispersion processes due to advection but neglect diffusion on the sub-grid scale. The method based on the concept of average age (AvA) of exogeneous water can include all such sources simultaneously not only boundary water bodies but also various (fresh)-water discharges. Due to the inclusion of sub-grid diffusion this method gives a smoother measure of the water renewal. It is shown that backward in time trajectories and AvA-times are basically equipollent methods, yielding correlated results within the limits set by the diffusivity.
NASA Technical Reports Server (NTRS)
Taylor, Gregory E.; Zack, John W.; Manobianco, John
1994-01-01
NASA funded Mesoscale Environmental Simulations and Operations (MESO), Inc. to develop a version of the Mesoscale Atmospheric Simulation System (MASS). The model has been modified specifically for short-range forecasting in the vicinity of KSC/CCAS. To accomplish this, the model domain has been limited to increase the number of horizontal grid points (and therefore grid resolution) and the model' s treatment of precipitation, radiation, and surface hydrology physics has been enhanced to predict convection forced by local variations in surface heat, moisture fluxes, and cloud shading. The objective of this paper is to (1) provide an overview of MASS including the real-time initialization and configuration for running the data pre-processor and model, and (2) to summarize the preliminary evaluation of the model's forecasts of temperature, moisture, and wind at selected rawinsonde station locations during February 1994 and July 1994. MASS is a hydrostatic, three-dimensional modeling system which includes schemes to represent planetary boundary layer processes, surface energy and moisture budgets, free atmospheric long and short wave radiation, cloud microphysics, and sub-grid scale moist convection.
Benefits Analysis of Smart Grid Projects. White paper, 2014-2016
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marnay, Chris; Liu, Liping; Yu, JianCheng
Smart grids are rolling out internationally, with the United States (U.S.) nearing completion of a significant USD4-plus-billion federal program funded under the American Recovery and Reinvestment Act (ARRA-2009). The emergence of smart grids is widespread across developed countries. Multiple approaches to analyzing the benefits of smart grids have emerged. The goals of this white paper are to review these approaches and analyze examples of each to highlight their differences, advantages, and disadvantages. This work was conducted under the auspices of a joint U.S.-China research effort, the Climate Change Working Group (CCWG) Implementation Plan, Smart Grid. We present comparative benefits assessmentsmore » (BAs) of smart grid demonstrations in the U.S. and China along with a BA of a pilot project in Europe. In the U.S., we assess projects at two sites: (1) the University of California, Irvine campus (UCI), which consists of two distinct demonstrations: Southern California Edison’s (SCE) Irvine Smart Grid Demonstration Project (ISGD) and the UCI campus itself; and (2) the Navy Yard (TNY) area in Philadelphia, which has been repurposed as a mixed commercial-industrial, and possibly residential, development. In China, we cover several smart-grid aspects of the Sino-Singapore Tianjin Eco-city (TEC) and the Shenzhen Bay Technology and Ecology City (B-TEC). In Europe, we look at a BA of a pilot smart grid project in the Malagrotta area west of Rome, Italy, contributed by the Joint Research Centre (JRC) of the European Commission. The Irvine sub-project BAs use the U.S. Department of Energy (U.S. DOE) Smart Grid Computational Tool (SGCT), which is built on methods developed by the Electric Power Research Institute (EPRI). The TEC sub-project BAs apply Smart Grid Multi-Criteria Analysis (SG-MCA) developed by the State Grid Corporation of China (SGCC) based on the analytic hierarchy process (AHP) with fuzzy logic. The B-TEC and TNY sub-project BAs are evaluated using new approaches developed by those project teams. JRC has adopted an approach similar to EPRI’s but tailored to the Malagrotta distribution grid.« less
Conceptual Design of the Everglades Depth Estimation Network (EDEN) Grid
Jones, John W.; Price, Susan D.
2007-01-01
INTRODUCTION The Everglades Depth Estimation Network (EDEN) offers a consistent and documented dataset that can be used to guide large-scale field operations, to integrate hydrologic and ecological responses, and to support biological and ecological assessments that measure ecosystem responses to the Comprehensive Everglades Restoration Plan (Telis, 2006). Ground elevation data for the greater Everglades and the digital ground elevation models derived from them form the foundation for all EDEN water depth and associated ecologic/hydrologic modeling (Jones, 2004, Jones and Price, 2007). To use EDEN water depth and duration information most effectively, it is important to be able to view and manipulate information on elevation data quality and other land cover and habitat characteristics across the Everglades region. These requirements led to the development of the geographic data layer described in this techniques and methods report. Relying on extensive experience in GIS data development, distribution, and analysis, a great deal of forethought went into the design of the geographic data layer used to index elevation and other surface characteristics for the Greater Everglades region. To allow for simplicity of design and use, the EDEN area was broken into a large number of equal-sized rectangles ('Cells') that in total are referred to here as the 'grid'. Some characteristics of this grid, such as the size of its cells, its origin, the area of Florida it is designed to represent, and individual grid cell identifiers, could not be changed once the grid database was developed. Therefore, these characteristics were selected to design as robust a grid as possible and to ensure the grid's long-term utility. It is desirable to include all pertinent information known about elevation and elevation data collection as grid attributes. Also, it is very important to allow for efficient grid post-processing, sub-setting, analysis, and distribution. This document details the conceptual design of the EDEN grid spatial parameters and cell attribute-table content.
NASA Astrophysics Data System (ADS)
Torkelson, G. Q.; Stoll, R., II
2017-12-01
Large Eddy Simulation (LES) is a tool commonly used to study the turbulent transport of momentum, heat, and moisture in the Atmospheric Boundary Layer (ABL). For a wide range of ABL LES applications, representing the full range of turbulent length scales in the flow field is a challenge. This is an acute problem in regions of the ABL with strong velocity or scalar gradients, which are typically poorly resolved by standard computational grids (e.g., near the ground surface, in the entrainment zone). Most efforts to address this problem have focused on advanced sub-grid scale (SGS) turbulence model development, or on the use of massive computational resources. While some work exists using embedded meshes, very little has been done on the use of grid refinement. Here, we explore the benefits of grid refinement in a pseudo-spectral LES numerical code. The code utilizes both uniform refinement of the grid in horizontal directions, and stretching of the grid in the vertical direction. Combining the two techniques allows us to refine areas of the flow while maintaining an acceptable grid aspect ratio. In tests that used only refinement of the vertical grid spacing, large grid aspect ratios were found to cause a significant unphysical spike in the stream-wise velocity variance near the ground surface. This was especially problematic in simulations of stably-stratified ABL flows. The use of advanced SGS models was not sufficient to alleviate this issue. The new refinement technique is evaluated using a series of idealized simulation test cases of neutrally and stably stratified ABLs. These test cases illustrate the ability of grid refinement to increase computational efficiency without loss in the representation of statistical features of the flow field.
NASA Astrophysics Data System (ADS)
Lee, J.; Zhang, Y.; Klein, S. A.
2017-12-01
The triggering of the land breeze, and hence the development of deep convection over heterogeneous land should be understood as a consequence of the complex processes involving various factors from land surface and atmosphere simultaneously. That is a sub-grid scale process that many large-scale models have difficulty incorporating it into the parameterization scheme partly due to lack of our understanding. Thus, it is imperative that we approach the problem using a high-resolution modeling framework. In this study, we use SAM-SLM (Lee and Khairoutdinov, 2015), a large-eddy simulation model coupled to a land model, to explore the cloud effect such as cold pool, the cloud shading and the soil moisture memory on the land breeze structure and the further development of cloud and precipitation over a heterogeneous land surface. The atmospheric large scale forcing and the initial sounding are taken from the new composite case study of the fair-weather, non-precipitating shallow cumuli at ARM SGP (Zhang et al., 2017). We model the land surface as a chess board pattern with alternating leaf area index (LAI). The patch contrast of the LAI is adjusted to encompass the weak to strong heterogeneity amplitude. The surface sensible- and latent heat fluxes are computed according to the given LAI representing the differential surface heating over a heterogeneous land surface. Separate from the surface forcing imposed from the originally modeled surface, the cases that transition into the moist convection can induce another layer of the surface heterogeneity from the 1) radiation shading by clouds, 2) adjusted soil moisture pattern by the rain, 3) spreading cold pool. First, we assess and quantifies the individual cloud effect on the land breeze and the moist convection under the weak wind to simplify the feedback processes. And then, the same set of experiments is repeated under sheared background wind with low level jet, a typical summer time wind pattern at ARM SGP site, to account for more realistic situations. Our goal is to assist answering the question: "Do the sub-grid scale land surface heterogeneity matter for the weather and climate modeling?" This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344. LLNL-ABS- 736011.
Calculations of High-Temperature Jet Flow Using Hybrid Reynolds-Average Navier-Stokes Formulations
NASA Technical Reports Server (NTRS)
Abdol-Hamid, Khaled S.; Elmiligui, Alaa; Giriamaji, Sharath S.
2008-01-01
Two multiscale-type turbulence models are implemented in the PAB3D solver. The models are based on modifying the Reynolds-averaged Navier Stokes equations. The first scheme is a hybrid Reynolds-averaged- Navier Stokes/large-eddy-simulation model using the two-equation k(epsilon) model with a Reynolds-averaged-Navier Stokes/large-eddy-simulation transition function dependent on grid spacing and the computed turbulence length scale. The second scheme is a modified version of the partially averaged Navier Stokes model in which the unresolved kinetic energy parameter f(sub k) is allowed to vary as a function of grid spacing and the turbulence length scale. This parameter is estimated based on a novel two-stage procedure to efficiently estimate the level of scale resolution possible for a given flow on a given grid for partially averaged Navier Stokes. It has been found that the prescribed scale resolution can play a major role in obtaining accurate flow solutions. The parameter f(sub k) varies between zero and one and is equal to one in the viscous sublayer and when the Reynolds-averaged Navier Stokes turbulent viscosity becomes smaller than the large-eddy-simulation viscosity. The formulation, usage methodology, and validation examples are presented to demonstrate the enhancement of PAB3D's time-accurate turbulence modeling capabilities. The accurate simulations of flow and turbulent quantities will provide a valuable tool for accurate jet noise predictions. Solutions from these models are compared with Reynolds-averaged Navier Stokes results and experimental data for high-temperature jet flows. The current results show promise for the capability of hybrid Reynolds-averaged Navier Stokes and large eddy simulation and partially averaged Navier Stokes in simulating such flow phenomena.
NASA Astrophysics Data System (ADS)
Haverd, V.; Smith, B.; Nieradzik, L. P.; Briggs, P. R.
2014-08-01
Poorly constrained rates of biomass turnover are a key limitation of Earth system models (ESMs). In light of this, we recently proposed a new approach encoded in a model called Populations-Order-Physiology (POP), for the simulation of woody ecosystem stand dynamics, demography and disturbance-mediated heterogeneity. POP is suitable for continental to global applications and designed for coupling to the terrestrial ecosystem component of any ESM. POP bridges the gap between first-generation dynamic vegetation models (DVMs) with simple large-area parameterisations of woody biomass (typically used in current ESMs) and complex second-generation DVMs that explicitly simulate demographic processes and landscape heterogeneity of forests. The key simplification in the POP approach, compared with second-generation DVMs, is to compute physiological processes such as assimilation at grid-scale (with CABLE (Community Atmosphere Biosphere Land Exchange) or a similar land surface model), but to partition the grid-scale biomass increment among age classes defined at sub-grid-scale, each subject to its own dynamics. POP was successfully demonstrated along a savanna transect in northern Australia, replicating the effects of strong rainfall and fire disturbance gradients on observed stand productivity and structure. Here, we extend the application of POP to wide-ranging temporal and boreal forests, employing paired observations of stem biomass and density from forest inventory data to calibrate model parameters governing stand demography and biomass evolution. The calibrated POP model is then coupled to the CABLE land surface model, and the combined model (CABLE-POP) is evaluated against leaf-stem allometry observations from forest stands ranging in age from 3 to 200 year. Results indicate that simulated biomass pools conform well with observed allometry. We conclude that POP represents an ecologically plausible and efficient alternative to large-area parameterisations of woody biomass turnover, typically used in current ESMs.
NASA Technical Reports Server (NTRS)
Shen, Bo-Wen; Tao, Wei-Kuo; Lin, Yuh-Lang; Laing, Arlene
2012-01-01
In this study, it is proposed that twin tropical cyclones (TCs), Kesiny and 01A, in May 2002 formed in association with the scale interactions of three gyres that appeared as a convectively coupled mixed Rossby gravity (ccMRG) wave during an active phase of the Madden-Julian Oscillation (MJO). This is shown by analyzing observational data, including NCEP reanalysis data and METEOSAT 7 IR satellite imagery, and performing numerical simulations using a global mesoscale model. A 10-day control run is initialized at 0000 UTC 1 May 2002 with grid-scale condensation but no sub-grid cumulus parameterizations. The ccMRG wave was identified as encompassing two developing and one non-developing gyres, the first two of which intensified and evolved into the twin TCs. The control run is able to reproduce the evolution of the ccMRG wave and thus the formation of the twin TCs about two and five days in advance as well as their subsequent intensity evolution and movement within an 8-10 day period. Five additional 10-day sensitivity experiments with different model configurations are conducted to help understand the interaction of the three gyres, leading to the formation of the TCs. These experiments suggest the improved lead time in the control run may be attributed to the realistic simulation of the ccMRG wave with the following processes: (1) wave deepening (intensification) associated with a reduction in wavelength and/or the intensification of individual gyres, (2) poleward movement of gyres that may be associated with boundary layer processes, (3) realistic simulation of moist processes at regional scales in association with each of the gyres, and (4) the vertical phasing of low- and mid-level cyclonic circulations associated with a specific gyre.
NASA Astrophysics Data System (ADS)
Garrett, T. J.; Alva, S.; Glenn, I. B.; Krueger, S. K.
2015-12-01
There are two possible approaches for parameterizing sub-grid cloud dynamics in a coarser grid model. The most common is to use a fine scale model to explicitly resolve the mechanistic details of clouds to the best extent possible, and then to parameterize these behaviors cloud state for the coarser grid. A second is to invoke physical intuition and some very general theoretical principles from equilibrium statistical mechanics. This approach avoids any requirement to resolve time-dependent processes in order to arrive at a suitable solution. The second approach is widely used elsewhere in the atmospheric sciences: for example the Planck function for blackbody radiation is derived this way, where no mention is made of the complexities of modeling a large ensemble of time-dependent radiation-dipole interactions in order to obtain the "grid-scale" spectrum of thermal emission by the blackbody as a whole. We find that this statistical approach may be equally suitable for modeling convective clouds. Specifically, we make the physical argument that the dissipation of buoyant energy in convective clouds is done through mixing across a cloud perimeter. From thermodynamic reasoning, one might then anticipate that vertically stacked isentropic surfaces are characterized by a power law dlnN/dlnP = -1, where N(P) is the number clouds of perimeter P. In a Giga-LES simulation of convective clouds within a 100 km square domain we find that such a power law does appear to characterize simulated cloud perimeters along isentropes, provided a sufficient cloudy sample. The suggestion is that it may be possible to parameterize certain important aspects of cloud state without appealing to computationally expensive dynamic simulations.
Multigrid solution of the Navier-Stokes equations on highly stretched grids with defect correction
NASA Technical Reports Server (NTRS)
Sockol, Peter M.
1993-01-01
Relaxation-based multigrid solvers for the steady incompressible Navier-Stokes equations are examined to determine their computational speed and robustness. Four relaxation methods with a common discretization have been used as smoothers in a single tailored multigrid procedure. The equations are discretized on a staggered grid with first order upwind used for convection in the relaxation process on all grids and defect correction to second order central on the fine grid introduced once per multigrid cycle. A fixed W(1,1) cycle with full weighting of residuals is used in the FAS multigrid process. The resulting solvers have been applied to three 2D flow problems, over a range of Reynolds numbers, on both uniform and highly stretched grids. In all cases the L(sub 2) norm of the velocity changes is reduced to 10(exp -6) in a few 10's of fine grid sweeps. The results from this study are used to draw conclusions on the strengths and weaknesses of the individual relaxation schemes as well as those of the overall multigrid procedure when used as a solver on highly stretched grids.
Experimental Study of Vane Heat Transfer and Aerodynamics at Elevated Levels of Turbulence
NASA Technical Reports Server (NTRS)
Ames, Forrest E.
1994-01-01
A four vane subsonic cascade was used to investigate how free stream turbulence influences pressure surface heat transfer. A simulated combustor turbulence generator was built to generate high level (13 percent) large scale (Lu approximately 44 percent inlet span) turbulence. The mock combustor was also moved upstream to generate a moderate level (8.3 percent) of turbulence for comparison to smaller scale grid generated turbulence (7.8 percent). The high level combustor turbulence caused an average pressure surface heat transfer augmentation of 56 percent above the low turbulence baseline. The smaller scale grid turbulence produced the next greatest effect on heat transfer and demonstrated the importance of scale on heat transfer augmentation. In general, the heat transfer scaling parameter U(sub infinity) TU(sub infinity) LU(sub infinity)(exp -1/3) was found to hold for the turbulence. Heat transfer augmentation was also found to scale approximately on Re(sub ex)(exp 1/3) at constant turbulence conditions. Some evidence of turbulence intensification in terms of elevated dissipation rates was found along the pressure surface outside the boundary layer. However, based on the level of dissipation and the resulting heat transfer augmentation, the amplification of turbulence has only a moderate effect on pressure surface heat transfer. The flow field turbulence does drive turbulent production within the boundary layer which in turn causes the high levels of heat transfer augmentation. Unlike heat transfer, the flow field straining was found to have a significant effect on turbulence isotropy. On examination of the one dimensional spectra for u' and v', the effect to isotropy was largely limited to lower wavenumber spectra. The higher wavenumber spectra showed little or no change. The high level large scale turbulence was found to have a strong influence on wake development. The free stream turbulence significantly enhanced mixing resulting in broader and shallower wakes than the baseline case. High levels of flow field turbulence were found to correlate with a significant increase in total pressure loss in the core of the flow. Documenting the wake growth and characteristics provides boundary conditions for the downstream rotor.
NASA Astrophysics Data System (ADS)
Shang, H.; Chen, L.; Bréon, F. M.; Letu, H.; Li, S.; Wang, Z.; Su, L.
2015-11-01
The principles of cloud droplet size retrieval via Polarization and Directionality of the Earth's Reflectance (POLDER) requires that clouds be horizontally homogeneous. The retrieval is performed by combining all measurements from an area of 150 km × 150 km to compensate for POLDER's insufficient directional sampling. Using POLDER-like data simulated with the RT3 model, we investigate the impact of cloud horizontal inhomogeneity and directional sampling on the retrieval and analyze which spatial resolution is potentially accessible from the measurements. Case studies show that the sub-grid-scale variability in droplet effective radius (CDR) can significantly reduce valid retrievals and introduce small biases to the CDR (~ 1.5 μm) and effective variance (EV) estimates. Nevertheless, the sub-grid-scale variations in EV and cloud optical thickness (COT) only influence the EV retrievals and not the CDR estimate. In the directional sampling cases studied, the retrieval using limited observations is accurate and is largely free of random noise. Several improvements have been made to the original POLDER droplet size retrieval. For example, measurements in the primary rainbow region (137-145°) are used to ensure retrievals of large droplet (> 15 μm) and to reduce the uncertainties caused by cloud heterogeneity. We apply the improved method using the POLDER global L1B data from June 2008, and the new CDR results are compared with the operational CDRs. The comparison shows that the operational CDRs tend to be underestimated for large droplets because the cloudbow oscillations in the scattering angle region of 145-165° are weak for cloud fields with CDR > 15 μm. Finally, a sub-grid-scale retrieval case demonstrates that a higher resolution, e.g., 42 km × 42 km, can be used when inverting cloud droplet size distribution parameters from POLDER measurements.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Som, Sibendu; Wang, Zihan; Pei, Yuanjiang
A state-of-the-art spray modeling methodology, recently presented by Senecal et al. [ , , ], is applied to Large Eddy Simulations (LES) of vaporizing gasoline sprays. Simulations of non-combusting Spray G (gasoline fuel) from the Engine Combustion Network are performed. Adaptive mesh refinement (AMR) with cell sizes from 0.09 mm to 0.5 mm are utilized to further demonstrate grid convergence of the dynamic structure LES model for the gasoline sprays. Grid settings are recommended to optimize the accuracy/runtime tradeoff for LES-based spray simulations at different injection pressure conditions typically encountered in gasoline direct injection (GDI) applications. The influence of LESmore » sub-grid scale (SGS) models is explored by comparing the results from dynamic structure and Smagorinsky based models against simulations without any SGS model. Twenty different realizations are simulated by changing the random number seed used in the spray sub-models. It is shown that for global quantities such as spray penetration, comparing a single LES simulation to experimental data is reasonable. Through a detailed analysis using the relevance index (RI) criteria, recommendations are made regarding the minimum number of LES realizations required for accurate prediction of the gasoline sprays.« less
Electron acceleration by turbulent plasmoid reconnection
NASA Astrophysics Data System (ADS)
Zhou, X.; Büchner, J.; Widmer, F.; Muñoz, P. A.
2018-04-01
In space and astrophysical plasmas, like in planetary magnetospheres, as that of Mercury, energetic electrons are often found near current sheets, which hint at electron acceleration by magnetic reconnection. Unfortunately, electron acceleration by reconnection is not well understood yet, in particular, acceleration by turbulent plasmoid reconnection. We have investigated electron acceleration by turbulent plasmoid reconnection, described by MHD simulations, via test particle calculations. In order to avoid resolving all relevant turbulence scales down to the dissipation scales, a mean-field turbulence model is used to describe the turbulence of sub-grid scales and their effects via a turbulent electromotive force (EMF). The mean-field model describes the turbulent EMF as a function of the mean values of current density, vorticity, magnetic field as well as of the energy, cross-helicity, and residual helicity of the turbulence. We found that, mainly around X-points of turbulent reconnection, strongly enhanced localized EMFs most efficiently accelerated electrons and caused the formation of power-law spectra. Magnetic-field-aligned EMFs, caused by the turbulence, dominate the electron acceleration process. Scaling the acceleration processes to parameters of the Hermean magnetotail, electron energies up to 60 keV can be reached by turbulent plasmoid reconnection through the thermal plasma.
NASA Astrophysics Data System (ADS)
Stefanski, Douglas Lawrence
A finite volume method for solving the Reynolds Averaged Navier-Stokes (RANS) equations on unstructured hybrid grids is presented. Capabilities for handling arbitrary mixtures of reactive gas species within the unstructured framework are developed. The modeling of turbulent effects is carried out via the 1998 Wilcox k -- o model. This unstructured solver is incorporated within VULCAN -- a multi-block structured grid code -- as part of a novel patching procedure in which non-matching interfaces between structured blocks are replaced by transitional unstructured grids. This approach provides a fully-conservative alternative to VULCAN's non-conservative patching methods for handling such interfaces. In addition, the further development of the standalone unstructured solver toward large-eddy simulation (LES) applications is also carried out. Dual time-stepping using a Crank-Nicholson formulation is added to recover time-accuracy, and modeling of sub-grid scale effects is incorporated to provide higher fidelity LES solutions for turbulent flows. A switch based on the work of Ducros, et al., is implemented to transition from a monotonicity-preserving flux scheme near shocks to a central-difference method in vorticity-dominated regions in order to better resolve small-scale turbulent structures. The updated unstructured solver is used to carry out large-eddy simulations of a supersonic constrained mixing layer.
NASA Astrophysics Data System (ADS)
Undapalli, Satish
A new combustor referred to as Stagnation Point Reverse Flow (SPRF) combustor has been developed at Georgia Tech to meet the increasingly stringent emission regulations. The combustor incorporates a novel design to meet the conflicting requirements of low pollution and high stability in both premixed and non-premixed modes. The objective of this thesis work is to perform Large Eddy Simulations (LES) on this lab-scale combustor and elucidate the underlying physics that has resulted in its excellent performance. To achieve this, numerical simulations have been performed in both the premixed and non-premixed combustion modes, and velocity field, species field, entrainment characteristics, flame structure, emissions, and mixing characteristics have been analyzed. Simulations have been carried out first for a non-reactive case to resolve relevant fluid mechanics without heat release by the computational grid. The computed mean and RMS quantities in the non-reacting case compared well with the experimental data. Next, the simulations were extended for the premixed reactive case by employing different sub-grid scale combustion chemistry closures: Eddy Break Up (EBU), Artificially Thickened Flame (TF) and Linear Eddy Mixing (LEM) models. Results from the EBU and TF models exhibit reasonable agreement with the experimental velocity field. However, the computed thermal and species fields have noticeable discrepancies. Only LEM with LES (LEMLES), which is an advanced scalar approach, has been able to accurately predict both the velocity and species fields. Scalar mixing plays an important role in combustion, and this is solved directly at the sub-grid scales in LEM. As a result, LEM accurately predicts the scalar fields. Due to the two way coupling between the super-grid and sub-grid quantities, the velocity predictions also compare very well with the experiments. In other approaches, the sub-grid effects have been either modeled using conventional approaches (EBU) or need some ad hoc adjustments to account these effects accurately (TF). The results from LEMLES, using a reduced chemical mechanism, have been analyzed in the premixed mode. The results show that mass entrainment occurs along the shear layer in the combustor. The entrained mass carries products into the reactant stream and provides reactant preheating. Thus, product entrainment enhances the reaction rates and help stabilize the flame even at very lean conditions. These products have been shown to enter into the flame through local extinction zones present on the flame surface. The flame structure has been further analyzed, and the combustion mode was found to be primarily in thin reaction zones. Closer to the injector, there are isolated regions, where the combustion mode is in broken reaction zones, while the downstream flame structure is closer to a flamelet regime. The emissions in the combustor have been studied using simple global mechanisms for NO x. Computations have shown extremely low NOx values, comparable to the measured emissions. These low emissions have been shown to be primarily due to the low temperatures in the combustor. LEMLES computations have also been performed with a detailed chemistry to capture more accurate flame structure. The flame in the detailed chemistry case shows more extinction zones close to the injector than that in the reduced chemical mechanism. The LEMLES approach has also been used to resolve the combustion mode in the non-premixed case. The studies have indicated that the mixing of the fuel and air close to the injector controls the combustion process. The predictions in the near field have been shown to be very sensitive to the inflow conditions. Analysis has shown that the fuel and air mixing occurs to lean proportions in the combustor before any burning takes place. The flame structure in the non-premixed mode was very similar to the premixed mode. Along with the fuel air mixing, the products also mixed with the reactants and provided the preheating effects to stabilize the flame in the downstream region of the combustor.
NASA Astrophysics Data System (ADS)
Marques, Gustavo; Stern, Alon; Harrison, Matthew; Sergienko, Olga; Hallberg, Robert
2017-04-01
Dense shelf water (DSW) is formed in coastal polynyas around Antarctica as a result of intense cooling and brine rejection. A fraction of this water reaches ice shelves cavities and is modified due to interactions with sub-ice-shelf melt water. This modified water mass contributes to the formation of Antarctic Bottom Water, and consequently, influences the large-scale ocean circulation. Here, we investigate the role of sub-ice-shelf melting in the formation and export of DSW using idealized simulations with an isopycnal ocean model (MOM6) coupled with a sea ice model (SIS2) and a thermodynamic active ice shelf. A set of experiments is conducted with variable horizontal grid resolutions (0.5, 1.0 and 2.0 km), ice shelf geometries and atmospheric forcing. In all simulations DSW is spontaneously formed in coastal polynyas due to the combined effect of the imposed atmospheric forcing and the ocean state. Our results show that sub-ice-shelf melting can significantly change the rate of dense shelf water outflows, highlighting the importance of this process to correctly represent bottom water formation.
Analysis of Surface Heterogeneity Effects with Mesoscale Terrestrial Modeling Platforms
NASA Astrophysics Data System (ADS)
Simmer, C.
2015-12-01
An improved understanding of the full variability in the weather and climate system is crucial for reducing the uncertainty in weather forecasting and climate prediction, and to aid policy makers to develop adaptation and mitigation strategies. A yet unknown part of uncertainty in the predictions from the numerical models is caused by the negligence of non-resolved land surface heterogeneity and the sub-surface dynamics and their potential impact on the state of the atmosphere. At the same time, mesoscale numerical models using finer horizontal grid resolution [O(1)km] can suffer from inconsistencies and neglected scale-dependencies in ABL parameterizations and non-resolved effects of integrated surface-subsurface lateral flow at this scale. Our present knowledge suggests large-eddy-simulation (LES) as an eventual solution to overcome the inadequacy of the physical parameterizations in the atmosphere in this transition scale, yet we are constrained by the computational resources, memory management, big-data, when using LES for regional domains. For the present, there is a need for scale-aware parameterizations not only in the atmosphere but also in the land surface and subsurface model components. In this study, we use the recently developed Terrestrial Systems Modeling Platform (TerrSysMP) as a numerical tool to analyze the uncertainty in the simulation of surface exchange fluxes and boundary layer circulations at grid resolutions of the order of 1km, and explore the sensitivity of the atmospheric boundary layer evolution and convective rainfall processes on land surface heterogeneity.
A New Stellar Atmosphere Grid and Comparisons with HST /STIS CALSPEC Flux Distributions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bohlin, Ralph C.; Fleming, Scott W.; Gordon, Karl D.
The Space Telescope Imaging Spectrograph has measured the spectral energy distributions for several stars of types O, B, A, F, and G. These absolute fluxes from the CALSPEC database are fit with a new spectral grid computed from the ATLAS-APOGEE ATLAS9 model atmosphere database using a chi-square minimization technique in four parameters. The quality of the fits are compared for complete LTE grids by Castelli and Kurucz (CK04) and our new comprehensive LTE grid (BOSZ). For the cooler stars, the fits with the MARCS LTE grid are also evaluated, while the hottest stars are also fit with the NLTE Lanzmore » and Hubeny OB star grids. Unfortunately, these NLTE models do not transition smoothly in the infrared to agree with our new BOSZ LTE grid at the NLTE lower limit of T {sub eff} = 15,000 K. The new BOSZ grid is available via the Space Telescope Institute MAST archive and has a much finer sampled IR wavelength scale than CK04, which will facilitate the modeling of stars observed by the James Webb Space Telescope . Our result for the angular diameter of Sirius agrees with the ground-based interferometric value.« less
Turbulence Enhancement by Fractal Square Grids: Effects of the Number of Fractal Scales
NASA Astrophysics Data System (ADS)
Omilion, Alexis; Ibrahim, Mounir; Zhang, Wei
2017-11-01
Fractal square grids offer a unique solution for passive flow control as they can produce wakes with a distinct turbulence intensity peak and a prolonged turbulence decay region at the expense of only minimal pressure drop. While previous studies have solidified this characteristic of fractal square grids, how the number of scales (or fractal iterations N) affect turbulence production and decay of the induced wake is still not well understood. The focus of this research is to determine the relationship between the fractal iteration N and the turbulence produced in the wake flow using well-controlled water-tunnel experiments. Particle Image Velocimetry (PIV) is used to measure the instantaneous velocity fields downstream of four different fractal grids with increasing number of scales (N = 1, 2, 3, and 4) and a conventional single-scale grid. By comparing the turbulent scales and statistics of the wake, we are able to determine how each iteration affects the peak turbulence intensity and the production/decay of turbulence from the grid. In light of the ability of these fractal grids to increase turbulence intensity with low pressure drop, this work can potentially benefit a wide variety of applications where energy efficient mixing or convective heat transfer is a key process.
Sensitivity of an Antarctic Ice Sheet Model to Sub-Ice-Shelf Melting
NASA Astrophysics Data System (ADS)
Lipscomb, W. H.; Leguy, G.; Urban, N. M.; Berdahl, M.
2017-12-01
Theory and observations suggest that marine-based sectors of the Antarctic ice sheet could retreat rapidly under ocean warming and increased melting beneath ice shelves. Numerical models of marine ice sheets vary widely in sensitivity, depending on grid resolution and the parameterization of key processes (e.g., calving and hydrofracture). Here we study the sensitivity of the Antarctic ice sheet to ocean warming and sub-shelf melting in standalone simulations of the Community Ice Sheet Model (CISM). Melt rates either are prescribed based on observations and high-resolution ocean model output, or are derived from a plume model forced by idealized ocean temperature profiles. In CISM, we vary the model resolution (between 1 and 8 km), Stokes approximation (shallow-shelf, depth-integrated higher-order, or 3D higher-order) and calving scheme to create an ensemble of plausible responses to sub-shelf melting. This work supports a broader goal of building statistical and reduced models that can translate large-scale Earth-system model projections to changes in Antarctic ocean temperatures and ice sheet discharge, thus better quantifying uncertainty in Antarctic-sourced sea-level rise.
Simulation of Deep Convective Clouds with the Dynamic Reconstruction Turbulence Closure
NASA Astrophysics Data System (ADS)
Shi, X.; Chow, F. K.; Street, R. L.; Bryan, G. H.
2017-12-01
The terra incognita (TI), or gray zone, in simulations is a range of grid spacing comparable to the most energetic eddy diameter. Spacing in mesoscale and simulations is much larger than the eddies, and turbulence is parameterized with one-dimensional vertical-mixing. Large eddy simulations (LES) have grid spacing much smaller than the energetic eddies, and use three-dimensional models of turbulence. Studies of convective weather use convection-permitting resolutions, which are in the TI. Neither mesoscale-turbulence nor LES models are designed for the TI, so TI turbulence parameterization needs to be discussed. Here, the effects of sub-filter scale (SFS) closure schemes on the simulation of deep tropical convection are evaluated by comparing three closures, i.e. Smagorinsky model, Deardorff-type TKE model and the dynamic reconstruction model (DRM), which partitions SFS turbulence into resolvable sub-filter scales (RSFS) and unresolved sub-grid scales (SGS). The RSFS are reconstructed, and the SGS are modeled with a dynamic eddy viscosity/diffusivity model. The RSFS stresses/fluxes allow backscatter of energy/variance via counter-gradient stresses/fluxes. In high-resolution (100m) simulations of tropical convection use of these turbulence models did not lead to significant differences in cloud water/ice distribution, precipitation flux, or vertical fluxes of momentum and heat. When model resolutions are coarsened, the Smagorinsky and TKE models overestimate cloud ice and produces large-amplitude downward heat flux in the middle troposphere (not found in the high-resolution simulations). This error is a result of unrealistically large eddy diffusivities, i.e., the eddy diffusivity of the DRM is on the order of 1 for the coarse resolution simulations, the eddy diffusivity of the Smagorinsky and TKE model is on the order of 100. Splitting the eddy viscosity/diffusivity scalars into vertical and horizontal components by using different length scales and strain rate components helps to reduce the errors, but does not completely remedy the problem. In contrast, the coarse resolution simulations using the DRM produce results that are more consistent with the high-resolution results, suggesting that the DRM is a more appropriate turbulence model for simulating convection in the TI.
A Priori Subgrid Scale Modeling for a Droplet Laden Temporal Mixing Layer
NASA Technical Reports Server (NTRS)
Okongo, Nora; Bellan, Josette
2000-01-01
Subgrid analysis of a transitional temporal mixing layer with evaporating droplets has been performed using a direct numerical simulation (DNS) database. The DNS is for a Reynolds number (based on initial vorticity thickness) of 600, with droplet mass loading of 0.2. The gas phase is computed using a Eulerian formulation, with Lagrangian droplet tracking. Since Large Eddy Simulation (LES) of this flow requires the computation of unfiltered gas-phase variables at droplet locations from filtered gas-phase variables at the grid points, it is proposed to model these by assuming the gas-phase variables to be given by the filtered variables plus a correction based on the filtered standard deviation, which can be computed from the sub-grid scale (SGS) standard deviation. This model predicts unfiltered variables at droplet locations better than simply interpolating the filtered variables. Three methods are investigated for modeling the SGS standard deviation: Smagorinsky, gradient and scale-similarity. When properly calibrated, the gradient and scale-similarity methods give results in excellent agreement with the DNS.
A high throughput geocomputing system for remote sensing quantitative retrieval and a case study
NASA Astrophysics Data System (ADS)
Xue, Yong; Chen, Ziqiang; Xu, Hui; Ai, Jianwen; Jiang, Shuzheng; Li, Yingjie; Wang, Ying; Guang, Jie; Mei, Linlu; Jiao, Xijuan; He, Xingwei; Hou, Tingting
2011-12-01
The quality and accuracy of remote sensing instruments have been improved significantly, however, rapid processing of large-scale remote sensing data becomes the bottleneck for remote sensing quantitative retrieval applications. The remote sensing quantitative retrieval is a data-intensive computation application, which is one of the research issues of high throughput computation. The remote sensing quantitative retrieval Grid workflow is a high-level core component of remote sensing Grid, which is used to support the modeling, reconstruction and implementation of large-scale complex applications of remote sensing science. In this paper, we intend to study middleware components of the remote sensing Grid - the dynamic Grid workflow based on the remote sensing quantitative retrieval application on Grid platform. We designed a novel architecture for the remote sensing Grid workflow. According to this architecture, we constructed the Remote Sensing Information Service Grid Node (RSSN) with Condor. We developed a graphic user interface (GUI) tools to compose remote sensing processing Grid workflows, and took the aerosol optical depth (AOD) retrieval as an example. The case study showed that significant improvement in the system performance could be achieved with this implementation. The results also give a perspective on the potential of applying Grid workflow practices to remote sensing quantitative retrieval problems using commodity class PCs.
Regional climates in the GISS general circulation model: Surface air temperature
NASA Technical Reports Server (NTRS)
Hewitson, Bruce
1994-01-01
One of the more viable research techniques into global climate change for the purpose of understanding the consequent environmental impacts is based on the use of general circulation models (GCMs). However, GCMs are currently unable to reliably predict the regional climate change resulting from global warming, and it is at the regional scale that predictions are required for understanding human and environmental responses. Regional climates in the extratropics are in large part governed by the synoptic-scale circulation and the feasibility of using this interscale relationship is explored to provide a way of moving to grid cell and sub-grid cell scales in the model. The relationships between the daily circulation systems and surface air temperature for points across the continental United States are first developed in a quantitative form using a multivariate index based on principal components analysis (PCA) of the surface circulation. These relationships are then validated by predicting daily temperature using observed circulation and comparing the predicted values with the observed temperatures. The relationships predict surface temperature accurately over the major portion of the country in winter, and for half the country in summer. These relationships are then applied to the surface synoptic circulation of the Goddard Institute for Space Studies (GISS) GCM control run, and a set of surface grid cell temperatures are generated. These temperatures, based on the larger-scale validated circulation, may now be used with greater confidence at the regional scale. The generated temperatures are compared to those of the model and show that the model has regional errors of up to 10 C in individual grid cells.
Multiscale image processing and antiscatter grids in digital radiography.
Lo, Winnie Y; Hornof, William J; Zwingenberger, Allison L; Robertson, Ian D
2009-01-01
Scatter radiation is a source of noise and results in decreased signal-to-noise ratio and thus decreased image quality in digital radiography. We determined subjectively whether a digitally processed image made without a grid would be of similar quality to an image made with a grid but without image processing. Additionally the effects of exposure dose and of a using a grid with digital radiography on overall image quality were studied. Thoracic and abdominal radiographs of five dogs of various sizes were made. Four acquisition techniques were included (1) with a grid, standard exposure dose, digital image processing; (2) without a grid, standard exposure dose, digital image processing; (3) without a grid, half the exposure dose, digital image processing; and (4) with a grid, standard exposure dose, no digital image processing (to mimic a film-screen radiograph). Full-size radiographs as well as magnified images of specific anatomic regions were generated. Nine reviewers rated the overall image quality subjectively using a five-point scale. All digitally processed radiographs had higher overall scores than nondigitally processed radiographs regardless of patient size, exposure dose, or use of a grid. The images made at half the exposure dose had a slightly lower quality than those made at full dose, but this was only statistically significant in magnified images. Using a grid with digital image processing led to a slight but statistically significant increase in overall quality when compared with digitally processed images made without a grid but whether this increase in quality is clinically significant is unknown.
Liu, Yu; Wu, Bing-Fang; Zeng, Yuan; Zhang, Lei
2013-09-01
The integration of the effects of landscape pattern to the assessment of the impacts of soil erosion on eco-environmental is of practical significance in methodological prospect, being able to provide an approach for identifying water body's sediment source area, assessing the potential risks of sediment export of on-site soil erosion to the target water body, and evaluating the capacity of regional landscape pattern in preventing soil loss. In this paper, the RUSLE model was applied to simulate the on-site soil erosion rate. With the consideration of the soil retention potential of vegetation cover and topography, a quantitative assessment was conducted on the impacts of soil erosion in the water source region of the middle route for South-to-North Water Transfer Project on rivers and reservoirs by delineating landscape pattern at point (or cell) scale and sub-watershed level. At point (or grid cell) scale, the index of soil erosion impact intensity (I) was developed as an indicator of the potential risk of sediment export to the water bodies. At sub-watershed level, the landscape leakiness index (LI) was employed to indicate the sediment retention capacity of a given landscape pattern. The results revealed that integrating the information of landscape pattern and the indices of soil erosion process could spatially effectively reflect the impact intensity of in situ soil erosion on water bodies. The LI was significantly exponentially correlated to the mean sediment retention capacity of landscape and the mean vegetation coverage of watershed, and the sediment yield at sub-watershed scale was significantly correlated to the LI in an exponential regression. It could be concluded that the approach of delineating landscape pattern based on soil erosion process and the integration of the information of landscape pattern with its soil retention potential could provide a new approach for the risk evaluation of soil erosion.
Vorticity imbalance and stability in relation to convection
NASA Technical Reports Server (NTRS)
Read, W. L.; Scoggins, J. R.
1977-01-01
A complete synoptic-scale vorticity budget was related to convection storm development in the eastern two-thirds of the United States. The 3-h sounding interval permitted a study of time changes of the vorticity budget in areas of convective storms. Results of analyses revealed significant changes in values of terms in the vorticity equation at different stages of squall line development. Average budgets for all areas of convection indicate systematic imbalance in the terms in the vorticity equation. This imbalance resulted primarily from sub-grid scale processes. Potential instability in the lower troposphere was analyzed in relation to the development of convective activity. Instability was related to areas of convection; however, instability alone was inadequate for forecast purposes. Combinations of stability and terms in the vorticity equation in the form of indices succeeded in depicting areas of convection better than any one item separately.
Gasdynamic model of turbulent combustion in an explosion
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuhl, A.L.; Ferguson, R.E.; Chien, K.Y.
1994-08-31
Proposed here is a gasdynamic model of turbulent combustion in explosions. It is used to investigate turbulent mixing aspects of afterburning found in TNT charges detonated in air. Evolution of the turbulent velocity field was calculated by a high-order Godunov solution of the gasdynamic equations. Adaptive Mesh Refinement (AMR) was used to follow convective-mixing processes on the computational grid. Combustion was then taken into account by a simplified sub-grid model, demonstrating that it was controlled by turbulent mixing. The rate of fuel consumption decayed inversely with time, and was shown to be insensitive to grid resolution.
Scale-dependent coupling of hysteretic capillary pressure, trapping, and fluid mobilities
NASA Astrophysics Data System (ADS)
Doster, F.; Celia, M. A.; Nordbotten, J. M.
2012-12-01
Many applications of multiphase flow in porous media, including CO2-storage and enhanced oil recovery, require mathematical models that span a large range of length scales. In the context of numerical simulations, practical grid sizes are often on the order of tens of meters, thereby de facto defining a coarse model scale. Under particular conditions, it is possible to approximate the sub-grid-scale distribution of the fluid saturation within a grid cell; that reconstructed saturation can then be used to compute effective properties at the coarse scale. If both the density difference between the fluids and the vertical extend of the grid cell are large, and buoyant segregation within the cell on a sufficiently shorte time scale, then the phase pressure distributions are essentially hydrostatic and the saturation profile can be reconstructed from the inferred capillary pressures. However, the saturation reconstruction may not be unique because the parameters and parameter functions of classical formulations of two-phase flow in porous media - the relative permeability functions, the capillary pressure -saturation relationship, and the residual saturations - show path dependence, i.e. their values depend not only on the state variables but also on their drainage and imbibition histories. In this study we focus on capillary pressure hysteresis and trapping and show that the contribution of hysteresis to effective quantities is dependent on the vertical length scale. By studying the transition from the two extreme cases - the homogeneous saturation distribution for small vertical extents and the completely segregated distribution for large extents - we identify how hysteretic capillary pressure at the local scale induces hysteresis in all coarse-scale quantities for medium vertical extents and finally vanishes for large vertical extents. Our results allow for more accurate vertically integrated modeling while improving our understanding of the coupling of capillary pressure and relative permeabilities over larger length scales.
PHOTOCHEMICAL SIMULATIONS OF POINT SOURCE EMISSIONS WITH THE MODELS-3 CMAQ PLUME-IN-GRID APPROACH
A plume-in-grid (PinG) approach has been designed to provide a realistic treatment for the simulation the dynamic and chemical processes impacting pollutant species in major point source plumes during a subgrid scale phase within an Eulerian grid modeling framework. The PinG sci...
Sensitivity simulations of superparameterised convection in a general circulation model
NASA Astrophysics Data System (ADS)
Rybka, Harald; Tost, Holger
2015-04-01
Cloud Resolving Models (CRMs) covering a horizontal grid spacing from a few hundred meters up to a few kilometers have been used to explicitly resolve small-scale and mesoscale processes. Special attention has been paid to realistically represent cloud dynamics and cloud microphysics involving cloud droplets, ice crystals, graupel and aerosols. The entire variety of physical processes on the small-scale interacts with the larger-scale circulation and has to be parameterised on the coarse grid of a general circulation model (GCM). Since more than a decade an approach to connect these two types of models which act on different scales has been developed to resolve cloud processes and their interactions with the large-scale flow. The concept is to use an ensemble of CRM grid cells in a 2D or 3D configuration in each grid cell of the GCM to explicitly represent small-scale processes avoiding the use of convection and large-scale cloud parameterisations which are a major source for uncertainties regarding clouds. The idea is commonly known as superparameterisation or cloud-resolving convection parameterisation. This study presents different simulations of an adapted Earth System Model (ESM) connected to a CRM which acts as a superparameterisation. Simulations have been performed with the ECHAM/MESSy atmospheric chemistry (EMAC) model comparing conventional GCM runs (including convection and large-scale cloud parameterisations) with the improved superparameterised EMAC (SP-EMAC) modeling one year with prescribed sea surface temperatures and sea ice content. The sensitivity of atmospheric temperature, precipiation patterns, cloud amount and types is observed changing the embedded CRM represenation (orientation, width, no. of CRM cells, 2D vs. 3D). Additionally, we also evaluate the radiation balance with the new model configuration, and systematically analyse the impact of tunable parameters on the radiation budget and hydrological cycle. Furthermore, the subgrid variability (individual CRM cell output) is analysed in order to illustrate the importance of a highly varying atmospheric structure inside a single GCM grid box. Finally, the convective transport of Radon is observed comparing different transport procedures and their influence on the vertical tracer distribution.
NASA Astrophysics Data System (ADS)
Yang, Xiang
2017-11-01
The sizes of fluid motions in wall-bounded flows scale approximately as their distances from the wall. At high Reynolds numbers, resolving near-wall, small-scale, yet momentum-transferring eddies are computationally intensive, and to alleviate the strict near-wall grid resolution requirement, a wall model is usually used. The wall model of interest here is the integral wall model. This model parameterizes the near-wall sub-grid velocity profile as being comprised of a linear inner-layer and a logarithmic meso-layer with one additional term that accounts for the effects of flow acceleration, pressure gradients etc. We use the integral wall model for wall-modeled large-eddy simulations (WMLES) of turbulent boundary layers over rough walls. The effects of rough-wall topology on drag forces are investigated. A rough-wall model is then developed based on considerations of such effects, which are now known as mutual sheltering among roughness elements. Last, we discuss briefly a new interpretation of the Townsend attached eddy hypothesis-the hierarchical random additive process model (HRAP). The analogy between the energy cascade and the momentum cascade is mathematically formal as HRAP follows the multi-fractal formulism, which was extensively used for the energy cascade.
Evaluation of the Momentum Closure Schemes in MPAS-Ocean
NASA Astrophysics Data System (ADS)
Zhao, Shimei; Liu, Yudi; Liu, Wei
2018-04-01
In order to compare and evaluate the performances of the Laplacian viscosity closure, the biharmonic viscosity closure, and the Leith closure momentum schemes in the MPAS-Ocean model, a variety of physical quantities, such as the relative reference potential energy (RPE) change, the RPE time change rate (RPETCR), the grid Reynolds number, the root mean square (RMS) of kinetic energy, and the spectra of kinetic energy and enstrophy, are calculated on the basis of results of a 3D baroclinic periodic channel. Results indicate that: 1) The RPETCR demonstrates a saturation phenomenon in baroclinic eddy tests. The critical grid Reynolds number corresponding to RPETCR saturation differs between the three closures: the largest value is in the biharmonic viscosity closure, followed by that in the Laplacian viscosity closure, and that in the Leith closure is the smallest. 2) All three closures can effectively suppress spurious dianeutral mixing by reducing the grid Reynolds number under sub-saturation conditions of the RPETCR, but they can also damage certain physical processes. Generally, the damage to the rotation process is greater than that to the advection process. 3) The dissipation in the biharmonic viscosity closure is strongly dependent on scales. Most dissipation concentrates on small scales, and the energy of small-scale eddies is often transferred to large-scale kinetic energy. The viscous dissipation in the Laplacian viscosity closure is the strongest on various scales, followed by that in the Leith closure. Note that part of the small-scale kinetic energy is also transferred to large-scale kinetic energy in the Leith closure. 4) The characteristic length scale L and the dimensionless parameter D in the Leith closure are inherently coupled. The RPETCR is inversely proportional to the product of D and L. When the product of D and L is constant, both the simulated RPETCR and the inhibition of spurious dianeutral mixing are the same in all tests using the Leith closure. The dissipative scale in the Leith closure depends on the parameter L, and the dissipative intensity depends on the parameter D. 5) Although optimal results may not be achieved by using the optimal parameters obtained from the 2D barotropic model in the 3D baroclinic simulation, the total energies are dissipative in all three closures. Dissipation is the strongest in the biharmonic viscosity closure, followed by that in the Leith closure, and that in the Laplacian viscosity closure is the weakest. Mesoscale eddies develop the fastest in the biharmonic viscosity closure after the baroclinic adjustment process finishes, and the kinetic energy reaches its maximum, which is attributed to the smallest dissipation of enstrophy in the biharmonic viscosity closure. Mesoscale eddies develop the slowest, and the kinetic energy peak value is the smallest in the Laplacian viscosity closure. Results in the Leith closure are between that in the biharmonic viscosity closure and the Laplacian viscosity closure.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mottez, F.; Chanteur, G.; Roux, A.
1992-07-01
A two-dimensional, explicit, electrostatic particle code is used to investigate the nonlinear behavior of electrostatic ion waves generated by an ion beam flowing through a thermal ion and electron background in a strongly magnetized plasma ({omega}{sub ce} {much gt} {omega}{sub pe} where {omega}{sub ce} and {omega}{sub pe} are the electron gyrofrequency and the plasma frequency). To follow the nonlinear evolution of these ions waves, a long-lasting simulation is run with a large simulation grid: 128 {times} 512{lambda}{sub d}. Beam ions are shown to generate oblique waves. The nonlinear beatings between these oblique waves produce purely transverse waves, which leads tomore » a strong modulation of the density and of the electric potential in a direction transverse to the magnetic field. The transverse scale of these essentially field-aligned filaments is L{sub {perpendicular}} = 10 {rho}{sub i} where {rho}{sub i} is the ion Larmor radius of beam ions. Within these filaments, relatively stable field-aligned density and potential structures develop. The typical size, along the magnetic field, of these structures is L{sub {parallel}} = 10 {lambda}{sub d}, the density is modulated by 30%, and the electric potential is as large as T{sub e} within these structures. Unlike the potential structures that develop in a two-component plasma with downgoing electrons, these structures move upward. These characteristics are in good agreement with the weak double layers recently detected by Viking.« less
NASA Astrophysics Data System (ADS)
Leutwyler, David; Fuhrer, Oliver; Cumming, Benjamin; Lapillonne, Xavier; Gysi, Tobias; Lüthi, Daniel; Osuna, Carlos; Schär, Christoph
2014-05-01
The representation of moist convection is a major shortcoming of current global and regional climate models. State-of-the-art global models usually operate at grid spacings of 10-300 km, and therefore cannot fully resolve the relevant upscale and downscale energy cascades. Therefore parametrization of the relevant sub-grid scale processes is required. Several studies have shown that this approach entails major uncertainties for precipitation processes, which raises concerns about the model's ability to represent precipitation statistics and associated feedback processes, as well as their sensitivities to large-scale conditions. Further refining the model resolution to the kilometer scale allows representing these processes much closer to first principles and thus should yield an improved representation of the water cycle including the drivers of extreme events. Although cloud-resolving simulations are very useful tools for climate simulations and numerical weather prediction, their high horizontal resolution and consequently the small time steps needed, challenge current supercomputers to model large domains and long time scales. The recent innovations in the domain of hybrid supercomputers have led to mixed node designs with a conventional CPU and an accelerator such as a graphics processing unit (GPU). GPUs relax the necessity for cache coherency and complex memory hierarchies, but have a larger system memory-bandwidth. This is highly beneficial for low compute intensity codes such as atmospheric stencil-based models. However, to efficiently exploit these hybrid architectures, climate models need to be ported and/or redesigned. Within the framework of the Swiss High Performance High Productivity Computing initiative (HP2C) a project to port the COSMO model to hybrid architectures has recently come to and end. The product of these efforts is a version of COSMO with an improved performance on traditional x86-based clusters as well as hybrid architectures with GPUs. We present our redesign and porting approach as well as our experience and lessons learned. Furthermore, we discuss relevant performance benchmarks obtained on the new hybrid Cray XC30 system "Piz Daint" installed at the Swiss National Supercomputing Centre (CSCS), both in terms of time-to-solution as well as energy consumption. We will demonstrate a first set of short cloud-resolving climate simulations at the European-scale using the GPU-enabled COSMO prototype and elaborate our future plans on how to exploit this new model capability.
Grid Scale Energy Storage (Symposium EE8)
2016-06-01
27709-2211 Grid-Scale Energy Storage, electrolytes, systems ntegration, Lithium - ion chemistry, Redox flow batteries REPORT DOCUMENTATION PAGE 11... Lithium - Ion Chemistry (4) Redox Flow Batteries Christopher J. Orendorff from Sandia National Laboratories kicked off the symposium on Tuesday...for redox flow batteries . SEI formation is a well-known process in standard lithium - ion battery operation; however, using aqueous electrolytes does
Challenges of Representing Sub-Grid Physics in an Adaptive Mesh Refinement Atmospheric Model
NASA Astrophysics Data System (ADS)
O'Brien, T. A.; Johansen, H.; Johnson, J. N.; Rosa, D.; Benedict, J. J.; Keen, N. D.; Collins, W.; Goodfriend, E.
2015-12-01
Some of the greatest potential impacts from future climate change are tied to extreme atmospheric phenomena that are inherently multiscale, including tropical cyclones and atmospheric rivers. Extremes are challenging to simulate in conventional climate models due to existing models' coarse resolutions relative to the native length-scales of these phenomena. Studying the weather systems of interest requires an atmospheric model with sufficient local resolution, and sufficient performance for long-duration climate-change simulations. To this end, we have developed a new global climate code with adaptive spatial and temporal resolution. The dynamics are formulated using a block-structured conservative finite volume approach suitable for moist non-hydrostatic atmospheric dynamics. By using both space- and time-adaptive mesh refinement, the solver focuses computational resources only where greater accuracy is needed to resolve critical phenomena. We explore different methods for parameterizing sub-grid physics, such as microphysics, macrophysics, turbulence, and radiative transfer. In particular, we contrast the simplified physics representation of Reed and Jablonowski (2012) with the more complex physics representation used in the System for Atmospheric Modeling of Khairoutdinov and Randall (2003). We also explore the use of a novel macrophysics parameterization that is designed to be explicitly scale-aware.
The Applied Mathematics for Power Systems (AMPS)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chertkov, Michael
2012-07-24
Increased deployment of new technologies, e.g., renewable generation and electric vehicles, is rapidly transforming electrical power networks by crossing previously distinct spatiotemporal scales and invalidating many traditional approaches for designing, analyzing, and operating power grids. This trend is expected to accelerate over the coming years, bringing the disruptive challenge of complexity, but also opportunities to deliver unprecedented efficiency and reliability. Our Applied Mathematics for Power Systems (AMPS) Center will discover, enable, and solve emerging mathematics challenges arising in power systems and, more generally, in complex engineered networks. We will develop foundational applied mathematics resulting in rigorous algorithms and simulation toolboxesmore » for modern and future engineered networks. The AMPS Center deconstruction/reconstruction approach 'deconstructs' complex networks into sub-problems within non-separable spatiotemporal scales, a missing step in 20th century modeling of engineered networks. These sub-problems are addressed within the appropriate AMPS foundational pillar - complex systems, control theory, and optimization theory - and merged or 'reconstructed' at their boundaries into more general mathematical descriptions of complex engineered networks where important new questions are formulated and attacked. These two steps, iterated multiple times, will bridge the growing chasm between the legacy power grid and its future as a complex engineered network.« less
NASA Astrophysics Data System (ADS)
Wang, Qing; Zhao, Xinyu; Ihme, Matthias
2017-11-01
Particle-laden turbulent flows are important in numerous industrial applications, such as spray combustion engines, solar energy collectors etc. It is of interests to study this type of flows numerically, especially using large-eddy simulations (LES). However, capturing the turbulence-particle interaction in LES remains challenging due to the insufficient representation of the effect of sub-grid scale (SGS) dispersion. In the present work, a closure technique for the SGS dispersion using regularized deconvolution method (RDM) is assessed. RDM was proposed as the closure for the SGS dispersion in a counterflow spray that is studied numerically using finite difference method on a structured mesh. A presumed form of LES filter is used in the simulations. In the present study, this technique has been extended to finite volume method with an unstructured mesh, where no presumption on the filter form is required. The method is applied to a series of particle-laden turbulent jets. Parametric analyses of the model performance are conducted for flows with different Stokes numbers and Reynolds numbers. The results from LES will be compared against experiments and direct numerical simulations (DNS).
Thematic mapper-derived mineral distribution maps of Idaho, Nevada, and western Montana
Raines, Gary L.
2006-01-01
This report provides mineral distribution maps based on TM spectral information of minerals commonly associated with hydrothermal alteration in Nevada, Idaho, and western Montana. The product of the processing is provided as four ESRI GRID files with 30 m resolution by state. UTM Zone 11 projection is used for Nevada (grid clsnv) and western Idaho (grid clsid), UTM Zone 12 is used for eastern Idaho and western Montana (grid clsid_mt). A fourth grid with a special Albers projection is used for the Headwaters project covering Idaho and western Montana (grid crccls_hs). Symbolization for all four grids is stored in the ESRI layer or LYR files and color or CLR files. Objectives of the analyses were to cover a large area very quickly and to provide data that could be used at a scale of 1:100,000 or smaller. Thus, the image processing was standardized for speed while still achieving the desired 1:100,000-scale level of detail. Consequently, some subtle features of mineralogy may be missed. The hydrothermal alteration data were not field checked to separate mineral occurrences due to hydrothermal alteration from those due to other natural occurrences. The data were evaluated by overlaying the results with 1:100,000 scale topographic maps to confirm correlation with known mineralized areas. The data were also tested in the Battle Mountain area of north-central Nevada by a weights-of-evidence correlation analysis with metallic mineral sites from the USGS Mineral Resources Data System and were found to have significant spatial correlation. On the basis of on these analyses, the data are considered useful for regional studies at scales of 1:100,000.
Error Estimates of the Ares I Computed Turbulent Ascent Longitudinal Aerodynamic Analysis
NASA Technical Reports Server (NTRS)
Abdol-Hamid, Khaled S.; Ghaffari, Farhad
2012-01-01
Numerical predictions of the longitudinal aerodynamic characteristics for the Ares I class of vehicles, along with the associated error estimate derived from an iterative convergence grid refinement, are presented. Computational results are based on an unstructured grid, Reynolds-averaged Navier-Stokes analysis. The validity of the approach to compute the associated error estimates, derived from a base grid to an extrapolated infinite-size grid, was first demonstrated on a sub-scaled wind tunnel model at representative ascent flow conditions for which the experimental data existed. Such analysis at the transonic flow conditions revealed a maximum deviation of about 23% between the computed longitudinal aerodynamic coefficients with the base grid and the measured data across the entire roll angles. This maximum deviation from the wind tunnel data was associated with the computed normal force coefficient at the transonic flow condition and was reduced to approximately 16% based on the infinite-size grid. However, all the computed aerodynamic coefficients with the base grid at the supersonic flow conditions showed a maximum deviation of only about 8% with that level being improved to approximately 5% for the infinite-size grid. The results and the error estimates based on the established procedure are also presented for the flight flow conditions.
Siciliano, Mattia; Raimo, Simona; Tufano, Dario; Basile, Giuseppe; Grossi, Dario; Santangelo, Franco; Trojano, Luigi; Santangelo, Gabriella
2016-03-01
The Addenbrooke's Cognitive Examination Revised (ACE-R) is a rapid screening battery, including five sub-scales to explore different cognitive domains: attention/orientation, memory, fluency, language and visuospatial. ACE-R is considered useful in discriminating cognitively normal subjects from patients with mild dementia. The aim of present study was to provide normative values for ACE-R total score and sub-scale scores in a large sample of Italian healthy subjects. Five hundred twenty-six Italian healthy subjects (282 women and 246 men) of different ages (age range 20-93 years) and educational level (from primary school to university) underwent ACE-R and Montreal Cognitive Assessment (MoCA). Multiple linear regression analysis revealed that age and education significantly influenced performance on ACE-R total score and sub-scale scores. A significant effect of gender was found only in sub-scale attention/orientation. From the derived linear equation, a correction grid for raw scores was built. Inferential cut-offs score were estimated using a non-parametric technique and equivalent scores (ES) were computed. Correlation analysis showed a good significant correlation between ACE-R adjusted scores with MoCA adjusted scores (r = 0.612, p < 0.001). The present study provided normative data for the ACE-R in an Italian population useful for both clinical and research purposes.
The modelling of dispersion in 2-D tidal flow over an uneven bed
NASA Astrophysics Data System (ADS)
Kalkwijk, Jan P. Th.
This paper deals with the effective mixing by topographic induced velocity variations in 2-D tidal flow. This type of mixing is characterized by tidally-averaged dispersion coefficients, which depend on the magnitude of the depth variations with respect to a mean depth, the velocity variations and the basic dispersion coefficients. The analysis is principally based on a Taylor type approximation (large clouds, small concentration variations) of the 2-D advection diffusion equation and a 2-D velocity field that behaves harmonically both in time and in space. Neglecting transient phenomena and applying time and space averaging the effective dispersion coefficients can be derived. Under certain circumstances it is possible to relate the velocity variations to the depth variations, so that finally effective dispersion coefficients can be determined using the power spectrum of the depth variations. In a special paragraph attention is paid to the modelling of sub-grid mixing in case of numerical integration of the advection-diffusion equation. It appears that the dispersion coefficients taking account of the sub-grid mixing are not only determined by the velocity variations within a certain grid cell, but also by the velocity variations at a larger scale.
Utility photovoltaic group: Status report
NASA Astrophysics Data System (ADS)
Serfass, Jeffrey A.; Hester, Stephen L.; Wills, Bethany N.
1996-01-01
The Utility PhotoVoltaic Group (UPVG) was formed in October of 1992 with a mission to accelerate the use of cost-effective small-scale and emerging grid-connected applications of photovoltaics for the benefit of electric utilities and their customers. The UPVG is now implementing a program to install up to 50 megawatts of photovoltaics in small-scale and grid-connected applications. This program, called TEAM-UP, is a partnership of the U.S. electric utility industry and the U.S. Department of Energy to help develop utility PV markets. TEAM-UP is a utility-directed program to significantly increase utility PV experience by promoting installations of utility PV systems. Two primary program areas are proposed for TEAM-UP: (1) Small-Scale Applications (SSA)—an initiative to aggregate utility purchases of small-scale, grid-independent applications; and (2) Grid-Connected Applications (GCA)—an initiative to identify and competitively award cost-sharing contracts for grid-connected PV systems with high market growth potential, or collective purchase programs involving multiple buyers. This paper describes these programs and outlines the schedule, the procurement status, and the results of the TEAM-UP process.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Makwana, K. D., E-mail: kirit.makwana@gmx.com; Cattaneo, F.; Zhdankin, V.
Simulations of decaying magnetohydrodynamic (MHD) turbulence are performed with a fluid and a kinetic code. The initial condition is an ensemble of long-wavelength, counter-propagating, shear-Alfvén waves, which interact and rapidly generate strong MHD turbulence. The total energy is conserved and the rate of turbulent energy decay is very similar in both codes, although the fluid code has numerical dissipation, whereas the kinetic code has kinetic dissipation. The inertial range power spectrum index is similar in both the codes. The fluid code shows a perpendicular wavenumber spectral slope of k{sub ⊥}{sup −1.3}. The kinetic code shows a spectral slope of k{submore » ⊥}{sup −1.5} for smaller simulation domain, and k{sub ⊥}{sup −1.3} for larger domain. We estimate that collisionless damping mechanisms in the kinetic code can account for the dissipation of the observed nonlinear energy cascade. Current sheets are geometrically characterized. Their lengths and widths are in good agreement between the two codes. The length scales linearly with the driving scale of the turbulence. In the fluid code, their thickness is determined by the grid resolution as there is no explicit diffusivity. In the kinetic code, their thickness is very close to the skin-depth, irrespective of the grid resolution. This work shows that kinetic codes can reproduce the MHD inertial range dynamics at large scales, while at the same time capturing important kinetic physics at small scales.« less
Evaluation of subgrid-scale turbulence models using a fully simulated turbulent flow
NASA Technical Reports Server (NTRS)
Clark, R. A.; Ferziger, J. H.; Reynolds, W. C.
1977-01-01
An exact turbulent flow field was calculated on a three-dimensional grid with 64 points on a side. The flow simulates grid-generated turbulence from wind tunnel experiments. In this simulation, the grid spacing is small enough to include essentially all of the viscous energy dissipation, and the box is large enough to contain the largest eddy in the flow. The method is limited to low-turbulence Reynolds numbers, in our case R sub lambda = 36.6. To complete the calculation using a reasonable amount of computer time with reasonable accuracy, a third-order time-integration scheme was developed which runs at about the same speed as a simple first-order scheme. It obtains this accuracy by saving the velocity field and its first-time derivative at each time step. Fourth-order accurate space-differencing is used.
Computational System For Rapid CFD Analysis In Engineering
NASA Technical Reports Server (NTRS)
Barson, Steven L.; Ascoli, Edward P.; Decroix, Michelle E.; Sindir, Munir M.
1995-01-01
Computational system comprising modular hardware and software sub-systems developed to accelerate and facilitate use of techniques of computational fluid dynamics (CFD) in engineering environment. Addresses integration of all aspects of CFD analysis process, including definition of hardware surfaces, generation of computational grids, CFD flow solution, and postprocessing. Incorporates interfaces for integration of all hardware and software tools needed to perform complete CFD analysis. Includes tools for efficient definition of flow geometry, generation of computational grids, computation of flows on grids, and postprocessing of flow data. System accepts geometric input from any of three basic sources: computer-aided design (CAD), computer-aided engineering (CAE), or definition by user.
NASA Astrophysics Data System (ADS)
Tian, Zhang; Yanfeng, Gong
2017-05-01
In order to solve the contradiction between demand and distribution range of primary energy resource, Ultra High Voltage (UHV) power grids should be developed rapidly to meet development of energy bases and accessing of large-scale renewable energy. This paper reviewed the latest research processes of AC/DC transmission technologies, summarized the characteristics of AC/DC power grids, concluded that China’s power grids certainly enter a new period of large -scale hybrid UHV AC/DC power grids and characteristics of “strong DC and weak AC” becomes increasingly pro minent; possible problems in operation of AC/DC power grids was discussed, and interaction or effect between AC/DC power grids was made an intensive study of; according to above problems in operation of power grids, preliminary scheme is summarized as fo llows: strengthening backbone structures, enhancing AC/DC transmission technologies, promoting protection measures of clean energ y accessing grids, and taking actions to solve stability problems of voltage and frequency etc. It’s valuable for making hybrid UHV AC/DC power grids adapt to operating mode of large power grids, thus guaranteeing security and stability of power system.
NASA Technical Reports Server (NTRS)
Zhang, Yao; Xiao, Xiangming; Jin, Cui; Dong, Jinwei; Zhou, Sha; Wagle, Pradeep; Joiner, Joanna; Guanter, Luis; Zhang, Yongguang; Zhang , Geli;
2016-01-01
Accurate estimation of the gross primary production (GPP) of terrestrial ecosystems is vital for a better understanding of the spatial-temporal patterns of the global carbon cycle. In this study,we estimate GPP in North America (NA) using the satellite-based Vegetation Photosynthesis Model (VPM), MODIS (Moderate Resolution Imaging Spectrometer) images at 8-day temporal and 500 meter spatial resolutions, and NCEP-NARR (National Center for Environmental Prediction-North America Regional Reanalysis) climate data. The simulated GPP (GPP (sub VPM)) agrees well with the flux tower derived GPP (GPPEC) at 39 AmeriFlux sites (155 site-years). The GPP (sub VPM) in 2010 is spatially aggregated to 0.5 by 0.5-degree grid cells and then compared with sun-induced chlorophyll fluorescence (SIF) data from Global Ozone Monitoring Instrument 2 (GOME-2), which is directly related to vegetation photosynthesis. Spatial distribution and seasonal dynamics of GPP (sub VPM) and GOME-2 SIF show good consistency. At the biome scale, GPP (sub VPM) and SIF shows strong linear relationships (R (sup 2) is greater than 0.95) and small variations in regression slopes ((4.60-5.55 grams Carbon per square meter per day) divided by (milliwatts per square meter per nanometer per square radian)). The total annual GPP (sub VPM) in NA in 2010 is approximately 13.53 petagrams Carbon per year, which accounts for approximately 11.0 percent of the global terrestrial GPP and is within the range of annual GPP estimates from six other process-based and data-driven models (11.35-22.23 petagrams Carbon per year). Among the seven models, some models did not capture the spatial pattern of GOME-2 SIF data at annual scale, especially in Midwest cropland region. The results from this study demonstrate the reliable performance of VPM at the continental scale, and the potential of SIF data being used as a benchmark to compare with GPP models.
NASA Astrophysics Data System (ADS)
Juvela, Mika J.
The relationship between physical conditions of an interstellar cloud and the observed radiation is defined by the radiative transfer problem. Radiative transfer calculations are needed if, e.g., one wants to disentangle abundance variations from excitation effects or wants to model variations of dust properties inside an interstellar cloud. New observational facilities (e.g., ALMA and Herschel) will bring improved accuracy both in terms of intensity and spatial resolution. This will enable detailed studies of the densest sub-structures of interstellar clouds and star forming regions. Such observations must be interpreted with accurate radiative transfer methods and realistic source models. In many cases this will mean modelling in three dimensions. High optical depths and observed wide range of linear scales are, however, challenging for radiative transfer modelling. A large range of linear scales can be accessed only with hierarchical models. Figure 1 shows an example of the use of a hierarchical grid for radiative transfer calculations when the original model cloud (L=10 pc,
Grossberg, Stephen; Pilly, Praveen K
2014-02-05
A neural model proposes how entorhinal grid cells and hippocampal place cells may develop as spatial categories in a hierarchy of self-organizing maps (SOMs). The model responds to realistic rat navigational trajectories by learning both grid cells with hexagonal grid firing fields of multiple spatial scales, and place cells with one or more firing fields, that match neurophysiological data about their development in juvenile rats. Both grid and place cells can develop by detecting, learning and remembering the most frequent and energetic co-occurrences of their inputs. The model's parsimonious properties include: similar ring attractor mechanisms process linear and angular path integration inputs that drive map learning; the same SOM mechanisms can learn grid cell and place cell receptive fields; and the learning of the dorsoventral organization of multiple spatial scale modules through medial entorhinal cortex to hippocampus (HC) may use mechanisms homologous to those for temporal learning through lateral entorhinal cortex to HC ('neural relativity'). The model clarifies how top-down HC-to-entorhinal attentional mechanisms may stabilize map learning, simulates how hippocampal inactivation may disrupt grid cells, and explains data about theta, beta and gamma oscillations. The article also compares the three main types of grid cell models in the light of recent data.
Influence of grid resolution, parcel size and drag models on bubbling fluidized bed simulation
Lu, Liqiang; Konan, Arthur; Benyahia, Sofiane
2017-06-02
Here in this paper, a bubbling fluidized bed is simulated with different numerical parameters, such as grid resolution and parcel size. We examined also the effect of using two homogeneous drag correlations and a heterogeneous drag based on the energy minimization method. A fast and reliable bubble detection algorithm was developed based on the connected component labeling. The radial and axial solids volume fraction profiles are compared with experiment data and previous simulation results. These results show a significant influence of drag models on bubble size and voidage distributions and a much less dependence on numerical parameters. With a heterogeneousmore » drag model that accounts for sub-scale structures, the void fraction in the bubbling fluidized bed can be well captured with coarse grid and large computation parcels. Refining the CFD grid and reducing the parcel size can improve the simulation results but with a large increase in computation cost.« less
NASA Technical Reports Server (NTRS)
Selkirk, Henry B.; Molod, Andrea M.
2014-01-01
Large-scale models such as GEOS-5 typically calculate grid-scale fractional cloudiness through a PDF parameterization of the sub-gridscale distribution of specific humidity. The GEOS-5 moisture routine uses a simple rectangular PDF varying in height that follows a tanh profile. While below 10 km this profile is informed by moisture information from the AIRS instrument, there is relatively little empirical basis for the profile above that level. ATTREX provides an opportunity to refine the profile using estimates of the horizontal variability of measurements of water vapor, total water and ice particles from the Global Hawk aircraft at or near the tropopause. These measurements will be compared with estimates of large-scale cloud fraction from CALIPSO and lidar retrievals from the CPL on the aircraft. We will use the variability measurements to perform studies of the sensitivity of the GEOS-5 cloud-fraction to various modifications to the PDF shape and to its vertical profile.
Automatic digital surface model (DSM) generation from aerial imagery data
NASA Astrophysics Data System (ADS)
Zhou, Nan; Cao, Shixiang; He, Hongyan; Xing, Kun; Yue, Chunyu
2018-04-01
Aerial sensors are widely used to acquire imagery for photogrammetric and remote sensing application. In general, the images have large overlapped region, which provide a lot of redundant geometry and radiation information for matching. This paper presents a POS supported dense matching procedure for automatic DSM generation from aerial imagery data. The method uses a coarse-to-fine hierarchical strategy with an effective combination of several image matching algorithms: image radiation pre-processing, image pyramid generation, feature point extraction and grid point generation, multi-image geometrically constraint cross-correlation (MIG3C), global relaxation optimization, multi-image geometrically constrained least squares matching (MIGCLSM), TIN generation and point cloud filtering. The image radiation pre-processing is used in order to reduce the effects of the inherent radiometric problems and optimize the images. The presented approach essentially consists of 3 components: feature point extraction and matching procedure, grid point matching procedure and relational matching procedure. The MIGCLSM method is used to achieve potentially sub-pixel accuracy matches and identify some inaccurate and possibly false matches. The feasibility of the method has been tested on different aerial scale images with different landcover types. The accuracy evaluation is based on the comparison between the automatic extracted DSMs derived from the precise exterior orientation parameters (EOPs) and the POS.
NASA Astrophysics Data System (ADS)
Michioka, Takenobu; Sato, Ayumu; Sada, Koichi
2011-10-01
Large-scale turbulent motions enhancing horizontal gas spread in an atmospheric boundary layer are simulated in a wind-tunnel experiment. The large-scale turbulent motions can be generated using an active grid installed at the front of the test section in the wind tunnel, when appropriate parameters for the angular deflection and the rotation speed are chosen. The power spectra of vertical velocity fluctuations are unchanged with and without the active grid because they are strongly affected by the surface. The power spectra of both streamwise and lateral velocity fluctuations with the active grid increase in the low frequency region, and are closer to the empirical relations inferred from field observations. The large-scale turbulent motions do not affect the Reynolds shear stress, but change the balance of the processes involved. The relative contributions of ejections to sweeps are suppressed by large-scale turbulent motions, indicating that the motions behave as sweep events. The lateral gas spread is enhanced by the lateral large-scale turbulent motions generated by the active grid. The large-scale motions, however, do not affect the vertical velocity fluctuations near the surface, resulting in their having a minimal effect on the vertical gas spread. The peak concentration normalized using the root-mean-squared value of concentration fluctuation is remarkably constant over most regions of the plume irrespective of the operation of the active grid.
NASA Astrophysics Data System (ADS)
Baker, Kirk R.; Hawkins, Andy; Kelly, James T.
2014-12-01
Near source modeling is needed to assess primary and secondary pollutant impacts from single sources and single source complexes. Source-receptor relationships need to be resolved from tens of meters to tens of kilometers. Dispersion models are typically applied for near-source primary pollutant impacts but lack complex photochemistry. Photochemical models provide a realistic chemical environment but are typically applied using grid cell sizes that may be larger than the distance between sources and receptors. It is important to understand the impacts of grid resolution and sub-grid plume treatments on photochemical modeling of near-source primary pollution gradients. Here, the CAMx photochemical grid model is applied using multiple grid resolutions and sub-grid plume treatment for SO2 and compared with a receptor mesonet largely impacted by nearby sources approximately 3-17 km away in a complex terrain environment. Measurements are compared with model estimates of SO2 at 4- and 1-km resolution, both with and without sub-grid plume treatment and inclusion of finer two-way grid nests. Annual average estimated SO2 mixing ratios are highest nearest the sources and decrease as distance from the sources increase. In general, CAMx estimates of SO2 do not compare well with the near-source observations when paired in space and time. Given the proximity of these sources and receptors, accuracy in wind vector estimation is critical for applications that pair pollutant predictions and observations in time and space. In typical permit applications, predictions and observations are not paired in time and space and the entire distributions of each are directly compared. Using this approach, model estimates using 1-km grid resolution best match the distribution of observations and are most comparable to similar studies that used dispersion and Lagrangian modeling systems. Model-estimated SO2 increases as grid cell size decreases from 4 km to 250 m. However, it is notable that the 1-km model estimates using 1-km meteorological model input are higher than the 1-km model simulation that used interpolated 4-km meteorology. The inclusion of sub-grid plume treatment did not improve model skill in predicting SO2 in time and space and generally acts to keep emitted mass aloft.
NASA Astrophysics Data System (ADS)
Barbera, Roberto; Donvit, Giacinto; Falzone, Alberto; Rocca, Giuseppe La; Maggi, Giorgio Pietro; Milanesi, Luciano; Vicarioicario, Saverio
This paper depicts the solution proposed by INFN to allow users, not owning a personal digital certificate and therefore not belonging to any specific Virtual Organization (VO), to access Grid infrastructures via the GENIUS Grid portal enabled with robot certificates. Robot certificates, also known as portal certificates, are associated with a specific application that the user wants to share with the whole Grid community and have recently been introduced by the EUGridPMA (European Policy Management Authority for Grid Authentication) to perform automated tasks on Grids on behalf of users. They are proven to be extremely useful to automate grid service monitoring, data processing production, distributed data collection systems, etc. In this paper, robot certificates have been used to allow bioinformaticians involved in the Italian LIBI project to perform large scale phylogenetic analyses. The distributed environment set up in this work strongly simplify the grid access of occasional users and represents a valuable step forward to wide the communities of users.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rai, Raj K.; Berg, Larry K.; Pekour, Mikhail
The assumption of sub-grid scale (SGS) horizontal homogeneity within a model grid cell, which forms the basis of SGS turbulence closures used by mesoscale models, becomes increasingly tenuous as grid spacing is reduced to a few kilometers or less, such as in many emerging high-resolution applications. Herein, we use the turbulence kinetic energy (TKE) budget equation to study the spatio-temporal variability in two types of terrain—complex (Columbia Basin Wind Energy Study [CBWES] site, north-eastern Oregon) and flat (ScaledWind Farm Technologies [SWiFT] site, west Texas) using the Weather Research and Forecasting (WRF) model. In each case six-nested domains (three domains eachmore » for mesoscale and large-eddy simulation [LES]) are used to downscale the horizontal grid spacing from 10 km to 10 m using the WRF model framework. The model output was used to calculate the values of the TKE budget terms in vertical and horizontal planes as well as the averages of grid cells contained in the four quadrants (a quarter area) of the LES domain. The budget terms calculated along the planes and the mean profile of budget terms show larger spatial variability at CBWES site than at the SWiFT site. The contribution of the horizontal derivative of the shear production term to the total production shear was found to be 45% and 15% of the total shear, at the CBWES and SWiFT sites, respectively, indicating that the horizontal derivatives applied in the budget equation should not be ignored in mesoscale model parameterizations, especially for cases with complex terrain with <10 km scale.« less
NASA Astrophysics Data System (ADS)
Gan, Chee Kwan; Challacombe, Matt
2003-05-01
Recently, early onset linear scaling computation of the exchange-correlation matrix has been achieved using hierarchical cubature [J. Chem. Phys. 113, 10037 (2000)]. Hierarchical cubature differs from other methods in that the integration grid is adaptive and purely Cartesian, which allows for a straightforward domain decomposition in parallel computations; the volume enclosing the entire grid may be simply divided into a number of nonoverlapping boxes. In our data parallel approach, each box requires only a fraction of the total density to perform the necessary numerical integrations due to the finite extent of Gaussian-orbital basis sets. This inherent data locality may be exploited to reduce communications between processors as well as to avoid memory and copy overheads associated with data replication. Although the hierarchical cubature grid is Cartesian, naive boxing leads to irregular work loads due to strong spatial variations of the grid and the electron density. In this paper we describe equal time partitioning, which employs time measurement of the smallest sub-volumes (corresponding to the primitive cubature rule) to load balance grid-work for the next self-consistent-field iteration. After start-up from a heuristic center of mass partitioning, equal time partitioning exploits smooth variation of the density and grid between iterations to achieve load balance. With the 3-21G basis set and a medium quality grid, equal time partitioning applied to taxol (62 heavy atoms) attained a speedup of 61 out of 64 processors, while for a 110 molecule water cluster at standard density it achieved a speedup of 113 out of 128. The efficiency of equal time partitioning applied to hierarchical cubature improves as the grid work per processor increases. With a fine grid and the 6-311G(df,p) basis set, calculations on the 26 atom molecule α-pinene achieved a parallel efficiency better than 99% with 64 processors. For more coarse grained calculations, superlinear speedups are found to result from reduced computational complexity associated with data parallelism.
NASA Astrophysics Data System (ADS)
Dipankar, A.; Stevens, B. B.; Zängl, G.; Pondkule, M.; Brdar, S.
2014-12-01
The effect of clouds on large scale dynamics is represented in climate models through parameterization of various processes, of which the parameterization of shallow and deep convection are particularly uncertain. The atmospheric boundary layer, which controls the coupling to the surface, and which defines the scale of shallow convection, is typically 1 km in depth. Thus, simulations on a O(100 m) grid largely obviate the need for such parameterizations. By crossing this threshold of O(100m) grid resolution one can begin thinking of large-eddy simulation (LES), wherein the sub-grid scale parameterization have a sounder theoretical foundation. Substantial initiatives have been taken internationally to approach this threshold. For example, Miura et al., 2007 and Mirakawa et al., 2014 approach this threshold by doing global simulations, with (gradually) decreasing grid resolution, to understand the effect of cloud-resolving scales on the general circulation. Our strategy, on the other hand, is to take a big leap forward by fixing the resolution at O(100 m), and gradually increasing the domain size. We believe that breaking this threshold would greatly help in improving the parameterization schemes and reducing the uncertainty in climate predictions. To take this forward, the German Federal Ministry of Education and Research has initiated a project on HD(CP)2 that aims for a limited area LES at resolution O(100 m) using the new unified modeling system ICON (Zängl et al., 2014). In the talk, results from the HD(CP)2 evaluation simulation will be shown that targets high resolution simulation over a small domain around Jülich, Germany. This site is chosen because high resolution HD(CP)2 Observational Prototype Experiment took place in this region from 1.04.2013 to 31.05.2013, in order to critically evaluate the model. Nesting capabilities of ICON is used to gradually increase the resolution from the outermost domain, which is forced from the COSMO-DE data, to the innermost and finest resolution domain centered around Jülich (see Fig. 1 top panel). Furthermore, detailed analyses of the simulation results against the observation data will be presented. A reprsentative figure showing time series of column integrated water vapor (IWV) for both model and observation on 24.04.2013 is shown in bottom panel of Fig. 1.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wardle, Kent E.; Frey, Kurt; Pereira, Candido
2014-02-02
This task is aimed at predictive modeling of solvent extraction processes in typical extraction equipment through multiple simulation methods at various scales of resolution. We have conducted detailed continuum fluid dynamics simulation on the process unit level as well as simulations of the molecular-level physical interactions which govern extraction chemistry. Through combination of information gained through simulations at each of these two tiers along with advanced techniques such as the Lattice Boltzmann Method (LBM) which can bridge these two scales, we can develop the tools to work towards predictive simulation for solvent extraction on the equipment scale (Figure 1). Themore » goal of such a tool-along with enabling optimized design and operation of extraction units-would be to allow prediction of stage extraction effrciency under specified conditions. Simulation efforts on each of the two scales will be described below. As the initial application of FELBM in the work performed during FYl0 has been on annular mixing it will be discussed in context of the continuum-scale. In the future, however, it is anticipated that the real value of FELBM will be in its use as a tool for sub-grid model development through highly refined DNS-like multiphase simulations facilitating exploration and development of droplet models including breakup and coalescence which will be needed for the large-scale simulations where droplet level physics cannot be resolved. In this area, it can have a significant advantage over traditional CFD methods as its high computational efficiency allows exploration of significantly greater physical detail especially as computational resources increase in the future.« less
NeuroGrid: recording action potentials from the surface of the brain.
Khodagholy, Dion; Gelinas, Jennifer N; Thesen, Thomas; Doyle, Werner; Devinsky, Orrin; Malliaras, George G; Buzsáki, György
2015-02-01
Recording from neural networks at the resolution of action potentials is critical for understanding how information is processed in the brain. Here, we address this challenge by developing an organic material-based, ultraconformable, biocompatible and scalable neural interface array (the 'NeuroGrid') that can record both local field potentials(LFPs) and action potentials from superficial cortical neurons without penetrating the brain surface. Spikes with features of interneurons and pyramidal cells were simultaneously acquired by multiple neighboring electrodes of the NeuroGrid, allowing for the isolation of putative single neurons in rats. Spiking activity demonstrated consistent phase modulation by ongoing brain oscillations and was stable in recordings exceeding 1 week's duration. We also recorded LFP-modulated spiking activity intraoperatively in patients undergoing epilepsy surgery. The NeuroGrid constitutes an effective method for large-scale, stable recording of neuronal spikes in concert with local population synaptic activity, enhancing comprehension of neural processes across spatiotemporal scales and potentially facilitating diagnosis and therapy for brain disorders.
Renormalization group analysis of turbulence
NASA Technical Reports Server (NTRS)
Smith, Leslie M.
1989-01-01
The objective is to understand and extend a recent theory of turbulence based on dynamic renormalization group (RNG) techniques. The application of RNG methods to hydrodynamic turbulence was explored most extensively by Yakhot and Orszag (1986). An eddy viscosity was calculated which was consistent with the Kolmogorov inertial range by systematic elimination of the small scales in the flow. Further, assumed smallness of the nonlinear terms in the redefined equations for the large scales results in predictions for important flow constants such as the Kolmogorov constant. It is emphasized that no adjustable parameters are needed. The parameterization of the small scales in a self-consistent manner has important implications for sub-grid modeling.
Grid cell hexagonal patterns formed by fast self-organized learning within entorhinal cortex.
Mhatre, Himanshu; Gorchetchnikov, Anatoli; Grossberg, Stephen
2012-02-01
Grid cells in the dorsal segment of the medial entorhinal cortex (dMEC) show remarkable hexagonal activity patterns, at multiple spatial scales, during spatial navigation. It has previously been shown how a self-organizing map can convert firing patterns across entorhinal grid cells into hippocampal place cells that are capable of representing much larger spatial scales. Can grid cell firing fields also arise during navigation through learning within a self-organizing map? This article describes a simple and general mathematical property of the trigonometry of spatial navigation which favors hexagonal patterns. The article also develops a neural model that can learn to exploit this trigonometric relationship. This GRIDSmap self-organizing map model converts path integration signals into hexagonal grid cell patterns of multiple scales. GRIDSmap creates only grid cell firing patterns with the observed hexagonal structure, predicts how these hexagonal patterns can be learned from experience, and can process biologically plausible neural input and output signals during navigation. These results support an emerging unified computational framework based on a hierarchy of self-organizing maps for explaining how entorhinal-hippocampal interactions support spatial navigation. Copyright © 2010 Wiley Periodicals, Inc.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Donner, Leo J.; O'Brien, Travis A.; Rieger, Daniel
Both climate forcing and climate sensitivity persist as stubborn uncertainties limiting the extent to which climate models can provide actionable scientific scenarios for climate change. A key, explicit control on cloud-aerosol interactions, the largest uncertainty in climate forcing, is the vertical velocity of cloud-scale updrafts. Model-based studies of climate sensitivity indicate that convective entrainment, which is closely related to updraft speeds, is an important control on climate sensitivity. Updraft vertical velocities also drive many physical processes essential to numerical weather prediction. Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climatemore » and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying vertical velocities, and parameterizations which do provide vertical velocities have been subject to limited evaluation against what have until recently been scant observations. Atmospheric observations imply that the distribution of vertical velocities depends on the areas over which the vertical velocities are averaged. Distributions of vertical velocities in climate models may capture this behavior, but it has not been accounted for when parameterizing cloud and precipitation processes in current models. New observations of convective vertical velocities offer a potentially promising path toward developing process-level cloud models and parameterizations for climate and numerical weather prediction. Taking account of scale-dependence of resolved vertical velocities offers a path to matching cloud-scale physical processes and their driving dynamics more realistically, with a prospect of reduced uncertainty in both climate forcing and sensitivity.« less
Part 2 of a Computational Study of a Drop-Laden Mixing Layer
NASA Technical Reports Server (NTRS)
Okongo, Nora; Bellan, Josette
2004-01-01
This second of three reports on a computational study of a mixing layer laden with evaporating liquid drops presents the evaluation of Large Eddy Simulation (LES) models. The LES models were evaluated on an existing database that had been generated using Direct Numerical Simulation (DNS). The DNS method and the database are described in the first report of this series, Part 1 of a Computational Study of a Drop-Laden Mixing Layer (NPO-30719), NASA Tech Briefs, Vol. 28, No.7 (July 2004), page 59. The LES equations, which are derived by applying a spatial filter to the DNS set, govern the evolution of the larger scales of the flow and can therefore be solved on a coarser grid. Consistent with the reduction in grid points, the DNS drops would be represented by fewer drops, called computational drops in the LES context. The LES equations contain terms that cannot be directly computed on the coarser grid and that must instead be modeled. Two types of models are necessary: (1) those for the filtered source terms representing the effects of drops on the filtered flow field and (2) those for the sub-grid scale (SGS) fluxes arising from filtering the convective terms in the DNS equations. All of the filtered-sourceterm models that were developed were found to overestimate the filtered source terms. For modeling the SGS fluxes, constant-coefficient Smagorinsky, gradient, and scale-similarity models were assessed and calibrated on the DNS database. The Smagorinsky model correlated poorly with the SGS fluxes, whereas the gradient and scale-similarity models were well correlated with the SGS quantities that they represented.
Di Sarli, Valeria; Di Benedetto, Almerinda; Russo, Gennaro
2010-08-15
In this work, an assessment of different sub-grid scale (sgs) combustion models proposed for large eddy simulation (LES) of steady turbulent premixed combustion (Colin et al., Phys. Fluids 12 (2000) 1843-1863; Flohr and Pitsch, Proc. CTR Summer Program, 2000, pp. 61-82; Kim and Menon, Combust. Sci. Technol. 160 (2000) 119-150; Charlette et al., Combust. Flame 131 (2002) 159-180; Pitsch and Duchamp de Lageneste, Proc. Combust. Inst. 29 (2002) 2001-2008) was performed to identify the model that best predicts unsteady flame propagation in gas explosions. Numerical results were compared to the experimental data by Patel et al. (Proc. Combust. Inst. 29 (2002) 1849-1854) for premixed deflagrating flame in a vented chamber in the presence of three sequential obstacles. It is found that all sgs combustion models are able to reproduce qualitatively the experiment in terms of step of flame acceleration and deceleration around each obstacle, and shape of the propagating flame. Without adjusting any constants and parameters, the sgs model by Charlette et al. also provides satisfactory quantitative predictions for flame speed and pressure peak. Conversely, the sgs combustion models other than Charlette et al. give correct predictions only after an ad hoc tuning of constants and parameters. Copyright 2010 Elsevier B.V. All rights reserved.
Stochastic four-way coupling of gas-solid flows for Large Eddy Simulations
NASA Astrophysics Data System (ADS)
Curran, Thomas; Denner, Fabian; van Wachem, Berend
2017-11-01
The interaction of solid particles with turbulence has for long been a topic of interest for predicting the behavior of industrially relevant flows. For the turbulent fluid phase, Large Eddy Simulation (LES) methods are widely used for their low computational cost, leaving only the sub-grid scales (SGS) of turbulence to be modelled. Although LES has seen great success in predicting the behavior of turbulent single-phase flows, the development of LES for turbulent gas-solid flows is still in its infancy. This contribution aims at constructing a model to describe the four-way coupling of particles in an LES framework, by considering the role particles play in the transport of turbulent kinetic energy across the scales. Firstly, a stochastic model reconstructing the sub-grid velocities for the particle tracking is presented. Secondly, to solve particle-particle interaction, most models involve a deterministic treatment of the collisions. We finally introduce a stochastic model for estimating the collision probability. All results are validated against fully resolved DNS-DPS simulations. The final goal of this contribution is to propose a global stochastic method adapted to two-phase LES simulation where the number of particles considered can be significantly increased. Financial support from PetroBras is gratefully acknowledged.
Toward a Unified Representation of Atmospheric Convection in Variable-Resolution Climate Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Walko, Robert
2016-11-07
The purpose of this project was to improve the representation of convection in atmospheric weather and climate models that employ computational grids with spatially-variable resolution. Specifically, our work targeted models whose grids are fine enough over selected regions that convection is resolved explicitly, while over other regions the grid is coarser and convection is represented as a subgrid-scale process. The working criterion for a successful scheme for representing convection over this range of grid resolution was that identical convective environments must produce very similar convective responses (i.e., the same precipitation amount, rate, and timing, and the same modification of themore » atmospheric profile) regardless of grid scale. The need for such a convective scheme has increased in recent years as more global weather and climate models have adopted variable resolution meshes that are often extended into the range of resolving convection in selected locations.« less
Matching soil grid unit resolutions with polygon unit scales for DNDC modelling of regional SOC pool
NASA Astrophysics Data System (ADS)
Zhang, H. D.; Yu, D. S.; Ni, Y. L.; Zhang, L. M.; Shi, X. Z.
2015-03-01
Matching soil grid unit resolution with polygon unit map scale is important to minimize uncertainty of regional soil organic carbon (SOC) pool simulation as their strong influences on the uncertainty. A series of soil grid units at varying cell sizes were derived from soil polygon units at the six map scales of 1:50 000 (C5), 1:200 000 (D2), 1:500 000 (P5), 1:1 000 000 (N1), 1:4 000 000 (N4) and 1:14 000 000 (N14), respectively, in the Tai lake region of China. Both format soil units were used for regional SOC pool simulation with DeNitrification-DeComposition (DNDC) process-based model, which runs span the time period 1982 to 2000 at the six map scales, respectively. Four indices, soil type number (STN) and area (AREA), average SOC density (ASOCD) and total SOC stocks (SOCS) of surface paddy soils simulated with the DNDC, were attributed from all these soil polygon and grid units, respectively. Subjecting to the four index values (IV) from the parent polygon units, the variation of an index value (VIV, %) from the grid units was used to assess its dataset accuracy and redundancy, which reflects uncertainty in the simulation of SOC. Optimal soil grid unit resolutions were generated and suggested for the DNDC simulation of regional SOC pool, matching with soil polygon units map scales, respectively. With the optimal raster resolution the soil grid units dataset can hold the same accuracy as its parent polygon units dataset without any redundancy, when VIV < 1% of all the four indices was assumed as criteria to the assessment. An quadratic curve regression model y = -8.0 × 10-6x2 + 0.228x + 0.211 (R2 = 0.9994, p < 0.05) was revealed, which describes the relationship between optimal soil grid unit resolution (y, km) and soil polygon unit map scale (1:x). The knowledge may serve for grid partitioning of regions focused on the investigation and simulation of SOC pool dynamics at certain map scale.
Grid scale drives the scale and long-term stability of place maps
Mallory, Caitlin S; Hardcastle, Kiah; Bant, Jason S; Giocomo, Lisa M
2018-01-01
Medial entorhinal cortex (MEC) grid cells fire at regular spatial intervals and project to the hippocampus, where place cells are active in spatially restricted locations. One feature of the grid population is the increase in grid spatial scale along the dorsal-ventral MEC axis. However, the difficulty in perturbing grid scale without impacting the properties of other functionally-defined MEC cell types has obscured how grid scale influences hippocampal coding and spatial memory. Here, we use a targeted viral approach to knock out HCN1 channels selectively in MEC, causing grid scale to expand while leaving other MEC spatial and velocity signals intact. Grid scale expansion resulted in place scale expansion in fields located far from environmental boundaries, reduced long-term place field stability and impaired spatial learning. These observations, combined with simulations of a grid-to-place cell model and position decoding of place cells, illuminate how grid scale impacts place coding and spatial memory. PMID:29335607
NASA Astrophysics Data System (ADS)
Fernández, V.; Dietrich, D. E.; Haney, R. L.; Tintoré, J.
In situ and satellite data obtained during the last ten years have shown that the circula- tion in the Mediterranean Sea is extremely complex in space, with significant features ranging from mesoscale to sub-basin and basin scale, and highly variable in time, with mesoscale to seasonal and interannual signals. Also, the steep bottom topography and the variable atmospheric conditions from one sub-basin to another, make the circula- tion to be composed of numerous energetic and narrow coastal currents, density fronts and mesoscale structures that interact at sub-basin scale with the large scale circula- tion. To simulate numerically and better understand these features, besides high grid resolution, a low numerical dispersion and low physical dissipation ocean model is required. We present the results from a 1/8z horizontal resolution numerical simula- tion of the Mediterranean Sea using DieCAST ocean model, which meets the above requirements since it is stable with low general dissipation and uses accurate fourth- order-accurate approximations with low numerical dispersion. The simulations are carried out with climatological surface forcing using monthly mean winds and relax- ation towards climatological values of temperature and salinity. The model reproduces the main features of the large basin scale circulation, as well as the seasonal variabil- ity of sub-basin scale currents that are well documented by observations in straits and channels. In addition, DieCAST brings out natural fronts and eddies that usually do not appear in numerical simulations of the Mediterranean and that lead to a natural interannual variability. The role of this intrinsic variability in the general circulation will be discussed.
RF Models for Plasma-Surface Interactions in VSim
NASA Astrophysics Data System (ADS)
Jenkins, Thomas G.; Smithe, D. N.; Pankin, A. Y.; Roark, C. M.; Zhou, C. D.; Stoltz, P. H.; Kruger, S. E.
2014-10-01
An overview of ongoing enhancements to the Plasma Discharge (PD) module of Tech-X's VSim software tool is presented. A sub-grid kinetic sheath model, developed for the accurate computation of sheath potentials near metal and dielectric-coated walls, enables the physical effects of DC and RF sheath physics to be included in macroscopic-scale plasma simulations that need not explicitly resolve sheath scale lengths. Sheath potential evolution, together with particle behavior near the sheath, can thus be simulated in complex geometries. Generalizations of the model to include sputtering, secondary electron emission, and effects from multiple ion species and background magnetic fields are summarized; related numerical results are also presented. In addition, improved tools for plasma chemistry and IEDF/EEDF visualization and modeling are discussed, as well as our initial efforts toward the development of hybrid fluid/kinetic transition capabilities within VSim. Ultimately, we aim to establish VSimPD as a robust, efficient computational tool for modeling industrial plasma processes. Supported by US DoE SBIR-I/II Award DE-SC0009501.
AgMIP 1.5°C Assessment: Mitigation and Adaptation at Coordinated Global and Regional Scales
NASA Astrophysics Data System (ADS)
Rosenzweig, C.
2016-12-01
The AgMIP 1.5°C Coordinated Global and Regional Integrated Assessments of Climate Change and Food Security (AgMIP 1.5 CGRA) is linking site-based crop and livestock models with similar models run on global grids, and then links these biophysical components with economics models and nutrition metrics at regional and global scales. The AgMIP 1.5 CGRA assessment brings together experts in climate, crop, livestock, economics, nutrition, and food security to define the 1.5°C Protocols and guide the process throughout the assessment. Scenarios are designed to consistently combine elements of intertwined storylines of future society including socioeconomic development (Shared Socioeconomic Pathways), greenhouse gas concentrations (Representative Concentration Pathways), and specific pathways of agricultural sector development (Representative Agricultural Pathways). Shared Climate Policy Assumptions will be extended to provide additional agricultural detail on mitigation and adaptation strategies. The multi-model, multi-disciplinary, multi-scale integrated assessment framework is using scenarios of economic development, adaptation, mitigation, food policy, and food security. These coordinated assessments are grounded in the expertise of AgMIP partners around the world, leading to more consistent results and messages for stakeholders, policymakers, and the scientific community. The early inclusion of nutrition and food security experts has helped to ensure that assessment outputs include important metrics upon which investment and policy decisions may be based. The CGRA builds upon existing AgMIP research groups (e.g., the AgMIP Wheat Team and the AgMIP Global Gridded Crop Modeling Initiative; GGCMI) and regional programs (e.g., AgMIP Regional Teams in Sub-Saharan Africa and South Asia), with new protocols for cross-scale and cross-disciplinary linkages to ensure the propagation of expert judgment and consistent assumptions.
NASA Technical Reports Server (NTRS)
Wood, Eric F.
1993-01-01
The objectives of the research were as follows: (1) Extend the Representative Elementary Area (RE) concept, first proposed and developed in Wood et al, (1988), to the water balance fluxes of the interstorm period (redistribution, evapotranspiration and baseflow) necessary for the analysis of long-term water balance processes. (2) Derive spatially averaged water balance model equations for spatially variable soil, topography and vegetation, over A RANGE OF CLIMATES. This is a necessary step in our goal to derive consistent hydrologic results up to GCM grid scales necessary for global climate modeling. (3) Apply the above macroscale water balance equations with remotely sensed data and begin to explore the feasibility of parameterizing the water balance constitutive equations at GCM grid scale.
Navier-Stokes simulation of rotor-body flowfield in hover using overset grids
NASA Technical Reports Server (NTRS)
Srinivasan, G. R.; Ahmad, J. U.
1993-01-01
A free-wake Navier-Stokes numerical scheme and multiple Chimera overset grids have been utilized for calculating the quasi-steady hovering flowfield of a Boeing-360 rotor mounted on an axisymmetric whirl-tower. The entire geometry of this rotor-body configuration is gridded-up with eleven different overset grids. The composite grid has 1.3 million grid points for the entire flow domain. The numerical results, obtained using coarse grids and a rigid rotor assumption, show a thrust value that is within 5% of the experimental value at a flow condition of M(sub tip) = 0.63, Theta(sub c) = 8 deg, and Re = 2.5 x 10(exp 6). The numerical method thus demonstrates the feasibility of using a multi-block scheme for calculating the flowfields of complex configurations consisting of rotating and non-rotating components.
HiPEP Ion Optics System Evaluation Using Gridlets
NASA Technical Reports Server (NTRS)
Willliams, John D.; Farnell, Cody C.; Laufer, D. Mark; Martinez, Rafael A.
2004-01-01
Experimental measurements are presented for sub-scale ion optics systems comprised of 7 and 19 aperture pairs with geometrical features that are similar to the HiPEP ion optics system. Effects of hole diameter and grid-to-grid spacing are presented as functions of applied voltage and beamlet current. Recommendations are made for the beamlet current range where the ion optics system can be safely operated without experiencing direct impingement of high energy ions on the accelerator grid surface. Measurements are also presented of the accelerator grid voltage where beam plasma electrons backstream through the ion optics system. Results of numerical simulations obtained with the ffx code are compared to both the impingement limit and backstreaming measurements. An emphasis is placed on identifying differences between measurements and simulation predictions to highlight areas where more research is needed. Relatively large effects are observed in simulations when the discharge chamber plasma properties and ion optics geometry are varied. Parameters investigated using simulations include the applied voltages, grid spacing, hole-to-hole spacing, doubles-to-singles ratio, plasma potential, and electron temperature; and estimates are provided for the sensitivity of impingement limits on these parameters.
OVERSMART Reporting Tool for Flow Computations Over Large Grid Systems
NASA Technical Reports Server (NTRS)
Kao, David L.; Chan, William M.
2012-01-01
Structured grid solvers such as NASA's OVERFLOW compressible Navier-Stokes flow solver can generate large data files that contain convergence histories for flow equation residuals, turbulence model equation residuals, component forces and moments, and component relative motion dynamics variables. Most of today's large-scale problems can extend to hundreds of grids, and over 100 million grid points. However, due to the lack of efficient tools, only a small fraction of information contained in these files is analyzed. OVERSMART (OVERFLOW Solution Monitoring And Reporting Tool) provides a comprehensive report of solution convergence of flow computations over large, complex grid systems. It produces a one-page executive summary of the behavior of flow equation residuals, turbulence model equation residuals, and component forces and moments. Under the automatic option, a matrix of commonly viewed plots such as residual histograms, composite residuals, sub-iteration bar graphs, and component forces and moments is automatically generated. Specific plots required by the user can also be prescribed via a command file or a graphical user interface. Output is directed to the user s computer screen and/or to an html file for archival purposes. The current implementation has been targeted for the OVERFLOW flow solver, which is used to obtain a flow solution on structured overset grids. The OVERSMART framework allows easy extension to other flow solvers.
NASA Astrophysics Data System (ADS)
Good, Garrett; Gerashchenko, Sergiy; Warhaft, Zellman
2010-11-01
Water droplets of sub-Kolmogorov size are sprayed into the turbulence side of a shearless turbulent-non-turbulent interface (TNI) as well as a turbulent-turbulent interface (TTI). An active grid is used to form the mixing layer and a splitter plate separates the droplet-non droplet interface near the origin. Particle concentration, size and velocity are determined by Phase Doppler Particle Analyzer, the velocity field by hot wires, and the droplet accelerations by particle tracking. As for a passive scalar, for the TTI, the concentration profiles are described by an error function. For the TNI, the concentration profiles fall off more rapidly than for the TTI due to the large-scale intermittency. The profile evolution and effects of initial conditions are discussed, as are the relative importance of the large and small scales in the transport process. It is shown that the concentration statistics are better described in terms of the Stokes number based on the large scales than the small, but some features of the mixing are determined by the small scales, and these will be discussed. Sponsored by the U.S. NSF.
A Priori Analyses of Three Subgrid-Scale Models for One-Parameter Families of Filters
NASA Technical Reports Server (NTRS)
Pruett, C. David; Adams, Nikolaus A.
1998-01-01
The decay of isotropic turbulence a compressible flow is examined by direct numerical simulation (DNS). A priori analyses of the DNS data are then performed to evaluate three subgrid-scale (SGS) models for large-eddy simulation (LES): a generalized Smagorinsky model (M1), a stress-similarity model (M2), and a gradient model (M3). The models exploit one-parameter second- or fourth-order filters of Pade type, which permit the cutoff wavenumber k(sub c) to be tuned independently of the grid increment (delta)x. The modeled (M) and exact (E) SGS-stresses are compared component-wise by correlation coefficients of the form C(E,M) computed over the entire three-dimensional fields. In general, M1 correlates poorly against exact stresses (C < 0.2), M3 correlates moderately well (C approx. 0.6), and M2 correlates remarkably well (0.8 < C < 1.0). Specifically, correlations C(E, M2) are high provided the grid and test filters are of the same order. Moreover, the highest correlations (C approx.= 1.0) result whenever the grid and test filters are identical (in both order and cutoff). Finally, present results reveal the exact SGS stresses obtained by grid filters of differing orders to be only moderately well correlated. Thus, in LES the model should not be specified independently of the filter.
NASA Astrophysics Data System (ADS)
Behera, Abhinna; Rivière, Emmanuel; Marécal, Virginie; Rysman, Jean-François; Claud, Chantal; Burgalat, Jérémie
2017-04-01
The stratospheric water vapour (WV) has a conceding impact on the radiative and chemical budget of Earth's atmosphere. The convective overshooting (COV) at the tropics is well admitted for playing a role in transporting directly WV to the stratosphere. Nonetheless, its impact on the lower stratosphere is yet to be determined at global scale, as the satellite and other air-borne measurements are not of having fine enough resolution to quantify this impact at large scale. Therefore, efforts have been made to quantify the influence of COV over the WV budget in the tropical tropopause layer (TTL) through modelling. Our approach is to build two synthetic tropical wet-seasons; where one would be having only deep convection (DC) but no COV at all, and the second one would be having the COV, and in both cases the WV budget in the TTL would be estimated. Before that, a French-Brazilian TRO-pico campaign was carried out at Bauru, Brazil in order to understand the influence of COV on the WV budget in the TTL. The radio-sounding, and the small balloon-borne WV measurements from the campaign are being utilized to validate the model simulation. Brazilian version of Regional Atmospheric Modeling System (BRAMS) is used with a single grid system to simulate a WV variability in a wet-season. Grell's convective parameterization with ensemble closure, microphysics with double moment scheme and 7 types of hydrometeors are incorporated to simulate the WV variability for a wet-season at the tropics. The grid size of simulation is chosen to be 20 km x 20 km horizontally and from surface to 30 km altitude, so that there cannot be COV at all, only DC due to such a relatively coarse resolution. The European Centre for Medium-range Weather Forecasts (ECMWF) operational analyses data are used every 6 hours for grid initialization and boundary conditions, and grid center nudging. The simulation is carried out for a full wet-season (Nov 2012 - Mar 2013) at Brazilian scale, so that it would coincide with the TRO-pico campaign measurements. As of first step, we have already shown that, this model with only DC is well capable of producing key features of the TTL. Hence in the second step, keeping all the settings same in the model, a sub-grid scale process/parameterization is being developed in order to reproduce COV in the model. Then, we would be able to compare these two atmospheres, and it would describe quantitatively the impact of COV on the WV budget in the TTL at a continental scale. This on-going work reports about the further advancement done to introduce the COV parameterization in BRAMS by incorporating the information from satellite-borne and balloon-borne measurements. The preliminary results of the simulation with COV nudging, achieved till date of EGU assembly, will be presented.
Laser Induced Aluminum Surface Breakdown Model
NASA Technical Reports Server (NTRS)
Chen, Yen-Sen; Liu, Jiwen; Zhang, Sijun; Wang, Ten-See (Technical Monitor)
2002-01-01
Laser powered propulsion systems involve complex fluid dynamics, thermodynamics and radiative transfer processes. Based on an unstructured grid, pressure-based computational aerothermodynamics; platform, several sub-models describing such underlying physics as laser ray tracing and focusing, thermal non-equilibrium, plasma radiation and air spark ignition have been developed. This proposed work shall extend the numerical platform and existing sub-models to include the aluminum wall surface Inverse Bremsstrahlung (IB) effect from which surface ablation and free-electron generation can be initiated without relying on the air spark ignition sub-model. The following tasks will be performed to accomplish the research objectives.
NASA Astrophysics Data System (ADS)
Hershkovitz, Yaron; Anker, Yaakov; Ben-Dor, Eyal; Schwartz, Guy; Gasith, Avital
2010-05-01
In-stream vegetation is a key ecosystem component in many fluvial ecosystems, having cascading effects on stream conditions and biotic structure. Traditionally, ground-level surveys (e.g. grid and transect analyses) are commonly used for estimating cover of aquatic macrophytes. Nonetheless, this methodological approach is highly time consuming and usually yields information which is practically limited to habitat and sub-reach scales. In contrast, remote-sensing techniques (e.g. satellite imagery and airborne photography), enable collection of large datasets over section, stream and basin scales, in relatively short time and reasonable cost. However, the commonly used spatial high resolution (1m) is often inadequate for examining aquatic vegetation on habitat or sub-reach scales. We examined the utility of a pseudo-spectral methodology, using RGB digital photography for estimating the cover of in-stream vegetation in a small Mediterranean-climate stream. We compared this methodology with that obtained by traditional ground-level grid methodology and with an airborne hyper-spectral remote sensing survey (AISA-ES). The study was conducted along a 2 km section of an intermittent stream (Taninim stream, Israel). When studied, the stream was dominated by patches of watercress (Nasturtium officinale) and mats of filamentous algae (Cladophora glomerata). The extent of vegetation cover at the habitat and section scales (100 and 104 m, respectively) were estimated by the pseudo-spectral methodology, using an airborne Roli camera with a Phase-One P 45 (39 MP) CCD image acquisition unit. The swaths were taken in elevation of about 460 m having a spatial resolution of about 4 cm (NADIR). For measuring vegetation cover at the section scale (104 m) we also used a 'push-broom' AISA-ES hyper-spectral swath having a sensor configuration of 182 bands (350-2500 nm) at elevation of ca. 1,200 m (i.e. spatial resolution of ca. 1 m). Simultaneously, with every swath we used an Analytical Spectral Device (ASD) to measure hyper-spectral signatures (2150 bands configuration; 350-2500 nm) of selected ground-level targets (located by GPS) of soil, water; vegetation (common reed, watercress, filamentous algae) and standard EVA foam colored sheets (red, green, blue, black and white). Processing and analysis of the data were performed over an ITT ENVI platform. The hyper-spectral image underwent radiometric calibration according to the flight and sensor calibration parameters on CALIGEO platform and the raw DN scale was converted into radiance scale. Ground level visual survey of vegetation cover and height was applied at the habitat scale (100 m) by placing a 1m2 netted grids (10x10cm cells) along 'bank-to-bank' transect (in triplicates). Estimates of plant cover obtained by the pseudo-spectral methodology at the habitat scale were 35-61% for the watercress, 0.4-25% for the filamentous algae and 27-51% for plant-free patches. The respective estimates by ground level visual survey were 26-50, 14-43% and 36-50%. The pseudo-spectral methodology also yielded estimates for the section scale (104 m) of ca. 39% for the watercress, ca. 32% for the filamentous algae and 6% for plant-free patches. The respective estimates obtained by hyper-spectral swath were 38, 26 and 8%. Validation against ground-level measurements proved that pseudo-spectral methodology gives reasonably good estimates of in-stream plant cover. Therefore, this methodology can serve as a substitute for ground level estimates at small stream scales and for the low resolution hyper-spectral methodology at larger scales.
NASA Astrophysics Data System (ADS)
Toyota, Takenobu; Kimura, Noriaki
2018-02-01
The validity of the sea ice rheological model formulated by Hibler (1979), which is widely used in present numerical sea ice models, is examined for the Sea of Okhotsk as an example of the seasonal ice zone (SIZ), based on satellite-derived sea ice velocity, concentration and thickness. Our focus was the formulation of the yield curve, the shape of which can be estimated from ice drift pattern based on the energy equation of deformation, while the strength of the ice cover that determines its magnitude was evaluated using ice concentration and thickness data. Ice drift was obtained with a grid spacing of 37.5 km from the AMSR-E 89 GHz brightness temperature using a maximum cross-correlation method. The ice thickness was obtained with a spatial resolution of 100 m from a regression of the PALSAR backscatter coefficients with ice thickness. To assess scale dependence, the ice drift data derived from a coastal radar covering a 70 km range in the southernmost Sea of Okhotsk were similarly analyzed. The results obtained were mostly consistent with Hibler's formulation that was based on the Arctic Ocean on both scales with no dependence on a time scale, and justify the treatment of sea ice as a plastic material, with an elliptical shaped yield curve to some extent. However, it also highlights the difficulty in parameterizing sub-grid scale ridging in the model because grid scale ice velocities reduce the deformation magnitude by half due to the large variation of the deformation field in the SIZ.
How well do the GCMs replicate the historical precipitation variability in the Colorado River Basin?
NASA Astrophysics Data System (ADS)
Guentchev, G.; Barsugli, J. J.; Eischeid, J.; Raff, D. A.; Brekke, L.
2009-12-01
Observed precipitation variability measures are compared to measures obtained using the World Climate Research Programme (WCRP) Coupled Model Intercomparison Project (CMIP3) General Circulation Models (GCM) data from 36 model projections downscaled by Brekke at al. (2007) and 30 model projections downscaled by Jon Eischeid. Three groups of variability measures are considered in this historical period (1951-1999) comparison: a) basic variability measures, such as standard deviation, interdecadal standard deviation; b) exceedance probability values, i.e., 10% (extreme wet years) and 90% (extreme dry years) exceedance probability values of series of n-year running mean annual amounts, where n=1,12; 10% exceedance probability values of annual maximum monthly precipitation (extreme wet months); and c) runs variability measures, e.g., frequency of negative and positive runs of annual precipitation amounts, total number of the negative and positive runs. Two gridded precipitation data sets produced from observations are used: the Maurer et al. (2002) and the Daly et al. (1994) Precipitation Regression on Independent Slopes Method (PRISM) data sets. The data consist of monthly grid-point precipitation averaged on a United States Geological Survey (USGS) hydrological sub-region scale. The statistical significance of the obtained model minus observed measure differences is assessed using a block bootstrapping approach. The analyses were performed on annual, seasonal and monthly scale. The results indicate that the interdecadal standard deviation is underestimated, in general, on all time scales by the downscaled model data. The differences are statistically significant at a 0.05 significance level for several Lower Colorado Basin sub-regions on annual and seasonal scale, and for several sub-regions located mostly in the Upper Colorado River Basin for the months of March, June, July and November. Although the models simulate drier extreme wet years, wetter extreme dry years and drier extreme wet months for the Upper Colorado basin, the differences are mostly not-significant. Exceptions are the results about the extreme wet years for n=3 for sub-region White-Yampa, for n=6, 7, and 8 for sub-region Upper Colorado-Dolores, and about the extreme dry years for n=11 for sub-region Great Divide-Upper Green. None of the results for the sub-regions in the Lower Colorado Basin were significant. For most of the Upper Colorado sub-regions the models simulate significantly lower frequency of negative and positive 4-6 year runs, while for several sub-regions a significantly higher frequency of 2-year negative runs is evident in the model versus the Maurer data comparisons. The model projections versus the PRISM data comparison reveals similar results for the negative runs, while for the positive runs the results indicate that the models simulate higher frequency of the 2-6 year runs. The results for the Lower Colorado basin sub-regions are similar, in general, to these for the Upper Colorado sub-regions. The differences between the simulated and the observed total number of negative or positive runs were not significant for almost all of the sub-regions within the Colorado River Basin.
The development of an advanced vertical discretisation scheme for a regional ocean model
NASA Astrophysics Data System (ADS)
Bruciaferri, Diego; Shapiro, Georgy; Wobus, Fred
2017-04-01
When designing an ocean model, the choice of the vertical coordinate system must be pursued very carefully (Griffies et al., 2000); especially in those regional areas where local multi-scale processes interact with large-scale oceanographic features. Three main vertical coordinates are usually used in ocean modelling, namely the geopotential, terrain-following and isopycnic, but each one presents its own limitations and strengths. In the last decades, much research has been spent to investigate and develop hybrid approaches able to combine the advantages of each vertical coordinate system but minimising their disadvantages. Here we propose the hybrid s-s-z vertical discretisation scheme, an advanced version of the approach used by Shapiro et al. (2013). In our new scheme, the vertical domain is divided into three zones: in the upper and middle zones use s-coordinates while the deeper zone uses z-levels. The s-s-z vertical grid is introduced into the NEMO (Nucleus for European Modelling of the Ocean) model code and we compare the model skill of our new vertical discretisation scheme with the NEMO vertical grid using z-levels with partial steps through a set of idealized numerical experiments for which analytical solutions or theoretical models exist. Modelling results demonstrate that the magnitude of spurious currents arising from the horizontal pressure gradient errors are of the same order (10 ^ -3 m/s ) both with z-partial steps or with s-s-z vertical grids for the conditions favourable for the geopotential grids ( horizontal initial density levels). For a number of more realistic conditions representing a general cyclonic circulation in the sea, the new discretisation scheme produces smaller spurious currents and hence is more accurate than the z-level approach. Moreover, the enhanced capability of the s-s-z scheme to reproduce dense water cascades as compared to the z-partial steps grid is shown. Finally, we show how the new s-s-z grid can be useful to improve lateral sub-grid-physics parametrisation in ocean model with s-levels. References: Griffies, S. M., Boning, C., Bryan, F. O., Chassignet, E. P., Gerdes, R., Hasumi, H., Hirst, A., Treguier, A.-M., and Webb, D., 2000. Developments in Ocean Climate Modelling, Ocean Modelling, 2, 123-192. Shapiro, G., Luneva, M., Pickering, J., and Storkey, D.: The effect of various vertical discretisation schemes and horizontal diffusion parameterisation on the performance of a 3-D ocean model: the Black Sea case study, Ocean Sci., 9, 377-390, doi:10.5194/os-9-377-2013, 2013.
Simulation of Extreme Arctic Cyclones in IPCC AR5 Experiments
2014-05-15
atmospheric fields, including sea level pressure ( SLP ), on daily and sub-daily time scales at 2° horizontal resolution. A higher-resolution and more...its 21st-century simulation. Extreme cyclones were defined as occurrences of daily mean SLP at least 40 hPa below the climatological annual-average... SLP at a grid point. As such, no cyclone-tracking algorithm was employed, because the purpose here is to identify instances of extremely strong
NASA Technical Reports Server (NTRS)
Seth, Anji; Giorgi, Filippo; Dickinson, Robert E.
1994-01-01
A vectorized version of the biosphere-atmosphere transfer scheme (VBATS) is used to study moisture, energy, and momentum fluxes from heterogeneous land surfaces st the scale of an atmospheric model (AM) grid cells. To incorporate subgrid scale inhomogeneity, VBATS includes two important features: (1) characterization of the land surface (vegetation and soil parameters) at N subgrid points within an AM grid cell and (2) explicit distribution of climate forcing (precipitation, clouds, etc.) over the subgrid. In this study, VBATS is used in stand-alone mode to simulate a single AM grid cell and to evaluate the effects of subgrid scale vegetation and climate specification on the surface fluxes and hydrology. It is found that the partitioning of energy can be affected by up to 30%, runoff by 50%, and surface stress in excess of 60%. Distributing climate forcing over the AM grid cell increases the Bowen ratio, as a result of enhanced sensible heat flux and reduced latent heat flux. The combined effect of heterogeneous vegetation and distribution of climate is found to be dependent on the dominat vegetation class in the AM grid cell. Development of this method is part of a larger program to explore the importance of subgrid scale processes in regional and global climate simulations.
NASA Astrophysics Data System (ADS)
Fernández, Alfonso; Najafi, Mohammad Reza; Durand, Michael; Mark, Bryan G.; Moritz, Mark; Jung, Hahn Chul; Neal, Jeffrey; Shastry, Apoorva; Laborde, Sarah; Phang, Sui Chian; Hamilton, Ian M.; Xiao, Ningchuan
2016-08-01
Recent innovations in hydraulic modeling have enabled global simulation of rivers, including simulation of their coupled wetlands and floodplains. Accurate simulations of floodplains using these approaches may imply tremendous advances in global hydrologic studies and in biogeochemical cycling. One such innovation is to explicitly treat sub-grid channels within two-dimensional models, given only remotely sensed data in areas with limited data availability. However, predicting inundated area in floodplains using a sub-grid model has not been rigorously validated. In this study, we applied the LISFLOOD-FP hydraulic model using a sub-grid channel parameterization to simulate inundation dynamics on the Logone River floodplain, in northern Cameroon, from 2001 to 2007. Our goal was to determine whether floodplain dynamics could be simulated with sufficient accuracy to understand human and natural contributions to current and future inundation patterns. Model inputs in this data-sparse region include in situ river discharge, satellite-derived rainfall, and the shuttle radar topography mission (SRTM) floodplain elevation. We found that the model accurately simulated total floodplain inundation, with a Pearson correlation coefficient greater than 0.9, and RMSE less than 700 km2, compared to peak inundation greater than 6000 km2. Predicted discharge downstream of the floodplain matched measurements (Nash-Sutcliffe efficiency of 0.81), and indicated that net flow from the channel to the floodplain was modeled accurately. However, the spatial pattern of inundation was not well simulated, apparently due to uncertainties in SRTM elevations. We evaluated model results at 250, 500 and 1000-m spatial resolutions, and found that results are insensitive to spatial resolution. We also compared the model output against results from a run of LISFLOOD-FP in which the sub-grid channel parameterization was disabled, finding that the sub-grid parameterization simulated more realistic dynamics. These results suggest that analysis of global inundation is feasible using a sub-grid model, but that spatial patterns at sub-kilometer resolutions still need to be adequately predicted.
Global Swath and Gridded Data Tiling
NASA Technical Reports Server (NTRS)
Thompson, Charles K.
2012-01-01
This software generates cylindrically projected tiles of swath-based or gridded satellite data for the purpose of dynamically generating high-resolution global images covering various time periods, scaling ranges, and colors called "tiles." It reconstructs a global image given a set of tiles covering a particular time range, scaling values, and a color table. The program is configurable in terms of tile size, spatial resolution, format of input data, location of input data (local or distributed), number of processes run in parallel, and data conditioning.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Piri, Mohammad
2014-03-31
Under this project, a multidisciplinary team of researchers at the University of Wyoming combined state-of-the-art experimental studies, numerical pore- and reservoir-scale modeling, and high performance computing to investigate trapping mechanisms relevant to geologic storage of mixed scCO{sub 2} in deep saline aquifers. The research included investigations in three fundamental areas: (i) the experimental determination of two-phase flow relative permeability functions, relative permeability hysteresis, and residual trapping under reservoir conditions for mixed scCO{sub 2}-brine systems; (ii) improved understanding of permanent trapping mechanisms; (iii) scientifically correct, fine grid numerical simulations of CO{sub 2} storage in deep saline aquifers taking into account themore » underlying rock heterogeneity. The specific activities included: (1) Measurement of reservoir-conditions drainage and imbibition relative permeabilities, irreducible brine and residual mixed scCO{sub 2} saturations, and relative permeability scanning curves (hysteresis) in rock samples from RSU; (2) Characterization of wettability through measurements of contact angles and interfacial tensions under reservoir conditions; (3) Development of physically-based dynamic core-scale pore network model; (4) Development of new, improved high-performance modules for the UW-team simulator to provide new capabilities to the existing model to include hysteresis in the relative permeability functions, geomechanical deformation and an equilibrium calculation (Both pore- and core-scale models were rigorously validated against well-characterized core- flooding experiments); and (5) An analysis of long term permanent trapping of mixed scCO{sub 2} through high-resolution numerical experiments and analytical solutions. The analysis takes into account formation heterogeneity, capillary trapping, and relative permeability hysteresis.« less
Multiple Scales in Fluid Dynamics and Meteorology: The DFG Priority Programme 1276 MetStröm
NASA Astrophysics Data System (ADS)
von Larcher, Th; Klein, R.
2012-04-01
Geophysical fluid motions are characterized by a very wide range of length and time scales, and by a rich collection of varying physical phenomena. The mathematical description of these motions reflects this multitude of scales and mechanisms in that it involves strong non-linearities and various scale-dependent singular limit regimes. Considerable progress has been made in recent years in the mathematical modelling and numerical simulation of such flows in detailed process studies, numerical weather forecasting, and climate research. One task of outstanding importance in this context has been and will remain for the foreseeable future the subgrid scale parameterization of the net effects of non-resolved processes that take place on spacio-temporal scales not resolvable even by the largest most recent supercomputers. Since the advent of numerical weather forecasting some 60 years ago, one simple but efficient means to achieve improved forecasting skills has been increased spacio-temporal resolution. This seems quite consistent with the concept of convergence of numerical methods in Applied Mathematics and Computational Fluid Dynamics (CFD) at a first glance. Yet, the very notion of increased resolution in atmosphere-ocean science is very different from the one used in Applied Mathematics: For the mathematician, increased resolution provides the benefit of getting closer to the ideal of a converged solution of some given partial differential equations. On the other hand, the atmosphere-ocean scientist would naturally refine the computational grid and adjust his mathematical model, such that it better represents the relevant physical processes that occur at smaller scales. This conceptual contradiction remains largely irrelevant as long as geophysical flow models operate with fixed computational grids and time steps and with subgrid scale parameterizations being optimized accordingly. The picture changes fundamentally when modern techniques from CFD involving spacio-temporal grid adaptivity get invoked in order to further improve the net efficiency in exploiting the given computational resources. In the setting of geophysical flow simulation one must then employ subgrid scale parameterizations that dynamically adapt to the changing grid sizes and time steps, implement ways to judiciously control and steer the newly available flexibility of resolution, and invent novel ways of quantifying the remaining errors. The DFG priority program MetStröm covers the expertise of Meteorology, Fluid Dynamics, and Applied Mathematics to develop model- as well as grid-adaptive numerical simulation concepts in multidisciplinary projects. The goal of this priority programme is to provide simulation models which combine scale-dependent (mathematical) descriptions of key physical processes with adaptive flow discretization schemes. Deterministic continuous approaches and discrete and/or stochastic closures and their possible interplay are taken into consideration. Research focuses on the theory and methodology of multiscale meteorological-fluid mechanics modelling. Accompanying reference experiments support model validation.
NASA Astrophysics Data System (ADS)
Dumbser, Michael; Loubère, Raphaël
2016-08-01
In this paper we propose a simple, robust and accurate nonlinear a posteriori stabilization of the Discontinuous Galerkin (DG) finite element method for the solution of nonlinear hyperbolic PDE systems on unstructured triangular and tetrahedral meshes in two and three space dimensions. This novel a posteriori limiter, which has been recently proposed for the simple Cartesian grid case in [62], is able to resolve discontinuities at a sub-grid scale and is substantially extended here to general unstructured simplex meshes in 2D and 3D. It can be summarized as follows: At the beginning of each time step, an approximation of the local minimum and maximum of the discrete solution is computed for each cell, taking into account also the vertex neighbors of an element. Then, an unlimited discontinuous Galerkin scheme of approximation degree N is run for one time step to produce a so-called candidate solution. Subsequently, an a posteriori detection step checks the unlimited candidate solution at time t n + 1 for positivity, absence of floating point errors and whether the discrete solution has remained within or at least very close to the bounds given by the local minimum and maximum computed in the first step. Elements that do not satisfy all the previously mentioned detection criteria are flagged as troubled cells. For these troubled cells, the candidate solution is discarded as inappropriate and consequently needs to be recomputed. Within these troubled cells the old discrete solution at the previous time tn is scattered onto small sub-cells (Ns = 2 N + 1 sub-cells per element edge), in order to obtain a set of sub-cell averages at time tn. Then, a more robust second order TVD finite volume scheme is applied to update the sub-cell averages within the troubled DG cells from time tn to time t n + 1. The new sub-grid data at time t n + 1 are finally gathered back into a valid cell-centered DG polynomial of degree N by using a classical conservative and higher order accurate finite volume reconstruction technique. Consequently, if the number Ns is sufficiently large (Ns ≥ N + 1), the subscale resolution capability of the DG scheme is fully maintained, while preserving at the same time an essentially non-oscillatory behavior of the solution at discontinuities. Many standard DG limiters only adjust the discrete solution in troubled cells, based on the limiting of higher order moments or by applying a nonlinear WENO/HWENO reconstruction on the data at the new time t n + 1. Instead, our new DG limiter entirely recomputes the troubled cells by solving the governing PDE system again starting from valid data at the old time level tn, but using this time a more robust scheme on the sub-grid level. In other words, the piecewise polynomials produced by the new limiter are the result of a more robust solution of the PDE system itself, while most standard DG limiters are simply based on a mere nonlinear data post-processing of the discrete solution. Technically speaking, the new method corresponds to an element-wise checkpointing and restarting of the solver, using a lower order scheme on the sub-grid. As a result, the present DG limiter is even able to cure floating point errors like NaN values that have occurred after divisions by zero or after the computation of roots from negative numbers. This is a unique feature of our new algorithm among existing DG limiters. The new a posteriori sub-cell stabilization approach is developed within a high order accurate one-step ADER-DG framework on multidimensional unstructured meshes for hyperbolic systems of conservation laws as well as for hyperbolic PDE with non-conservative products. The method is applied to the Euler equations of compressible gas dynamics, to the ideal magneto-hydrodynamics equations (MHD) as well as to the seven-equation Baer-Nunziato model of compressible multi-phase flows. A large set of standard test problems is solved in order to assess the accuracy and robustness of the new limiter.
NASA Astrophysics Data System (ADS)
Barros, A. P.; Eghdami, M.
2017-12-01
High-resolution ( 1 km) numerical weather prediction models are capable of producing atmospheric spectra over synoptic and mesoscale ranges. Nogueira and Barros (2015) showed using high-resolution simulations in the Andes that the horizontal scale invariant behavior of atmospheric wind and water fields in the model is a process-dependent transient property that varies with the underlying dynamics. They found a sharp transition in the scaling parameters between non-convective and convective conditions. Spectral slopes around 2-2.3 arise under non-convective or very weak convective conditions, whereas in convective situations the transient scaling exponents remain under -5/3. Based on these results, Nogueira and Barros (2015) proposed a new sub-grid scale parameterization of clouds obtained from coarse resolution states alone. High Reynolds number direct numerical simulations of two-dimensional turbulence transfer shows that atmospheric flows involve concurrent direct (downscale) enstrophy transfer in the synoptic scales and inverse (upscale) kinetic energy transfer from the meso- to the synoptic-scales. In this study we use an analogy to investigate the transient behavior of kinetic energy spectra of winds over the Andes and Southern Appalachian Mountains representative of high and middle mountains, respectively. In the unstable conditions and particularly in the Planetary Boundary Layer (PBL) the spectral slopes approach -5/3 associated with the upscale KE turbulence transfer. However, in the stable conditions and above the planetary boundary layer, the spectra slopes approach steeper slopes about -3 associated with the downscale KE transfer. The underlying topography, surface roughness, diurnal heating and cooling and moist processes add to the complexity of the problem by introducing anisotropy and sources and sinks of energy. A comprehensive analysis and scaling of flow behavior conditional on stability regime for both KE and moist processes (total water, cloud water, rainfall) is necessary to elucidate scale-interactions among different processes.
Development of renormalization group analysis of turbulence
NASA Technical Reports Server (NTRS)
Smith, L. M.
1990-01-01
The renormalization group (RG) procedure for nonlinear, dissipative systems is now quite standard, and its applications to the problem of hydrodynamic turbulence are becoming well known. In summary, the RG method isolates self similar behavior and provides a systematic procedure to describe scale invariant dynamics in terms of large scale variables only. The parameterization of the small scales in a self consistent manner has important implications for sub-grid modeling. This paper develops the homogeneous, isotropic turbulence and addresses the meaning and consequence of epsilon-expansion. The theory is then extended to include a weak mean flow and application of the RG method to a sequence of models is shown to converge to the Navier-Stokes equations.
Photo-Carrier Multi-Dynamical Imaging at the Nanometer Scale in Organic and Inorganic Solar Cells.
Fernández Garrillo, Pablo A; Borowik, Łukasz; Caffy, Florent; Demadrille, Renaud; Grévin, Benjamin
2016-11-16
Investigating the photocarrier dynamics in nanostructured and heterogeneous energy materials is of crucial importance from both fundamental and technological points of view. Here, we demonstrate how noncontact atomic force microscopy combined with Kelvin probe force microscopy under frequency-modulated illumination can be used to simultaneously image the surface photopotential dynamics at different time scales with a sub-10 nm lateral resolution. The basic principle of the method consists in the acquisition of spectroscopic curves of the surface potential as a function of the illumination frequency modulation on a two-dimensional grid. We show how this frequency-spectroscopy can be used to probe simultaneously the charging rate and several decay processes involving short-lived and long-lived carriers. With this approach, dynamical images of the trap-filling, trap-delayed recombination and nongeminate recombination processes have been acquired in nanophase segregated organic donor-acceptor bulk heterojunction thin films. Furthermore, the spatial variation of the minority carrier lifetime has been imaged in polycrystalline silicon thin films. These results establish two-dimensional multidynamical photovoltage imaging as a universal tool for local investigations of the photocarrier dynamics in photoactive materials and devices.
Modeling RF-induced Plasma-Surface Interactions with VSim
NASA Astrophysics Data System (ADS)
Jenkins, Thomas G.; Smithe, David N.; Pankin, Alexei Y.; Roark, Christine M.; Stoltz, Peter H.; Zhou, Sean C.-D.; Kruger, Scott E.
2014-10-01
An overview of ongoing enhancements to the Plasma Discharge (PD) module of Tech-X's VSim software tool is presented. A sub-grid kinetic sheath model, developed for the accurate computation of sheath potentials near metal and dielectric-coated walls, enables the physical effects of DC and RF sheath dynamics to be included in macroscopic-scale plasma simulations that need not explicitly resolve sheath scale lengths. Sheath potential evolution, together with particle behavior near the sheath (e.g. sputtering), can thus be simulated in complex, experimentally relevant geometries. Simulations of RF sheath-enhanced impurity production near surfaces of the C-Mod field-aligned ICRF antenna are presented to illustrate the model; impurity mitigation techniques are also explored. Model extensions to capture the physics of secondary electron emission and of multispecies plasmas are summarized, together with a discussion of improved tools for plasma chemistry and IEDF/EEDF visualization and modeling. The latter tools are also highly relevant for commercial plasma processing applications. Ultimately, we aim to establish VSimPD as a robust, efficient computational tool for modeling fusion and industrial plasma processes. Supported by U.S. DoE SBIR Phase I/II Award DE-SC0009501.
Biography of a technology: North America's power grid through the twentieth century
NASA Astrophysics Data System (ADS)
Cohn, Julie A.
North Americans are among the world's most intense consumers of electricity. The vast majority in the United States and Canada access power from a network of transmission lines that stretch from the East Coast to the West Coast and from Canada to the Mexican Baja. This network, known as the largest interconnected machine in the world, evolved during the first two thirds of the twentieth century. With the very first link-ups occurring at the end of the 1890s, a wide variety of public and private utilities extended power lines to reach markets, access and manage energy resources, balance loads, realize economies of scale, provide backup power, and achieve economic stability. In 1967, utility managers and the Bureau of Reclamation connected the expansive eastern and western power pools to create the North American grid. Unlike other power grids around the world, built by single, centrally controlled entities, this large technological system emerged as the result of multiple decisions across eighty-five years of development, and negotiations for control at the economic, political, and technological levels. This dissertation describes the process of building the North American grid and the paradoxes the resulting system represents. While the grid functions as a single machine moving electricity across the continent, it is owned by many independent entities. Smooth operations suggest that the grid is a unified system; however, it operates under shared management and divided authority. In addition, although a single power network seems the logical outcome of electrification, in fact it was assembled through aggregation, not planning. Interconnections intentionally increase the robustness of individual sub-networks, yet the system itself is fragile, as demonstrated by major cascading power outages. Finally, the transmission network facilitates increased use of energy resources and consumption of power, but at certain points in the past, it also served as a technology of conservation. While this project explores the history of how and why North America has a huge interconnected power system, it also offers insights into the challenges the grid poses for our energy future.
Optimize of shrink process with X-Y CD bias on hole pattern
NASA Astrophysics Data System (ADS)
Koike, Kyohei; Hara, Arisa; Natori, Sakurako; Yamauchi, Shohei; Yamato, Masatoshi; Oyama, Kenichi; Yaegashi, Hidetami
2017-03-01
Gridded design rules[1] is major process in configuring logic circuit used 193-immersion lithography. In the scaling of grid patterning, we can make 10nm order line and space pattern by using multiple patterning techniques such as self-aligned multiple patterning (SAMP) and litho-etch- litho-etch (LELE)[2][3][4] . On the other hand, Line cut process has some error parameters such as pattern defect, placement error, roughness and X-Y CD bias with the decreasing scale. We tried to cure hole pattern roughness to use additional process such as Line smoothing[5] . Each smoothing process showed different effect. As the result, CDx shrink amount is smaller than CDy without one additional process. In this paper, we will report the pattern controllability comparison of EUV and 193-immersion. And we will discuss optimum method about CD bias on hole pattern.
ISINGLASS Auroral Sounding Rocket Campaign Data Synthesis: Radar, Imagery, and In Situ Observations
NASA Astrophysics Data System (ADS)
Clayton, R.; Lynch, K. A.; Evans, T.; Hampton, D. L.; Burleigh, M.; Zettergren, M. D.; Varney, R. H.; Reimer, A.; Hysell, D. L.; Michell, R.; Samara, M.; Grubbs, G. A., II
2017-12-01
E-field and flow variations across auroral arc boundaries are typically sub-grid measurements for ground based sensors such as radars and imagers, even for quiet stable arcs. In situ measurements can provide small scale resolution, but only provide a snapshot at a localized time and place. Using ground based and in situ measurements of the ISINGLASS auroral sounding rocket campaign in conjunction, we use the in situ measurements to validate ground based synthesis of these small scale observations based on the classification of auroral arcs in Marklund(1984). With validation of this technique, sub-grid information can be gained from radar data using particular visible auroral features during times where only ground based measurements are present. The ISINGLASS campaign (Poker Flat Alaska, Winter 2017) included the nights of Feb 22 2017 and Mar 02 2017, which possessed multiple stable arc boundaries that can be used for synthesis, including the two events into which the ISINGLASS rockets were launched. On Mar 02 from 0700 to 0800 UT, two stable slowly southward-propagating auroral arcs persisted within the instrument field of view, and lasted for a period of >15min. The second of these events contains the 36.304 rocket trajectory, while both events have full ground support from camera imagery and radar. Data synthesis from these events is accomplished using Butler (2010), Vennell (2009), and manually selected auroral boundaries from ground based cameras. With determination of the auroral arc boundaries from ground based imagery, a prediction of the fields along the length of a long straight arc boundary can be made using the ground based radar data, even on a sub-radar-grid scale, using the Marklund arc boundary classification. We assume that fields everywhere along a long stable arc boundary should be the same. Given a long stable arc, measurements anywhere along the arc (i.e. from PFISR) can be replicated along the length of the boundary. This prediction can then be validated from the in situ measurements of the fields from the ISINGLASS campaign. Upon successful synthesis and validation of the ground based data for the times where in situ data are present, the same analysis will be applied to similar long straight stable arcs during the campaign window when ground support is present to further explore the data synthesis method.
The construction of power grid operation index system considering the risk of maintenance
NASA Astrophysics Data System (ADS)
Tang, Jihong; Wang, Canlin; Jiang, Xinfan; Ye, Jianhui; Pan, Feilai
2018-02-01
In recent years, large-scale blackout occurred at home and abroad caused widespread concern about the operation of the grid in the world, and the maintenance risk is an important indicator of grid safety. The barrier operation of the circuit breaker exists in the process of overhaul of the power grid. The operation of the different barrier is of great significance to the change of the power flow, thus affecting the safe operation of the system. Most of the grid operating status evaluation index system did not consider the risk of maintenance, to this end, this paper from the security, economy, quality and cleanliness of the four angles, build the power grid operation index system considering the risk of maintenance.
Load Balancing Strategies for Multi-Block Overset Grid Applications
NASA Technical Reports Server (NTRS)
Djomehri, M. Jahed; Biswas, Rupak; Lopez-Benitez, Noe; Biegel, Bryan (Technical Monitor)
2002-01-01
The multi-block overset grid method is a powerful technique for high-fidelity computational fluid dynamics (CFD) simulations about complex aerospace configurations. The solution process uses a grid system that discretizes the problem domain by using separately generated but overlapping structured grids that periodically update and exchange boundary information through interpolation. For efficient high performance computations of large-scale realistic applications using this methodology, the individual grids must be properly partitioned among the parallel processors. Overall performance, therefore, largely depends on the quality of load balancing. In this paper, we present three different load balancing strategies far overset grids and analyze their effects on the parallel efficiency of a Navier-Stokes CFD application running on an SGI Origin2000 machine.
The Center for Multiscale Plasma Dynamics, Final Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gombosi, Tamas I.
The University of Michigan participated in the joint UCLA/Maryland fusion science center focused on plasma physics problems for which the traditional separation of the dynamics into microscale and macroscale processes breaks down. These processes involve large scale flows and magnetic fields tightly coupled to the small scale, kinetic dynamics of turbulence, particle acceleration and energy cascade. The interaction between these vastly disparate scales controls the evolution of the system. The enormous range of temporal and spatial scales associated with these problems renders direct simulation intractable even in computations that use the largest existing parallel computers. Our efforts focused on twomore » main problems: the development of Hall MHD solvers on solution adaptive grids and the development of solution adaptive grids using generalized coordinates so that the proper geometry of inertial confinement can be taken into account and efficient refinement strategies can be obtained.« less
Process for Descaling and Decontaminating Metals
Baybarz, R. D.
1961-04-25
The oxide scale on the surface of stainless steels and similar metals is removed by contacting the metal under an inert atmosphere with a dilute H/sub 2/ SO/sub 4/ solution containing CrSO/sub 4/. The removed oxide scale is either dissolved or disintegrated into a slurry by the solution. Preferred reagent concentrations are 0.3 to 0.5 M CrSO/sub 4/ and 0.5 to 0.6 M H/sub 2/SO/sub 4/. The process is particularly applicable to decontamination of aqueous homogeneous nuclear reactor systems. (AEC)
Are Atmospheric Updrafts a Key to Unlocking Climate Forcing and Sensitivity?
Donner, Leo J.; O'Brien, Travis A.; Rieger, Daniel; ...
2016-06-08
Both climate forcing and climate sensitivity persist as stubborn uncertainties limiting the extent to which climate models can provide actionable scientific scenarios for climate change. A key, explicit control on cloud-aerosol interactions, the largest uncertainty in climate forcing, is the vertical velocity of cloud-scale updrafts. Model-based studies of climate sensitivity indicate that convective entrainment, which is closely related to updraft speeds, is an important control on climate sensitivity. Updraft vertical velocities also drive many physical processes essential to numerical weather prediction. Vertical velocities and their role in atmospheric physical processes have been given very limited attention in models for climatemore » and numerical weather prediction. The relevant physical scales range down to tens of meters and are thus frequently sub-grid and require parameterization. Many state-of-science convection parameterizations provide mass fluxes without specifying vertical velocities, and parameterizations which do provide vertical velocities have been subject to limited evaluation against what have until recently been scant observations. Atmospheric observations imply that the distribution of vertical velocities depends on the areas over which the vertical velocities are averaged. Distributions of vertical velocities in climate models may capture this behavior, but it has not been accounted for when parameterizing cloud and precipitation processes in current models. New observations of convective vertical velocities offer a potentially promising path toward developing process-level cloud models and parameterizations for climate and numerical weather prediction. Taking account of scale-dependence of resolved vertical velocities offers a path to matching cloud-scale physical processes and their driving dynamics more realistically, with a prospect of reduced uncertainty in both climate forcing and sensitivity.« less
NASA Astrophysics Data System (ADS)
Stevens, R. G.; Lonsdale, C. L.; Brock, C. A.; Reed, M. K.; Crawford, J. H.; Holloway, J. S.; Ryerson, T. B.; Huey, L. G.; Nowak, J. B.; Pierce, J. R.
2012-04-01
New-particle formation in the plumes of coal-fired power plants and other anthropogenic sulphur sources may be an important source of particles in the atmosphere. It remains unclear, however, how best to reproduce this formation in global and regional aerosol models with grid-box lengths that are 10s of kilometres and larger. The predictive power of these models is thus limited by the resultant uncertainties in aerosol size distributions. In this presentation, we focus on sub-grid sulphate aerosol processes within coal-fired power plant plumes: the sub-grid oxidation of SO2 with condensation of H2SO4 onto newly-formed and pre-existing particles. Based on the results of the System for Atmospheric Modelling (SAM), a Large-Eddy Simulation/Cloud-Resolving Model (LES/CRM) with online TwO Moment Aerosol Sectional (TOMAS) microphysics, we develop a computationally efficient, but physically based, parameterization that predicts the characteristics of aerosol formed within coal-fired power plant plumes based on parameters commonly available in global and regional-scale models. Given large-scale mean meteorological parameters, emissions from the power plant, mean background condensation sink, and the desired distance from the source, the parameterization will predict the fraction of the emitted SO2 that is oxidized to H2SO4, the fraction of that H2SO4 that forms new particles instead of condensing onto preexisting particles, the median diameter of the newly-formed particles, and the number of newly-formed particles per kilogram SO2 emitted. We perform a sensitivity analysis of these characteristics of the aerosol size distribution to the meteorological parameters, the condensation sink, and the emissions. In general, new-particle formation and growth is greatly reduced during polluted conditions due to the large preexisting aerosol surface area for H2SO4 condensation and particle coagulation. The new-particle formation and growth rates are also a strong function of the amount of sunlight and NOx since both control OH concentrations. Decreases in NOx emissions without simultaneous decreases in SO2 emissions increase new-particle formation and growth due to increased oxidation of SO2. The parameterization we describe here should allow for more accurate predictions of aerosol size distributions and a greater confidence in the effects of aerosols in climate and health studies.
A CPT for Improving Turbulence and Cloud Processes in the NCEP Global Models
NASA Astrophysics Data System (ADS)
Krueger, S. K.; Moorthi, S.; Randall, D. A.; Pincus, R.; Bogenschutz, P.; Belochitski, A.; Chikira, M.; Dazlich, D. A.; Swales, D. J.; Thakur, P. K.; Yang, F.; Cheng, A.
2016-12-01
Our Climate Process Team (CPT) is based on the premise that the NCEP (National Centers for Environmental Prediction) global models can be improved by installing an integrated, self-consistent description of turbulence, clouds, deep convection, and the interactions between clouds and radiative and microphysical processes. The goal of our CPT is to unify the representation of turbulence and subgrid-scale (SGS) cloud processes and to unify the representation of SGS deep convective precipitation and grid-scale precipitation as the horizontal resolution decreases. We aim to improve the representation of small-scale phenomena by implementing a PDF-based SGS turbulence and cloudiness scheme that replaces the boundary layer turbulence scheme, the shallow convection scheme, and the cloud fraction schemes in the GFS (Global Forecast System) and CFS (Climate Forecast System) global models. We intend to improve the treatment of deep convection by introducing a unified parameterization that scales continuously between the simulation of individual clouds when and where the grid spacing is sufficiently fine and the behavior of a conventional parameterization of deep convection when and where the grid spacing is coarse. We will endeavor to improve the representation of the interactions of clouds, radiation, and microphysics in the GFS/CFS by using the additional information provided by the PDF-based SGS cloud scheme. The team is evaluating the impacts of the model upgrades with metrics used by the NCEP short-range and seasonal forecast operations.
NASA Astrophysics Data System (ADS)
Buaria, D.; Yeung, P. K.
2017-12-01
A new parallel algorithm utilizing a partitioned global address space (PGAS) programming model to achieve high scalability is reported for particle tracking in direct numerical simulations of turbulent fluid flow. The work is motivated by the desire to obtain Lagrangian information necessary for the study of turbulent dispersion at the largest problem sizes feasible on current and next-generation multi-petaflop supercomputers. A large population of fluid particles is distributed among parallel processes dynamically, based on instantaneous particle positions such that all of the interpolation information needed for each particle is available either locally on its host process or neighboring processes holding adjacent sub-domains of the velocity field. With cubic splines as the preferred interpolation method, the new algorithm is designed to minimize the need for communication, by transferring between adjacent processes only those spline coefficients determined to be necessary for specific particles. This transfer is implemented very efficiently as a one-sided communication, using Co-Array Fortran (CAF) features which facilitate small data movements between different local partitions of a large global array. The cost of monitoring transfer of particle properties between adjacent processes for particles migrating across sub-domain boundaries is found to be small. Detailed benchmarks are obtained on the Cray petascale supercomputer Blue Waters at the University of Illinois, Urbana-Champaign. For operations on the particles in a 81923 simulation (0.55 trillion grid points) on 262,144 Cray XE6 cores, the new algorithm is found to be orders of magnitude faster relative to a prior algorithm in which each particle is tracked by the same parallel process at all times. This large speedup reduces the additional cost of tracking of order 300 million particles to just over 50% of the cost of computing the Eulerian velocity field at this scale. Improving support of PGAS models on major compilers suggests that this algorithm will be of wider applicability on most upcoming supercomputers.
NASA Astrophysics Data System (ADS)
Grunloh, Timothy P.
The objective of this dissertation is to develop a 3-D domain-overlapping coupling method that leverages the superior flow field resolution of the Computational Fluid Dynamics (CFD) code STAR-CCM+ and the fast execution of the System Thermal Hydraulic (STH) code TRACE to efficiently and accurately model thermal hydraulic transport properties in nuclear power plants under complex conditions of regulatory and economic importance. The primary contribution is the novel Stabilized Inertial Domain Overlapping (SIDO) coupling method, which allows for on-the-fly correction of TRACE solutions for local pressures and velocity profiles inside multi-dimensional regions based on the results of the CFD simulation. The method is found to outperform the more frequently-used domain decomposition coupling methods. An STH code such as TRACE is designed to simulate large, diverse component networks, requiring simplifications to the fluid flow equations for reasonable execution times. Empirical correlations are therefore required for many sub-grid processes. The coarse grids used by TRACE diminish sensitivity to small scale geometric details such as Reactor Pressure Vessel (RPV) internals. A CFD code such as STAR-CCM+ uses much finer computational meshes that are sensitive to the geometric details of reactor internals. In turbulent flows, it is infeasible to fully resolve the flow solution, but the correlations used to model turbulence are at a low level. The CFD code can therefore resolve smaller scale flow processes. The development of a 3-D coupling method was carried out with the intention of improving predictive capabilities of transport properties in the downcomer and lower plenum regions of an RPV in reactor safety calculations. These regions are responsible for the multi-dimensional mixing effects that determine the distribution at the core inlet of quantities with reactivity implications, such as fluid temperature and dissolved neutron absorber concentration.
NASA Technical Reports Server (NTRS)
Pawson, Steven; Ott, Lesley E.; Zhu, Zhengxin; Bowman, Kevin; Brix, Holger; Collatz, G. James; Dutkiewicz, Stephanie; Fisher, Joshua B.; Gregg, Watson W.; Hill, Chris;
2011-01-01
Forward GEOS-5 AGCM simulations of CO2, with transport constrained by analyzed meteorology for 2009-2010, are examined. The CO2 distributions are evaluated using AIRS upper tropospheric CO2 and ACOS-GOSAT total column CO2 observations. Different combinations of surface C02 fluxes are used to generate ensembles of runs that span some uncertainty in surface emissions and uptake. The fluxes are specified in GEOS-5 from different inventories (fossil and biofuel), different data-constrained estimates of land biological emissions, and different data-constrained ocean-biology estimates. One set of fluxes is based on the established "Transcom" database and others are constructed using contemporary satellite observations to constrain land and ocean process models. Likewise, different approximations to sub-grid transport are employed, to construct an ensemble of CO2 distributions related to transport variability. This work is part of NASA's "Carbon Monitoring System Flux Pilot Project,"
NASA Technical Reports Server (NTRS)
Wey, Thomas
2017-01-01
This paper summarizes the reacting results of simulating a bluff body stabilized flame experiment of Volvo Validation Rig using a releasable edition of the National Combustion Code (NCC). The turbulence models selected to investigate the configuration are the sub-grid scaled kinetic energy coupled large eddy simulation (K-LES) and the time-filtered Navier-Stokes (TFNS) simulation. The turbulence chemistry interaction used is linear eddy mixing (LEM).
Dynamically reconfigurable photovoltaic system
Okandan, Murat; Nielson, Gregory N.
2016-05-31
A PV system composed of sub-arrays, each having a group of PV cells that are electrically connected to each other. A power management circuit for each sub-array has a communications interface and serves to connect or disconnect the sub-array to a programmable power grid. The power grid has bus rows and bus columns. A bus management circuit is positioned at a respective junction of a bus column and a bus row and is programmable through its communication interface to connect or disconnect a power path in the grid. As a result, selected sub-arrays are connected by selected power paths to be in parallel so as to produce a low system voltage, and, alternately in series so as to produce a high system voltage that is greater than the low voltage by at least a factor of ten.
Dynamically reconfigurable photovoltaic system
Okandan, Murat; Nielson, Gregory N.
2016-12-27
A PV system composed of sub-arrays, each having a group of PV cells that are electrically connected to each other. A power management circuit for each sub-array has a communications interface and serves to connect or disconnect the sub-array to a programmable power grid. The power grid has bus rows and bus columns. A bus management circuit is positioned at a respective junction of a bus column and a bus row and is programmable through its communication interface to connect or disconnect a power path in the grid. As a result, selected sub-arrays are connected by selected power paths to be in parallel so as to produce a low system voltage, and, alternately in series so as to produce a high system voltage that is greater than the low voltage by at least a factor of ten.
NASA Astrophysics Data System (ADS)
Hoch, Jannis M.; Neal, Jeffrey C.; Baart, Fedor; van Beek, Rens; Winsemius, Hessel C.; Bates, Paul D.; Bierkens, Marc F. P.
2017-10-01
We here present GLOFRIM, a globally applicable computational framework for integrated hydrological-hydrodynamic modelling. GLOFRIM facilitates spatially explicit coupling of hydrodynamic and hydrologic models and caters for an ensemble of models to be coupled. It currently encompasses the global hydrological model PCR-GLOBWB as well as the hydrodynamic models Delft3D Flexible Mesh (DFM; solving the full shallow-water equations and allowing for spatially flexible meshing) and LISFLOOD-FP (LFP; solving the local inertia equations and running on regular grids). The main advantages of the framework are its open and free access, its global applicability, its versatility, and its extensibility with other hydrological or hydrodynamic models. Before applying GLOFRIM to an actual test case, we benchmarked both DFM and LFP for a synthetic test case. Results show that for sub-critical flow conditions, discharge response to the same input signal is near-identical for both models, which agrees with previous studies. We subsequently applied the framework to the Amazon River basin to not only test the framework thoroughly, but also to perform a first-ever benchmark of flexible and regular grids on a large-scale. Both DFM and LFP produce comparable results in terms of simulated discharge with LFP exhibiting slightly higher accuracy as expressed by a Kling-Gupta efficiency of 0.82 compared to 0.76 for DFM. However, benchmarking inundation extent between DFM and LFP over the entire study area, a critical success index of 0.46 was obtained, indicating that the models disagree as often as they agree. Differences between models in both simulated discharge and inundation extent are to a large extent attributable to the gridding techniques employed. In fact, the results show that both the numerical scheme of the inundation model and the gridding technique can contribute to deviations in simulated inundation extent as we control for model forcing and boundary conditions. This study shows that the presented computational framework is robust and widely applicable. GLOFRIM is designed as open access and easily extendable, and thus we hope that other large-scale hydrological and hydrodynamic models will be added. Eventually, more locally relevant processes would be captured and more robust model inter-comparison, benchmarking, and ensemble simulations of flood hazard on a large scale would be allowed for.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mitchell, J., E-mail: JMitchell16@slb.com; Chandrasekera, T. C.
2014-12-14
The nuclear magnetic resonance transverse relaxation time T{sub 2}, measured using the Carr-Purcell-Meiboom-Gill (CPMG) experiment, is a powerful method for obtaining unique information on liquids confined in porous media. Furthermore, T{sub 2} provides structural information on the porous material itself and has many applications in petrophysics, biophysics, and chemical engineering. Robust interpretation of T{sub 2} distributions demands appropriate processing of the measured data since T{sub 2} is influenced by diffusion through magnetic field inhomogeneities occurring at the pore scale, caused by the liquid/solid susceptibility contrast. Previously, we introduced a generic model for the diffusion exponent of the form −ant{sub e}{supmore » k} (where n is the number and t{sub e} the temporal separation of spin echoes, and a is a composite diffusion parameter) in order to distinguish the influence of relaxation and diffusion in CPMG data. Here, we improve the analysis by introducing an automatic search for the optimum power k that best describes the diffusion behavior. This automated method is more efficient than the manual trial-and-error grid search adopted previously, and avoids variability through subjective judgments of experimentalists. Although our method does not avoid the inherent assumption that the diffusion exponent depends on a single k value, we show through simulation and experiment that it is robust in measurements of heterogeneous systems that violate this assumption. In this way, we obtain quantitative T{sub 2} distributions from complicated porous structures and demonstrate the analysis with examples of ceramics used for filtration and catalysis, and limestone of relevance to the construction and petroleum industries.« less
A code inspection process for security reviews
DOE Office of Scientific and Technical Information (OSTI.GOV)
Garzoglio, Gabriele; /Fermilab
2009-05-01
In recent years, it has become more and more evident that software threat communities are taking an increasing interest in Grid infrastructures. To mitigate the security risk associated with the increased numbers of attacks, the Grid software development community needs to scale up effort to reduce software vulnerabilities. This can be achieved by introducing security review processes as a standard project management practice. The Grid Facilities Department of the Fermilab Computing Division has developed a code inspection process, tailored to reviewing security properties of software. The goal of the process is to identify technical risks associated with an application andmore » their impact. This is achieved by focusing on the business needs of the application (what it does and protects), on understanding threats and exploit communities (what an exploiter gains), and on uncovering potential vulnerabilities (what defects can be exploited). The desired outcome of the process is an improvement of the quality of the software artifact and an enhanced understanding of possible mitigation strategies for residual risks. This paper describes the inspection process and lessons learned on applying it to Grid middleware.« less
A code inspection process for security reviews
NASA Astrophysics Data System (ADS)
Garzoglio, Gabriele
2010-04-01
In recent years, it has become more and more evident that software threat communities are taking an increasing interest in Grid infrastructures. To mitigate the security risk associated with the increased numbers of attacks, the Grid software development community needs to scale up effort to reduce software vulnerabilities. This can be achieved by introducing security review processes as a standard project management practice. The Grid Facilities Department of the Fermilab Computing Division has developed a code inspection process, tailored to reviewing security properties of software. The goal of the process is to identify technical risks associated with an application and their impact. This is achieved by focusing on the business needs of the application (what it does and protects), on understanding threats and exploit communities (what an exploiter gains), and on uncovering potential vulnerabilities (what defects can be exploited). The desired outcome of the process is an improvement of the quality of the software artifact and an enhanced understanding of possible mitigation strategies for residual risks. This paper describes the inspection process and lessons learned on applying it to Grid middleware.
QX MAN: Q and X file manipulation
NASA Technical Reports Server (NTRS)
Krein, Mark A.
1992-01-01
QX MAN is a grid and solution file manipulation program written primarily for the PARC code and the GRIDGEN family of grid generation codes. QX MAN combines many of the features frequently encountered in grid generation, grid refinement, the setting-up of initial conditions, and post processing. QX MAN allows the user to manipulate single block and multi-block grids (and their accompanying solution files) by splitting, concatenating, rotating, translating, re-scaling, and stripping or adding points. In addition, QX MAN can be used to generate an initial solution file for the PARC code. The code was written to provide several formats for input and output in order for it to be useful in a broad spectrum of applications.
Influence of topographic heterogeneity on the abandance of larch forest in eastern Siberia
NASA Astrophysics Data System (ADS)
Sato, H.; Kobayashi, H.
2016-12-01
In eastern Siberia, larches (Larix spp.) often exist in pure stands, constructing the world's largest coniferous forest, of which changes can significantly affect the earth's albedo and the global carbon balance. We have conducted simulation studies for this vegetation, aiming to forecast its structures and functions under changing climate (1, 2). In previous studies of simulating vegetation at large geographical scales, the examining area is divided into coarse grid cells such as 0.5 * 0.5 degree resolution, and topographical heterogeneities within each grid cell are just ignored. However, in Siberian larch area, which is located on the environmental edge of existence of forest ecosystem, abundance of larch trees largely depends on topographic condition at the scale of tens to hundreds meters. We, therefore, analyzed patterns of within-grid-scale heterogeneity of larch LAI as a function of topographic condition, and examined its underlying reason. For this analysis, larch LAI was estimated for each 1/112 degree from the SPOT-VEGETATION data, and topographic properties such as angularity and aspect direction were estimated form the ASTER-GDEM data. Through this analysis, we found that, for example, sign of correlation between angularity and larch LAI depends on hydrological condition on the grid cell. We then refined the hydrological sub-model of our vegetation model SEIB-DGVM, and validated whether the modified model can reconstruct these patterns, and examined its impact on the estimation of biomass and vegetation productivity of entire larch region. -- References --1. Sato, H., et al. (2010). "Simulation study of the vegetation structure and function in eastern Siberian larch forests using the individual-based vegetation model SEIB-DGVM." Forest Ecology and Management 259(3): 301-311.2. Sato, H., et al. (2016). "Endurance of larch forest ecosystems in eastern Siberia under warming trends." Ecology and Evolution
OpenMP parallelization of a gridded SWAT (SWATG)
NASA Astrophysics Data System (ADS)
Zhang, Ying; Hou, Jinliang; Cao, Yongpan; Gu, Juan; Huang, Chunlin
2017-12-01
Large-scale, long-term and high spatial resolution simulation is a common issue in environmental modeling. A Gridded Hydrologic Response Unit (HRU)-based Soil and Water Assessment Tool (SWATG) that integrates grid modeling scheme with different spatial representations also presents such problems. The time-consuming problem affects applications of very high resolution large-scale watershed modeling. The OpenMP (Open Multi-Processing) parallel application interface is integrated with SWATG (called SWATGP) to accelerate grid modeling based on the HRU level. Such parallel implementation takes better advantage of the computational power of a shared memory computer system. We conducted two experiments at multiple temporal and spatial scales of hydrological modeling using SWATG and SWATGP on a high-end server. At 500-m resolution, SWATGP was found to be up to nine times faster than SWATG in modeling over a roughly 2000 km2 watershed with 1 CPU and a 15 thread configuration. The study results demonstrate that parallel models save considerable time relative to traditional sequential simulation runs. Parallel computations of environmental models are beneficial for model applications, especially at large spatial and temporal scales and at high resolutions. The proposed SWATGP model is thus a promising tool for large-scale and high-resolution water resources research and management in addition to offering data fusion and model coupling ability.
A High Order Finite Difference Scheme with Sharp Shock Resolution for the Euler Equations
NASA Technical Reports Server (NTRS)
Gerritsen, Margot; Olsson, Pelle
1996-01-01
We derive a high-order finite difference scheme for the Euler equations that satisfies a semi-discrete energy estimate, and present an efficient strategy for the treatment of discontinuities that leads to sharp shock resolution. The formulation of the semi-discrete energy estimate is based on a symmetrization of the Euler equations that preserves the homogeneity of the flux vector, a canonical splitting of the flux derivative vector, and the use of difference operators that satisfy a discrete analogue to the integration by parts procedure used in the continuous energy estimate. Around discontinuities or sharp gradients, refined grids are created on which the discrete equations are solved after adding a newly constructed artificial viscosity. The positioning of the sub-grids and computation of the viscosity are aided by a detection algorithm which is based on a multi-scale wavelet analysis of the pressure grid function. The wavelet theory provides easy to implement mathematical criteria to detect discontinuities, sharp gradients and spurious oscillations quickly and efficiently.
IGMS: An Integrated ISO-to-Appliance Scale Grid Modeling System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Palmintier, Bryan; Hale, Elaine; Hansen, Timothy M.
This paper describes the Integrated Grid Modeling System (IGMS), a novel electric power system modeling platform for integrated transmission-distribution analysis that co-simulates off-the-shelf tools on high performance computing (HPC) platforms to offer unprecedented resolution from ISO markets down to appliances and other end uses. Specifically, the system simultaneously models hundreds or thousands of distribution systems in co-simulation with detailed Independent System Operator (ISO) markets and AGC-level reserve deployment. IGMS uses a new MPI-based hierarchical co-simulation framework to connect existing sub-domain models. Our initial efforts integrate opensource tools for wholesale markets (FESTIV), bulk AC power flow (MATPOWER), and full-featured distribution systemsmore » including physics-based end-use and distributed generation models (many instances of GridLAB-D[TM]). The modular IGMS framework enables tool substitution and additions for multi-domain analyses. This paper describes the IGMS tool, characterizes its performance, and demonstrates the impacts of the coupled simulations for analyzing high-penetration solar PV and price responsive load scenarios.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
VAN HEYST,B.J.
1999-10-01
Sulfur and nitrogen oxides emitted to the atmosphere have been linked to the acidification of water bodies and soils and perturbations in the earth's radiation balance. In order to model the global transport and transformation of SO{sub x} and NO{sub x}, detailed spatial and temporal emission inventories are required. Benkovitz et al. (1996) published the development of an inventory of 1985 global emissions of SO{sub x} and NO{sub x} from anthropogenic sources. The inventory was gridded to a 1{degree} x 1{degree} latitude-longitude grid and has served as input to several global modeling studies. There is now a need to providemore » modelers with an update of this inventory to a more recent year, with a split of the emissions into elevated and low level sources. This paper describes the development of a 1990 update of the SO{sub x} and NO{sub x} global inventories that also includes a breakdown of sources into 17 sector groups. The inventory development starts with a gridded global default EDGAR inventory (Olivier et al, 1996). In countries where more detailed national inventories are available, these are used to replace the emissions for those countries in the global default. The gridded emissions are distributed into two height levels (0-100m and >100m) based on the final plume heights that are estimated to be typical for the various sectors considered. The sources of data as well as some of the methodologies employed to compile and develop the 1990 global inventory for SO{sub x} and NO{sub x} are discussed. The results reported should be considered to be interim since the work is still in progress and additional data sets are expected to become available.« less
NASA Astrophysics Data System (ADS)
Liguori, Sara; O'Loughlin, Fiachra; Souvignet, Maxime; Coxon, Gemma; Freer, Jim; Woods, Ross
2014-05-01
This research presents a newly developed observed sub-daily gridded precipitation product for England and Wales. Importantly our analysis specifically allows a quantification of rainfall errors from grid to the catchment scale, useful for hydrological model simulation and the evaluation of prediction uncertainties. Our methodology involves the disaggregation of the current one kilometre daily gridded precipitation records available for the United Kingdom[1]. The hourly product is created using information from: 1) 2000 tipping-bucket rain gauges; and 2) the United Kingdom Met-Office weather radar network. These two independent datasets provide rainfall estimates at temporal resolutions much smaller than the current daily gridded rainfall product; thus allowing the disaggregation of the daily rainfall records to an hourly timestep. Our analysis is conducted for the period 2004 to 2008, limited by the current availability of the datasets. We analyse the uncertainty components affecting the accuracy of this product. Specifically we explore how these uncertainties vary spatially, temporally and with climatic regimes. Preliminary results indicate scope for improvement of hydrological model performance by the utilisation of this new hourly gridded rainfall product. Such product will improve our ability to diagnose and identify structural errors in hydrological modelling by including the quantification of input errors. References [1] Keller V, Young AR, Morris D, Davies H (2006) Continuous Estimation of River Flows. Technical Report: Estimation of Precipitation Inputs. in Agency E (ed.). Environmental Agency.
Mixing in 3D Sparse Multi-Scale Grid Generated Turbulence
NASA Astrophysics Data System (ADS)
Usama, Syed; Kopec, Jacek; Tellez, Jackson; Kwiatkowski, Kamil; Redondo, Jose; Malik, Nadeem
2017-04-01
Flat 2D fractal grids are known to alter turbulence characteristics downstream of the grid as compared to the regular grids with the same blockage ratio and the same mass inflow rates [1]. This has excited interest in the turbulence community for possible exploitation for enhanced mixing and related applications. Recently, a new 3D multi-scale grid design has been proposed [2] such that each generation of length scale of turbulence grid elements is held in its own frame, the overall effect is a 3D co-planar arrangement of grid elements. This produces a 'sparse' grid system whereby each generation of grid elements produces a turbulent wake pattern that interacts with the other wake patterns downstream. A critical motivation here is that the effective blockage ratio in the 3D Sparse Grid Turbulence (3DSGT) design is significantly lower than in the flat 2D counterpart - typically the blockage ratio could be reduced from say 20% in 2D down to 4% in the 3DSGT. If this idea can be realized in practice, it could potentially greatly enhance the efficiency of turbulent mixing and transfer processes clearly having many possible applications. Work has begun on the 3DSGT experimentally using Surface Flow Image Velocimetry (SFIV) [3] at the European facility in the Max Planck Institute for Dynamics and Self-Organization located in Gottingen, Germany and also at the Technical University of Catalonia (UPC) in Spain, and numerically using Direct Numerical Simulation (DNS) at King Fahd University of Petroleum & Minerals (KFUPM) in Saudi Arabia and in University of Warsaw in Poland. DNS is the most useful method to compare the experimental results with, and we are studying different types of codes such as Imcompact3d, and OpenFoam. Many variables will eventually be investigated for optimal mixing conditions. For example, the number of scale generations, the spacing between frames, the size ratio of grid elements, inflow conditions, etc. We will report upon the first set of findings from the 3DSGT by the time of the conference. {Acknowledgements}: This work has been supported partly by the EuHIT grant, 'Turbulence Generated by Sparse 3D Multi-Scale Grid (M3SG)', 2017. {References} [1] S. Laizet, J. C. Vassilicos. DNS of Fractal-Generated Turbulence. Flow Turbulence Combust 87:673705, (2011). [2] N. A. Malik. Sparse 3D Multi-Scale Grid Turbulence Generator. USPTO Application no. 14/710,531, Patent Pending, (2015). [3] J. Tellez, M. Gomez, B. Russo, J.M. Redondo. Surface Flow Image Velocimetry (SFIV) for hydraulics applications. 18th Int. Symposium on the Application of Laser Imaging Techniques in Fluid Mechanics, Lisbon, Portugal (2016).
NASA Astrophysics Data System (ADS)
Baroni, Gabriele; Zink, Matthias; Kumar, Rohini; Samaniego, Luis; Attinger, Sabine
2017-04-01
The advances in computer science and the availability of new detailed data-sets have led to a growing number of distributed hydrological models applied to finer and finer grid resolutions for larger and larger catchment areas. It was argued, however, that this trend does not necessarily guarantee better understanding of the hydrological processes or it is even not necessary for specific modelling applications. In the present study, this topic is further discussed in relation to the soil spatial heterogeneity and its effect on simulated hydrological state and fluxes. To this end, three methods are developed and used for the characterization of the soil heterogeneity at different spatial scales. The methods are applied at the soil map of the upper Neckar catchment (Germany), as example. The different soil realizations are assessed regarding their impact on simulated state and fluxes using the distributed hydrological model mHM. The results are analysed by aggregating the model outputs at different spatial scales based on the Representative Elementary Scale concept (RES) proposed by Refsgaard et al. (2016). The analysis is further extended in the present study by aggregating the model output also at different temporal scales. The results show that small scale soil variabilities are not relevant when the integrated hydrological responses are considered e.g., simulated streamflow or average soil moisture over sub-catchments. On the contrary, these small scale soil variabilities strongly affect locally simulated states and fluxes i.e., soil moisture and evapotranspiration simulated at the grid resolution. A clear trade-off is also detected by aggregating the model output by spatial and temporal scales. Despite the scale at which the soil variabilities are (or are not) relevant is not universal, the RES concept provides a simple and effective framework to quantify the predictive capability of distributed models and to identify the need for further model improvements e.g., finer resolution input. For this reason, the integration in this analysis of all the relevant input factors (e.g., precipitation, vegetation, geology) could provide a strong support for the definition of the right scale for each specific model application. In this context, however, the main challenge for a proper model assessment will be the correct characterization of the spatio- temporal variability of each input factor. Refsgaard, J.C., Højberg, A.L., He, X., Hansen, A.L., Rasmussen, S.H., Stisen, S., 2016. Where are the limits of model predictive capabilities?: Representative Elementary Scale - RES. Hydrol. Process. doi:10.1002/hyp.11029
Optimal configurations of spatial scale for grid cell firing under noise and uncertainty
Towse, Benjamin W.; Barry, Caswell; Bush, Daniel; Burgess, Neil
2014-01-01
We examined the accuracy with which the location of an agent moving within an environment could be decoded from the simulated firing of systems of grid cells. Grid cells were modelled with Poisson spiking dynamics and organized into multiple ‘modules’ of cells, with firing patterns of similar spatial scale within modules and a wide range of spatial scales across modules. The number of grid cells per module, the spatial scaling factor between modules and the size of the environment were varied. Errors in decoded location can take two forms: small errors of precision and larger errors resulting from ambiguity in decoding periodic firing patterns. With enough cells per module (e.g. eight modules of 100 cells each) grid systems are highly robust to ambiguity errors, even over ranges much larger than the largest grid scale (e.g. over a 500 m range when the maximum grid scale is 264 cm). Results did not depend strongly on the precise organization of scales across modules (geometric, co-prime or random). However, independent spatial noise across modules, which would occur if modules receive independent spatial inputs and might increase with spatial uncertainty, dramatically degrades the performance of the grid system. This effect of spatial uncertainty can be mitigated by uniform expansion of grid scales. Thus, in the realistic regimes simulated here, the optimal overall scale for a grid system represents a trade-off between minimizing spatial uncertainty (requiring large scales) and maximizing precision (requiring small scales). Within this view, the temporary expansion of grid scales observed in novel environments may be an optimal response to increased spatial uncertainty induced by the unfamiliarity of the available spatial cues. PMID:24366144
gProcess and ESIP Platforms for Satellite Imagery Processing over the Grid
NASA Astrophysics Data System (ADS)
Bacu, Victor; Gorgan, Dorian; Rodila, Denisa; Pop, Florin; Neagu, Gabriel; Petcu, Dana
2010-05-01
The Environment oriented Satellite Data Processing Platform (ESIP) is developed through the SEE-GRID-SCI (SEE-GRID eInfrastructure for regional eScience) co-funded by the European Commission through FP7 [1]. The gProcess Platform [2] is a set of tools and services supporting the development and the execution over the Grid of the workflow based processing, and particularly the satelite imagery processing. The ESIP [3], [4] is build on top of the gProcess platform by adding a set of satellite image processing software modules and meteorological algorithms. The satellite images can reveal and supply important information on earth surface parameters, climate data, pollution level, weather conditions that can be used in different research areas. Generally, the processing algorithms of the satellite images can be decomposed in a set of modules that forms a graph representation of the processing workflow. Two types of workflows can be defined in the gProcess platform: abstract workflow (PDG - Process Description Graph), in which the user defines conceptually the algorithm, and instantiated workflow (iPDG - instantiated PDG), which is the mapping of the PDG pattern on particular satellite image and meteorological data [5]. The gProcess platform allows the definition of complex workflows by combining data resources, operators, services and sub-graphs. The gProcess platform is developed for the gLite middleware that is available in EGEE and SEE-GRID infrastructures [6]. gProcess exposes the specific functionality through web services [7]. The Editor Web Service retrieves information on available resources that are used to develop complex workflows (available operators, sub-graphs, services, supported resources, etc.). The Manager Web Service deals with resources management (uploading new resources such as workflows, operators, services, data, etc.) and in addition retrieves information on workflows. The Executor Web Service manages the execution of the instantiated workflows on the Grid infrastructure. In addition, this web service monitors the execution and generates statistical data that are important to evaluate performances and to optimize execution. The Viewer Web Service allows access to input and output data. To prove and to validate the utility of the gProcess and ESIP platforms there were developed the GreenView and GreenLand applications. The GreenView related functionality includes the refinement of some meteorological data such as temperature, and the calibration of the satellite images based on field measurements. The GreenLand application performs the classification of the satellite images by using a set of vegetation indices. The gProcess and ESIP platforms are used as well in GiSHEO project [8] to support the processing of Earth Observation data over the Grid in eGLE (GiSHEO eLearning Environment). Experiments of performance assessment were conducted and they have revealed that the workflow-based execution could improve the execution time of a satellite image processing algorithm [9]. It is not a reliable solution to execute all the workflow nodes on different machines. The execution of some nodes can be more time consuming and they will be performed in a longer time than other nodes. The total execution time will be affected because some nodes will slow down the execution. It is important to correctly balance the workflow nodes. Based on some optimization strategy the workflow nodes can be grouped horizontally, vertically or in a hybrid approach. In this way, those operators will be executed on one machine and also the data transfer between workflow nodes will be lower. The dynamic nature of the Grid infrastructure makes it more exposed to the occurrence of failures. These failures can occur at worker node, services availability, storage element, etc. Currently gProcess has support for some basic error prevention and error management solutions. In future, some more advanced error prevention and management solutions will be integrated in the gProcess platform. References [1] SEE-GRID-SCI Project, http://www.see-grid-sci.eu/ [2] Bacu V., Stefanut T., Rodila D., Gorgan D., Process Description Graph Composition by gProcess Platform. HiPerGRID - 3rd International Workshop on High Performance Grid Middleware, 28 May, Bucharest. Proceedings of CSCS-17 Conference, Vol.2., ISSN 2066-4451, pp. 423-430, (2009). [3] ESIP Platform, http://wiki.egee-see.org/index.php/JRA1_Commonalities [4] Gorgan D., Bacu V., Rodila D., Pop Fl., Petcu D., Experiments on ESIP - Environment oriented Satellite Data Processing Platform. SEE-GRID-SCI User Forum, 9-10 Dec 2009, Bogazici University, Istanbul, Turkey, ISBN: 978-975-403-510-0, pp. 157-166 (2009). [5] Radu, A., Bacu, V., Gorgan, D., Diagrammatic Description of Satellite Image Processing Workflow. Workshop on Grid Computing Applications Development (GridCAD) at the SYNASC Symposium, 28 September 2007, Timisoara, IEEE Computer Press, ISBN 0-7695-3078-8, 2007, pp. 341-348 (2007). [6] Gorgan D., Bacu V., Stefanut T., Rodila D., Mihon D., Grid based Satellite Image Processing Platform for Earth Observation Applications Development. IDAACS'2009 - IEEE Fifth International Workshop on "Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications", 21-23 September, Cosenza, Italy, IEEE Published in Computer Press, 247-252 (2009). [7] Rodila D., Bacu V., Gorgan D., Integration of Satellite Image Operators as Workflows in the gProcess Application. Proceedings of ICCP2009 - IEEE 5th International Conference on Intelligent Computer Communication and Processing, 27-29 Aug, 2009 Cluj-Napoca. ISBN: 978-1-4244-5007-7, pp. 355-358 (2009). [8] GiSHEO consortium, Project site, http://gisheo.info.uvt.ro [9] Bacu V., Gorgan D., Graph Based Evaluation of Satellite Imagery Processing over Grid. ISPDC 2008 - 7th International Symposium on Parallel and Distributed Computing, July 1-5, 2008, Krakow, Poland. IEEE Computer Society 2008, ISBN: 978-0-7695-3472-5, pp. 147-154.
Pulsed laser-induced formation of silica nanogrids
2014-01-01
Silica grids with micron to sub-micron mesh sizes and wire diameters of 50 nm are fabricated on fused silica substrates. They are formed by single-pulse structured excimer laser irradiation of a UV-absorbing silicon suboxide (SiO x ) coating through the transparent substrate. A polydimethylsiloxane (PDMS) superstrate (cover layer) coated on top of the SiO x film prior to laser exposure serves as confinement for controlled laser-induced structure formation. At sufficiently high laser fluence, this process leads to grids consisting of a periodic loop network connected to the substrate at regular positions. By an additional high-temperature annealing, the residual SiO x is oxidized, and a pure SiO2 grid is obtained. PACS 81.07.-b; 81.07.Gf; 81.65.Cf PMID:24581305
Grid-Enabled High Energy Physics Research using a Beowulf Cluster
NASA Astrophysics Data System (ADS)
Mahmood, Akhtar
2005-04-01
At Edinboro University of Pennsylvania, we have built a 8-node 25 Gflops Beowulf Cluster with 2.5 TB of disk storage space to carry out grid-enabled, data-intensive high energy physics research for the ATLAS experiment via Grid3. We will describe how we built and configured our Cluster, which we have named the Sphinx Beowulf Cluster. We will describe the results of our cluster benchmark studies and the run-time plots of several parallel application codes. Once fully functional, the Cluster will be part of Grid3[www.ivdgl.org/grid3]. The current ATLAS simulation grid application, models the entire physical processes from the proton anti-proton collisions and detector's response to the collision debri through the complete reconstruction of the event from analyses of these responses. The end result is a detailed set of data that simulates the real physical collision event inside a particle detector. Grid is the new IT infrastructure for the 21^st century science -- a new computing paradigm that is poised to transform the practice of large-scale data-intensive research in science and engineering. The Grid will allow scientist worldwide to view and analyze huge amounts of data flowing from the large-scale experiments in High Energy Physics. The Grid is expected to bring together geographically and organizationally dispersed computational resources, such as CPUs, storage systems, communication systems, and data sources.
DICOMGrid: a middleware to integrate PACS and EELA-2 grid infrastructure
NASA Astrophysics Data System (ADS)
Moreno, Ramon A.; de Sá Rebelo, Marina; Gutierrez, Marco A.
2010-03-01
Medical images provide lots of information for physicians, but the huge amount of data produced by medical image equipments in a modern Health Institution is not completely explored in its full potential yet. Nowadays medical images are used in hospitals mostly as part of routine activities while its intrinsic value for research is underestimated. Medical images can be used for the development of new visualization techniques, new algorithms for patient care and new image processing techniques. These research areas usually require the use of huge volumes of data to obtain significant results, along with enormous computing capabilities. Such qualities are characteristics of grid computing systems such as EELA-2 infrastructure. The grid technologies allow the sharing of data in large scale in a safe and integrated environment and offer high computing capabilities. In this paper we describe the DicomGrid to store and retrieve medical images, properly anonymized, that can be used by researchers to test new processing techniques, using the computational power offered by grid technology. A prototype of the DicomGrid is under evaluation and permits the submission of jobs into the EELA-2 grid infrastructure while offering a simple interface that requires minimal understanding of the grid operation.
The Computing and Data Grid Approach: Infrastructure for Distributed Science Applications
NASA Technical Reports Server (NTRS)
Johnston, William E.
2002-01-01
With the advent of Grids - infrastructure for using and managing widely distributed computing and data resources in the science environment - there is now an opportunity to provide a standard, large-scale, computing, data, instrument, and collaboration environment for science that spans many different projects and provides the required infrastructure and services in a relatively uniform and supportable way. Grid technology has evolved over the past several years to provide the services and infrastructure needed for building 'virtual' systems and organizations. We argue that Grid technology provides an excellent basis for the creation of the integrated environments that can combine the resources needed to support the large- scale science projects located at multiple laboratories and universities. We present some science case studies that indicate that a paradigm shift in the process of science will come about as a result of Grids providing transparent and secure access to advanced and integrated information and technologies infrastructure: powerful computing systems, large-scale data archives, scientific instruments, and collaboration tools. These changes will be in the form of services that can be integrated with the user's work environment, and that enable uniform and highly capable access to these computers, data, and instruments, regardless of the location or exact nature of these resources. These services will integrate transient-use resources like computing systems, scientific instruments, and data caches (e.g., as they are needed to perform a simulation or analyze data from a single experiment); persistent-use resources. such as databases, data catalogues, and archives, and; collaborators, whose involvement will continue for the lifetime of a project or longer. While we largely address large-scale science in this paper, Grids, particularly when combined with Web Services, will address a broad spectrum of science scenarios. both large and small scale.
Capturing remote mixing due to internal tides using multi-scale modeling tool: SOMAR-LES
NASA Astrophysics Data System (ADS)
Santilli, Edward; Chalamalla, Vamsi; Scotti, Alberto; Sarkar, Sutanu
2016-11-01
Internal tides that are generated during the interaction of an oscillating barotropic tide with the bottom bathymetry dissipate only a fraction of their energy near the generation region. The rest is radiated away in the form of low- high-mode internal tides. These internal tides dissipate energy at remote locations when they interact with the upper ocean pycnocline, continental slope, and large scale eddies. Capturing the wide range of length and time scales involved during the life-cycle of internal tides is computationally very expensive. A recently developed multi-scale modeling tool called SOMAR-LES combines the adaptive grid refinement features of SOMAR with the turbulence modeling features of a Large Eddy Simulation (LES) to capture multi-scale processes at a reduced computational cost. Numerical simulations of internal tide generation at idealized bottom bathymetries are performed to demonstrate this multi-scale modeling technique. Although each of the remote mixing phenomena have been considered independently in previous studies, this work aims to capture remote mixing processes during the life cycle of an internal tide in more realistic settings, by allowing multi-level (coarse and fine) grids to co-exist and exchange information during the time stepping process.
Capel, P.D.; Zhang, H.
2000-01-01
In assessing the occurrence, behavior, and effects of agricultural chemicals in surface water, the scales of study (i.e., watershed, county, state, and regional areas) are usually much larger than the scale of agricultural fields, where much of the understanding of processes has been developed. Field-scale areas are characterized by relatively homogeneous conditions. The combination of process-based simulation models and geographic information system technology can be used to help extend our understanding of field processes to water-quality concerns at larger scales. To demonstrate this, the model "Groundwater Loading Effects of Agricultural Management Systems" was used to estimate the potential loss of two pesticides (atrazine and permethrin) in runoff to surface water in Fillmore County in southeastern Minnesota. The county was divided into field-scale areas on the basis of a 100 m by 100 m grid, and the influences of soil type and surface topography on the potential losses of the two pesticides in runoff was evaluated for each individual grid cell. The results could be used for guidance for agricultural management and regulatory decisions, for planning environmental monitoring programs, and as an educational tool for the public.
Multiscale Approach to Small River Plumes off California
NASA Astrophysics Data System (ADS)
Basdurak, N. B.; Largier, J. L.; Nidzieko, N.
2012-12-01
While larger scale plumes have received significant attention, the dynamics of plumes associated with small rivers typical of California are little studied. Since small streams are not dominated by a momentum flux, their plumes are more susceptible to conditions in the coastal ocean such as wind and waves. In order to correctly model water transport at smaller scales, there is a need to capture larger scale processes. To do this, one-way nested grids with varying grid resolution (1 km and 10 m for the parent and the child grid respectively) were constructed. CENCOOS (Central and Northern California Ocean Observing System) model results were used as boundary conditions to the parent grid. Semi-idealized model results for Santa Rosa Creek, California are presented from an implementation of the Regional Ocean Modeling System (ROMS v3.0), a three-dimensional, free-surface, terrain-following numerical model. In these preliminary results, the interaction between tides, winds, and buoyancy forcing in plume dynamics is explored for scenarios including different strengths of freshwater flow with different modes (steady and pulsed). Seasonal changes in transport dynamics and dispersion patterns are analyzed.
During running in place, grid cells integrate elapsed time and distance run
Kraus, Benjamin J.; Brandon, Mark P.; Robinson, Robert J.; Connerney, Michael A.; Hasselmo, Michael E.; Eichenbaum, Howard
2015-01-01
Summary The spatial scale of grid cells may be provided by self-generated motion information or by external sensory information from environmental cues. To determine whether grid cell activity reflects distance traveled or elapsed time independent of external information, we recorded grid cells as animals ran in place on a treadmill. Grid cell activity was only weakly influenced by location but most grid cells and other neurons recorded from the same electrodes strongly signaled a combination of distance and time, with some signaling only distance or time. Grid cells were more sharply tuned to time and distance than non-grid cells. Many grid cells exhibited multiple firing fields during treadmill running, parallel to the periodic firing fields observed in open fields, suggesting a common mode of information processing. These observations indicate that, in the absence of external dynamic cues, grid cells integrate self-generated distance and time information to encode a representation of experience. PMID:26539893
NASA Technical Reports Server (NTRS)
Haag, Thomas W. (Technical Monitor); Shivakumar, Kunigal N.
2003-01-01
Electric ion thrusters are the preferred engines for deep space missions, because of very high specific impulse. The ion engine consists of screen and accelerator grids containing thousands of concentric very small holes. The xenon gas accelerates between the two grids, thus developing the impulse force. The dominant life-limiting mechanism in the state-of-the-art molybdenum thrusters is the xenon ion sputter erosion of the accelerator grid. Carbon/carbon composites (CCC) have shown to be have less than 1/7 the erosion rates than the molybdenum, thus for interplanetary missions CCC engines are inevitable. Early effort to develop CCC composite thrusters had a limited success because of limitations of the drilling technology and the damage caused by drilling. The proposed is an in-situ manufacturing of holes while the CCC is made. Special low CTE molds will be used along with the NC A&T s patented resin transfer molding (RTM) technology to manufacture the CCC grids. First, a manufacture process for 10-cm diameter thruster grids will be developed and verified. Quality of holes, density, CTE, tension, flexure, transverse fatigue and sputter yield properties will be measured. After establishing the acceptable quality and properties, the process will be scaled to manufacture 30-cm diameter grids. The properties of the two grid sizes are compared with each other.
The impact of mesoscale convective systems on global precipitation: A modeling study
NASA Astrophysics Data System (ADS)
Tao, Wei-Kuo
2017-04-01
The importance of precipitating mesoscale convective systems (MCSs) has been quantified from TRMM precipitation radar and microwave imager retrievals. MCSs generate more than 50% of the rainfall in most tropical regions. Typical MCSs have horizontal scales of a few hundred kilometers (km); therefore, a large domain and high resolution are required for realistic simulations of MCSs in cloud-resolving models (CRMs). Almost all traditional global and climate models do not have adequate parameterizations to represent MCSs. Typical multi-scale modeling frameworks (MMFs) with 32 CRM grid points and 4 km grid spacing also might not have sufficient resolution and domain size for realistically simulating MCSs. In this study, the impact of MCSs on precipitation processes is examined by conducting numerical model simulations using the Goddard Cumulus Ensemble model (GCE) and Goddard MMF (GMMF). The results indicate that both models can realistically simulate MCSs with more grid points (i.e., 128 and 256) and higher resolutions (1 or 2 km) compared to those simulations with less grid points (i.e., 32 and 64) and low resolution (4 km). The modeling results also show that the strengths of the Hadley circulations, mean zonal and regional vertical velocities, surface evaporation, and amount of surface rainfall are either weaker or reduced in the GMMF when using more CRM grid points and higher CRM resolution. In addition, the results indicate that large-scale surface evaporation and wind feed back are key processes for determining the surface rainfall amount in the GMMF. A sensitivity test with reduced sea surface temperatures (SSTs) is conducted and results in both reduced surface rainfall and evaporation.
A multi-resolution approach to electromagnetic modeling.
NASA Astrophysics Data System (ADS)
Cherevatova, M.; Egbert, G. D.; Smirnov, M. Yu
2018-04-01
We present a multi-resolution approach for three-dimensional magnetotelluric forward modeling. Our approach is motivated by the fact that fine grid resolution is typically required at shallow levels to adequately represent near surface inhomogeneities, topography, and bathymetry, while a much coarser grid may be adequate at depth where the diffusively propagating electromagnetic fields are much smoother. This is especially true for forward modeling required in regularized inversion, where conductivity variations at depth are generally very smooth. With a conventional structured finite-difference grid the fine discretization required to adequately represent rapid variations near the surface are continued to all depths, resulting in higher computational costs. Increasing the computational efficiency of the forward modeling is especially important for solving regularized inversion problems. We implement a multi-resolution finite-difference scheme that allows us to decrease the horizontal grid resolution with depth, as is done with vertical discretization. In our implementation, the multi-resolution grid is represented as a vertical stack of sub-grids, with each sub-grid being a standard Cartesian tensor product staggered grid. Thus, our approach is similar to the octree discretization previously used for electromagnetic modeling, but simpler in that we allow refinement only with depth. The major difficulty arose in deriving the forward modeling operators on interfaces between adjacent sub-grids. We considered three ways of handling the interface layers and suggest a preferable one, which results in similar accuracy as the staggered grid solution, while retaining the symmetry of coefficient matrix. A comparison between multi-resolution and staggered solvers for various models show that multi-resolution approach improves on computational efficiency without compromising the accuracy of the solution.
On the Representation of Subgrid Microtopography Effects in Process-based Hydrologic Models
NASA Astrophysics Data System (ADS)
Jan, A.; Painter, S. L.; Coon, E. T.
2017-12-01
Increased availability of high-resolution digital elevation are enabling process-based hydrologic modeling on finer and finer scales. However, spatial variability in surface elevation (microtopography) exists below the scale of a typical hyper-resolution grid cell and has the potential to play a significant role in water retention, runoff, and surface/subsurface interactions. Though the concept of microtopographic features (depressions, obstructions) and the associated implications on flow and discharge are well established, representing those effects in watershed-scale integrated surface/subsurface hydrology models remains a challenge. Using the complex and coupled hydrologic environment of the Arctic polygonal tundra as an example, we study the effects of submeter topography and present a subgrid model parameterized by small-scale spatial heterogeneities for use in hyper-resolution models with polygons at a scale of 15-20 meters forming the surface cells. The subgrid model alters the flow and storage terms in the diffusion wave equation for surface flow. We compare our results against sub-meter scale simulations (acts as a benchmark for our simulations) and hyper-resolution models without the subgrid representation. The initiation of runoff in the fine-scale simulations is delayed and the recession curve is slowed relative to simulated runoff using the hyper-resolution model with no subgrid representation. Our subgrid modeling approach improves the representation of runoff and water retention relative to models that ignore subgrid topography. We evaluate different strategies for parameterizing subgrid model and present a classification-based method to efficiently move forward to larger landscapes. This work was supported by the Interoperable Design of Extreme-scale Application Software (IDEAS) project and the Next-Generation Ecosystem Experiments-Arctic (NGEE Arctic) project. NGEE-Arctic is supported by the Office of Biological and Environmental Research in the DOE Office of Science.
NASA Astrophysics Data System (ADS)
O'Brien, E.
2017-12-01
We have conducted an integration study on the origin and evolution of the tectonics and volcanism of seafloor in the Western Pacific Ocean that took place during the Cretaceous Normal Superchron (CNS) where sparse data has so far precluded detailed investigation. We have compiled the latest satellite-based gravity, gravity gradient, and magnetic grids (EMAG2 v.3) for this region. These crustal-scale high-resolution grids suggest that the CNS seafloor contains fossilized lithospheric morphology possibly attributed to the interaction between Cretaceous supervolcanism activity and Mid-Cretaceous Pacific mid ocean ridge systems that have continuously expanded the Pacific Plate. We recognize previously identified fossilized microplates west of the Magellan Rise, short-lived abandoned propagating rifts and fracture zones, all of which show significant rotation of seafloor fabric. In addition to these large scale observations, we have also compiled marine geological information from previously drilled cores and new data from a Kongsberg Topas PS18 Parametric Sub-Bottom Profiler collected on a transect from Honolulu, Hawaii to Apra, Guam acquired during research cruise SKQ2014S2. In particular, the narrow beam and high bandwidth signal of the Topas PS18 sub-bottom profiler provides sonar data of the seabed with a resolution and depth penetration that is unprecedented compared with previously available surveys in the region. A preliminary assessment of this high resolution Topas data allows us to better characterize sub-seafloor sediment properties and identify features, including the Upper Transparent Layer with identifiable pelagic clay and porcelanite-chert reflectors as well as tectonic features such as the westernmost tip of the Waghenaer Fracture Zone.
Upscaling of Hydraulic Conductivity using the Double Constraint Method
NASA Astrophysics Data System (ADS)
El-Rawy, Mustafa; Zijl, Wouter; Batelaan, Okke
2013-04-01
The mathematics and modeling of flow through porous media is playing an increasingly important role for the groundwater supply, subsurface contaminant remediation and petroleum reservoir engineering. In hydrogeology hydraulic conductivity data are often collected at a scale that is smaller than the grid block dimensions of a groundwater model (e.g. MODFLOW). For instance, hydraulic conductivities determined from the field using slug and packer tests are measured in the order of centimeters to meters, whereas numerical groundwater models require conductivities representative of tens to hundreds of meters of grid cell length. Therefore, there is a need for upscaling to decrease the number of grid blocks in a groundwater flow model. Moreover, models with relatively few grid blocks are simpler to apply, especially when the model has to run many times, as is the case when it is used to assimilate time-dependent data. Since the 1960s different methods have been used to transform a detailed description of the spatial variability of hydraulic conductivity to a coarser description. In this work we will investigate a relatively simple, but instructive approach: the Double Constraint Method (DCM) to identify the coarse-scale conductivities to decrease the number of grid blocks. Its main advantages are robustness and easy implementation, enabling to base computations on any standard flow code with some post processing added. The inversion step of the double constraint method is based on a first forward run with all known fluxes on the boundary and in the wells, followed by a second forward run based on the heads measured on the phreatic surface (i.e. measured in shallow observation wells) and in deeper observation wells. Upscaling, in turn is inverse modeling (DCM) to determine conductivities in coarse-scale grid blocks from conductivities in fine-scale grid blocks. In such a way that the head and flux boundary conditions applied to the fine-scale model are also honored at the coarse-scale. Exemplification will be presented for the Kleine Nete catchment, Belgium. As a result we identified coarse-scale conductivities while decreasing the number of grid blocks with the advantage that a model run costs less computation time and requires less memory space. In addition, ranking of models was investigated.
geoknife: Reproducible web-processing of large gridded datasets
Read, Jordan S.; Walker, Jordan I.; Appling, Alison P.; Blodgett, David L.; Read, Emily K.; Winslow, Luke A.
2016-01-01
Geoprocessing of large gridded data according to overlap with irregular landscape features is common to many large-scale ecological analyses. The geoknife R package was created to facilitate reproducible analyses of gridded datasets found on the U.S. Geological Survey Geo Data Portal web application or elsewhere, using a web-enabled workflow that eliminates the need to download and store large datasets that are reliably hosted on the Internet. The package provides access to several data subset and summarization algorithms that are available on remote web processing servers. Outputs from geoknife include spatial and temporal data subsets, spatially-averaged time series values filtered by user-specified areas of interest, and categorical coverage fractions for various land-use types.
Characterization of Cloud Water-Content Distribution
NASA Technical Reports Server (NTRS)
Lee, Seungwon
2010-01-01
The development of realistic cloud parameterizations for climate models requires accurate characterizations of subgrid distributions of thermodynamic variables. To this end, a software tool was developed to characterize cloud water-content distributions in climate-model sub-grid scales. This software characterizes distributions of cloud water content with respect to cloud phase, cloud type, precipitation occurrence, and geo-location using CloudSat radar measurements. It uses a statistical method called maximum likelihood estimation to estimate the probability density function of the cloud water content.
Self-modulating pressure gauge
Edwards, D. Jr.; Lanni, C.P.
1979-08-07
An ion gauge is disclosed having a reduced x-ray limit and means for measuring that limit. The gauge comprises an ion gauge of the Bayard-Alpert type having a short collector and having means for varying the grid-collector voltage. The x-ray limit (i.e. the collector current resulting from x-rays striking the collector) may then be determined by the formula: I/sub x/ = ..cap alpha..I/sub l/ - I/sub h//..cap alpha.. - l where: I/sub x/ = x-ray limit, I/sub l/ and I/sub h/ = the collector current at the lower and higher grid voltage respectively; and, ..cap alpha.. = the ratio of the collector current due to positive ions at the higher voltage to that at the lower voltage.
Using the High-Level Based Program Interface to Facilitate the Large Scale Scientific Computing
Shang, Yizi; Shang, Ling; Gao, Chuanchang; Lu, Guiming; Ye, Yuntao; Jia, Dongdong
2014-01-01
This paper is to make further research on facilitating the large-scale scientific computing on the grid and the desktop grid platform. The related issues include the programming method, the overhead of the high-level program interface based middleware, and the data anticipate migration. The block based Gauss Jordan algorithm as a real example of large-scale scientific computing is used to evaluate those issues presented above. The results show that the high-level based program interface makes the complex scientific applications on large-scale scientific platform easier, though a little overhead is unavoidable. Also, the data anticipation migration mechanism can improve the efficiency of the platform which needs to process big data based scientific applications. PMID:24574931
NASA Astrophysics Data System (ADS)
Adegoke, J. O.; Engelbrecht, F.; Vezhapparambu, S.
2013-12-01
In previous work demonstrated the application of a var¬iable-resolution global atmospheric model, the conformal-cubic atmospheric model (CCAM), across a wide range of spatial and time scales to investigate the ability of the model to provide realistic simulations of present-day climate and plausible projections of future climate change over sub-Saharan Africa. By applying the model in stretched-grid mode the versatility of the model dynamics, numerical formulation and physical parameterizations to function across a range of length scales over the region of interest, was also explored. We primarily used CCAM to illustrate the capability of the model to function as a flexible downscaling tool at the climate-change time scale. Here we report on additional long term climate projection studies performed by downscaling at much higher resolutions (8 Km) over an area that stretches from just south of Sahara desert to the southern coast of the Niger Delta and into the Gulf of Guinea. To perform these simulations, CCAM was provided with synoptic-scale forcing of atmospheric circulation from 2.5 deg resolution NCEP reanalysis at 6-hourly interval and SSTs from NCEP reanalysis data uses as lower boundary forcing. CCAM 60 Km resolution downscaled to 8 Km (Schmidt factor 24.75) then 8 Km resolution simulation downscaled to 1 Km (Schmidt factor 200) over an area approximately 50 Km x 50 Km in the southern Lake Chad Basin (LCB). Our intent in conducting these high resolution model runs was to obtain a deeper understanding of linkages between the projected future climate and the hydrological processes that control the surface water regime in this part of sub-Saharan Africa.
NASA Astrophysics Data System (ADS)
José Gómez-Navarro, Juan; María López-Romero, José; Palacios-Peña, Laura; Montávez, Juan Pedro; Jiménez-Guerrero, Pedro
2017-04-01
A critical challenge for assessing regional climate change projections relies on improving the estimate of atmospheric aerosol impact on clouds and reducing the uncertainty associated with the use of parameterizations. In this sense, the horizontal grid spacing implemented in state-of-the-art regional climate simulations is typically 10-25 kilometers, meaning that very important processes such as convective precipitation are smaller than a grid box, and therefore need to be parameterized. This causes large uncertainties, as closure assumptions and a number of parameters have to be established by model tuning. Convection is a physical process that may be strongly conditioned by atmospheric aerosols, although the solution of aerosol-cloud interactions in warm convective clouds remains nowadays a very important scientific challenge, rendering parametrization of these complex processes an important bottleneck that is responsible from a great part of the uncertainty in current climate change projections. Therefore, the explicit simulation of convective processes might improve the quality and reliability of the simulations of the aerosol-cloud interactions in a wide range of atmospheric phenomena. Particularly over the Mediterranean, the role of aerosol particles is very important, being this a crossroad that fuels the mixing of particles from different sources (sea-salt, biomass burning, anthropogenic, Saharan dust, etc). Still, the role of aerosols in extreme events in this area such as medicanes has been barely addressed. This work aims at assessing the role of aerosol-atmosphere interaction in medicanes with the help of the regional chemistry/climate on-line coupled model WRF-CHEM run at a convection-permitting resolution. The analysis is exemplary based on the "Rolf" medicane (6-8 November 2011). Using this case study as reference, four sets of simulations are run with two spatial resolutions: one at a convection-permitting configuration of 4 km, and other at the lower resolution of 12 km, in whose case the convection has to be parameterized. Each configuration is used to produce two simulations, including and not including aerosol-radiation-cloud interactions. The comparison of the simulated output at different scales allows to evaluate the impact of sub-grid scale mixing of precursors on aerosol production. By focusing on these processes at different resolutions, the differences between convection-permitting models running at resolutions of 4 km to 12 km can be explored. Preliminary results indicate that the inclusion of aerosol effects may indeed impact the severity of this simulated medicane, especially sea salt aerosols, and leads to important spatial shifts and differences in intensity of surface precipitation.
Synoptic scale wind field properties from the SEASAT SASS
NASA Technical Reports Server (NTRS)
Pierson, W. J., Jr.; Sylvester, W. B.; Salfi, R. E.
1984-01-01
Dealiased SEASAT SEASAT A Scatterometer System SASS vector winds obtained during the Gulf Of Alaska SEASAT Experiment GOASEX program are processed to obtain superobservations centered on a one degree by one degree grid. The grid. The results provide values for the combined effects of mesoscale variability and communication noise on the individual SASS winds. These superobservations winds are then processed further to obtain estimates of synoptic scale vector winds stress fields, the horizontal divergence of the wind, the curl of the wind stress and the vertical velocity at 200 m above the sea surface, each with appropriate standard deviations of the estimates for each grid point value. They also explain the concentration of water vapor, liquid water and precipitation found by means of the SMMR Scanning Multichannel Microwave Radiometer at fronts and occlusions in terms of strong warm, moist air advection in the warm air sector accompanied by convergence in the friction layer. Their quality is far superior to that of analyses based on conventional data, which are shown to yield many inconsistencies.
The power of structural modeling of sub-grid scales - application to astrophysical plasmas
NASA Astrophysics Data System (ADS)
Georgiev Vlaykov, Dimitar; Grete, Philipp
2015-08-01
In numerous astrophysical phenomena the dynamical range can span 10s of orders of magnitude. This implies more than billions of degrees-of-freedom and precludes direct numerical simulations from ever being a realistic possibility. A physical model is necessary to capture the unresolved physics occurring at the sub-grid scales (SGS).Structural modeling is a powerful concept which renders itself applicable to various physical systems. It stems from the idea of capturing the structure of the SGS terms in the evolution equations based on the scale-separation mechanism and independently of the underlying physics. It originates in the hydrodynamics field of large-eddy simulations. We apply it to the study of astrophysical MHD.Here, we present a non-linear SGS model for compressible MHD turbulence. The model is validated a priori at the tensorial, vectorial and scalar levels against of set of high-resolution simulations of stochastically forced homogeneous isotropic turbulence in a periodic box. The parameter space spans 2 decades in sonic Mach numbers (0.2 - 20) and approximately one decade in magnetic Mach number ~(1-8). This covers the super-Alfvenic sub-, trans-, and hyper-sonic regimes, with a range of plasma beta from 0.05 to 25. The Reynolds number is of the order of 103.At the tensor level, the model components correlate well with the turbulence ones, at the level of 0.8 and above. Vectorially, the alignment with the true SGS terms is encouraging with more than 50% of the model within 30° of the data. At the scalar level we look at the dynamics of the SGS energy and cross-helicity. The corresponding SGS flux terms have median correlations of ~0.8. Physically, the model represents well the two directions of the energy cascade.In comparison, traditional functional models exhibit poor local correlations with the data already at the scalar level. Vectorially, they are indifferent to the anisotropy of the SGS terms. They often struggle to represent the energy backscatter from small to large scales as well as the turbulent dynamo mechanism.Overall, the new model surpasses the traditional ones in all tests by a large margin.
Generating High Resolution Climate Scenarios Through Regional Climate Modelling Over Southern Africa
NASA Astrophysics Data System (ADS)
Ndhlovu, G. Z.; Woyessa, Y. E.; Vijayaraghavan, S.
2017-12-01
limate change has impacted the global environment and the Continent of Africa, especially Southern Africa, regarded as one of the most vulnerable regions in Africa, has not been spared from these impacts. Global Climate Models (GCMs) with coarse horizontal resolutions of 150-300 km do not provide sufficient details at the local basin scale due to mismatch between the size of river basins and the grid cell of the GCM. This makes it difficult to apply the outputs of GCMs directly to impact studies such as hydrological modelling. This necessitates the use of regional climate modelling at high resolutions that provide detailed information at regional and local scales to study both climate change and its impacts. To this end, an experiment was set up and conducted with PRECIS, a regional climate model, to generate climate scenarios at a high resolution of 25km for the local region in Zambezi River basin of Southern Africa. The major input data used included lateral and surface boundary conditions based on the GCMs. The data is processed, analysed and compared with CORDEX climate change project data generated for Africa. This paper, highlights the major differences of the climate scenarios generated by PRECIS Model and CORDEX Project for Africa and further gives recommendations for further research on generation of climate scenarios. The climatic variables such as precipitation and temperatures have been analysed for flood and droughts in the region. The paper also describes the setting up and running of an experiment using a high-resolution PRECIS model. In addition, a description has been made in running the model and generating the output variables on a sub basin scale. Regional climate modelling which provides information on climate change impact may lead to enhanced understanding of adaptive water resources management. Understanding the regional climate modelling results on sub basin scale is the first step in analysing complex hydrological processes and a basis for designing of adaptation and mitigation strategies in the region. Key words: Climate change, regional climate modelling, hydrological processes, extremes, scenarios [1] Corresponding author: Email:gndhlovu@cut.ac.za Tel:+27 (0) 51 507 3072
NASA Astrophysics Data System (ADS)
Talukdar, Karabi; Behera, Laxmidhar
2018-03-01
Imaging below the basalt for hydrocarbon exploration is a global problem because of poor penetration and significant loss of seismic energy due to scattering, attenuation, absorption and mode-conversion when the seismic waves encounter a highly heterogeneous and rugose basalt layer. The conventional (short offset) seismic data acquisition, processing and modeling techniques adopted by the oil industry generally fails to image hydrocarbon-bearing sub-trappean Mesozoic sediments hidden below the basalt and is considered as a serious problem for hydrocarbon exploration in the world. To overcome this difficulty of sub-basalt imaging, we have generated dense synthetic seismic data with the help of elastic finite-difference full-wave modeling using staggered-grid scheme for the model derived from ray-trace inversion using sparse wide-angle seismic data acquired along Sinor-Valod profile in the Deccan Volcanic Province of India. The full-wave synthetic seismic data generated have been processed and imaged using conventional seismic data processing technique with Kirchhoff pre-stack time and depth migrations. The seismic image obtained correlates with all the structural features of the model obtained through ray-trace inversion of wide-angle seismic data, validating the effectiveness of robust elastic finite-difference full-wave modeling approach for imaging below thick basalts. Using the full-wave modeling also allows us to decipher small-scale heterogeneities imposed in the model as a measure of the rugose basalt interfaces, which could not be dealt with ray-trace inversion. Furthermore, we were able to accurately image thin low-velocity hydrocarbon-bearing Mesozoic sediments sandwiched between and hidden below two thick sequences of high-velocity basalt layers lying above the basement.
Performance Enhancement Strategies for Multi-Block Overset Grid CFD Applications
NASA Technical Reports Server (NTRS)
Djomehri, M. Jahed; Biswas, Rupak
2003-01-01
The overset grid methodology has significantly reduced time-to-solution of highfidelity computational fluid dynamics (CFD) simulations about complex aerospace configurations. The solution process resolves the geometrical complexity of the problem domain by using separately generated but overlapping structured discretization grids that periodically exchange information through interpolation. However, high performance computations of such large-scale realistic applications must be handled efficiently on state-of-the-art parallel supercomputers. This paper analyzes the effects of various performance enhancement strategies on the parallel efficiency of an overset grid Navier-Stokes CFD application running on an SGI Origin2000 machinc. Specifically, the role of asynchronous communication, grid splitting, and grid grouping strategies are presented and discussed. Details of a sophisticated graph partitioning technique for grid grouping are also provided. Results indicate that performance depends critically on the level of latency hiding and the quality of load balancing across the processors.
ANFIS-based modelling for coagulant dosage in drinking water treatment plant: a case study.
Heddam, Salim; Bermad, Abdelmalek; Dechemi, Noureddine
2012-04-01
Coagulation is the most important stage in drinking water treatment processes for the maintenance of acceptable treated water quality and economic plant operation, which involves many complex physical and chemical phenomena. Moreover, coagulant dosing rate is non-linearly correlated to raw water characteristics such as turbidity, conductivity, pH, temperature, etc. As such, coagulation reaction is hard or even impossible to control satisfactorily by conventional methods. Traditionally, jar tests are used to determine the optimum coagulant dosage. However, this is expensive and time-consuming and does not enable responses to changes in raw water quality in real time. Modelling can be used to overcome these limitations. In this study, an Adaptive Neuro-Fuzzy Inference System (ANFIS) was used for modelling of coagulant dosage in drinking water treatment plant of Boudouaou, Algeria. Six on-line variables of raw water quality including turbidity, conductivity, temperature, dissolved oxygen, ultraviolet absorbance, and the pH of water, and alum dosage were used to build the coagulant dosage model. Two ANFIS-based Neuro-fuzzy systems are presented. The two Neuro-fuzzy systems are: (1) grid partition-based fuzzy inference system (FIS), named ANFIS-GRID, and (2) subtractive clustering based (FIS), named ANFIS-SUB. The low root mean square error and high correlation coefficient values were obtained with ANFIS-SUB method of a first-order Sugeno type inference. This study demonstrates that ANFIS-SUB outperforms ANFIS-GRID due to its simplicity in parameter selection and its fitness in the target problem.
Energy-Water-Land-Climate Nexus: Modeling Impacts from the Asset to Regional Scale
NASA Astrophysics Data System (ADS)
Tidwell, V. C.; Bennett, K. E.; Middleton, R. S.; Behery, S.; Macknick, J.; Corning-Padilla, A.; Brinkman, G.; Meng, M.
2016-12-01
A critical challenge for the energy-water-land nexus is understanding and modeling the connection between the natural system—including changes in climate, land use/cover, and streamflow—and the engineered system including water for energy, agriculture, and society. Equally important is understanding the linkage across scales; that is, how impacts at the asset level aggregate to influence behavior at the local to regional scale. Toward this need, a case study was conducted featuring multi-sector and multi-scale modeling centered on the San Juan River basin (a watershed that accounts for one-tenth of the Colorado River drainage area). Simulations were driven by statistically downscaled climate data from three global climate models (emission scenario RCP 8.5) and planned growth in regional water demand. The Variable Infiltration Capacity (VIC) hydrologic model was fitted with a custom vegetation mortality sub-model and used to estimate tributary inflows to the San Juan River and estimate reservoir evaporation. San Juan River operations, including releases from Navajo Reservoir, were subsequently modeled using RiverWare to estimate impacts on water deliveries out to the year 2100. Major water demands included two large coal-fired power plants, a local electric utility, river-side irrigation, the Navajo Indian Irrigation Project and instream flows managed for endangered aquatic species. Also tracked were basin exports, including water (downstream flows to the Colorado River and interbasin transfers to the Rio Grande) and interstate electric power transmission. Implications for the larger western electric grid were assessed using PLEXOS, a sub-hourly dispatch, electric production-cost model. Results highlight asset-level interactions at the energy-water-land nexus driven by climate and population dynamics; specifically, growing vulnerabilities to shorted water deliveries. Analyses also explored linkages across geographic scales from the San Juan to the larger Colorado River and Rio Grande basins as well as the western power grid.
NASA Astrophysics Data System (ADS)
Mukhopadhyay, P.; Phani Murali Krishna, R.; Goswami, Bidyut B.; Abhik, S.; Ganai, Malay; Mahakur, M.; Khairoutdinov, Marat; Dudhia, Jimmy
2016-05-01
Inspite of significant improvement in numerical model physics, resolution and numerics, the general circulation models (GCMs) find it difficult to simulate realistic seasonal and intraseasonal variabilities over global tropics and particularly over Indian summer monsoon (ISM) region. The bias is mainly attributed to the improper representation of physical processes. Among all the processes, the cloud and convective processes appear to play a major role in modulating model bias. In recent times, NCEP CFSv2 model is being adopted under Monsoon Mission for dynamical monsoon forecast over Indian region. The analyses of climate free run of CFSv2 in two resolutions namely at T126 and T382, show largely similar bias in simulating seasonal rainfall, in capturing the intraseasonal variability at different scales over the global tropics and also in capturing tropical waves. Thus, the biases of CFSv2 indicate a deficiency in model's parameterization of cloud and convective processes. Keeping this in background and also for the need to improve the model fidelity, two approaches have been adopted. Firstly, in the superparameterization, 32 cloud resolving models each with a horizontal resolution of 4 km are embedded in each GCM (CFSv2) grid and the conventional sub-grid scale convective parameterization is deactivated. This is done to demonstrate the role of resolving cloud processes which otherwise remain unresolved. The superparameterized CFSv2 (SP-CFS) is developed on a coarser version T62. The model is integrated for six and half years in climate free run mode being initialised from 16 May 2008. The analyses reveal that SP-CFS simulates a significantly improved mean state as compared to default CFS. The systematic bias of lesser rainfall over Indian land mass, colder troposphere has substantially been improved. Most importantly the convectively coupled equatorial waves and the eastward propagating MJO has been found to be simulated with more fidelity in SP-CFS. The reason of such betterment in model mean state has been found to be due to the systematic improvement in moisture field, temperature profile and moist instability. The model also has better simulated the cloud and rainfall relation. This initiative demonstrates the role of cloud processes on the mean state of coupled GCM. As the superparameterization approach is computationally expensive, so in another approach, the conventional Simplified Arakawa Schubert (SAS) scheme is replaced by a revised SAS scheme (RSAS) and also the old and simplified cloud scheme of Zhao-Karr (1997) has been replaced by WSM6 in CFSV2 (hereafter CFS-CR). The primary objective of such modifications is to improve the distribution of convective rain in the model by using RSAS and the grid-scale or the large scale nonconvective rain by WSM6. The WSM6 computes the tendency of six class (water vapour, cloud water, ice, snow, graupel, rain water) hydrometeors at each of the model grid and contributes in the low, middle and high cloud fraction. By incorporating WSM6, for the first time in a global climate model, we are able to show a reasonable simulation of cloud ice and cloud liquid water distribution vertically and spatially as compared to Cloudsat observations. The CFS-CR has also showed improvement in simulating annual rainfall cycle and intraseasonal variability over the ISM region. These improvements in CFS-CR are likely to be associated with improvement of the convective and stratiform rainfall distribution in the model. These initiatives clearly address a long standing issue of resolving the cloud processes in climate model and demonstrate that the improved cloud and convective process paramterizations can eventually reduce the systematic bias and improve the model fidelity.
NASA Astrophysics Data System (ADS)
Quinn, Niall; Freer, Jim; Coxon, Gemma; O'Loughlin, Fiachra; Woods, Ross; Liguori, Sara
2015-04-01
In Great Britain and many other regions of the world, flooding resulting from short duration, high intensity rainfall events can lead to significant economic losses and fatalities. At present, such extreme events are often poorly evaluated using hydrological models due, in part, to their rarity and relatively short duration and a lack of appropriate data. Such storm characteristics are not well represented by daily rainfall records currently available using volumetric gauges and/or derived gridded products. This research aims to address this important data gap by developing a sub-daily gridded precipitation product for Great Britain. Our focus is to better understand these storm events and some of the challenges and uncertainties in quantifying such data across catchment scales. Our goal is to both improve such rainfall characterisation and derive an input to drive hydrological model simulations. Our methodology involves the collation, error checking, and spatial interpolation of approximately 2000 rain gauges located across Great Britain, provided by the Scottish Environment Protection Agency (SEPA) and the Environment Agency (EA). Error checking was conducted over the entirety of the TBR data available, utilising a two stage approach. First, rain gauge data at each site were examined independently, with data exceeding reasonable thresholds marked as suspect. Second, potentially erroneous data were marked using a neighbourhood analysis approach whereby measurements at a given gauge were deemed suspect if they did not fall within defined bounds of measurements at neighbouring gauges. A total of eight error checks were conducted. To provide the user with the greatest flexibility possible, the error markers associated with each check have been recorded at every site. This approach aims to enable the user to choose which checks they deem most suitable for a particular application. The quality assured TBR dataset was then spatially interpolated to produce a national scale gridded rainfall product. Finally, radar rainfall data provided by the UK Met Office was assimilated, where available, to provide an optimal hourly estimate of rainfall, given the error variance associated with both datasets. This research introduces a sub-daily rainfall product that will be of particular value to hydrological modellers requiring rainfall inputs at higher temporal resolutions than those currently available nationally. Further research will aim to quantify the uncertainties in the rainfall product in order to improve our ability to diagnose and identify structural errors in hydrological modelling of extreme events. Here we present our initial findings.
NASA Astrophysics Data System (ADS)
Huang, Melin; Huang, Bormin; Huang, Allen H.
2014-10-01
The Weather Research and Forecasting (WRF) model provided operational services worldwide in many areas and has linked to our daily activity, in particular during severe weather events. The scheme of Yonsei University (YSU) is one of planetary boundary layer (PBL) models in WRF. The PBL is responsible for vertical sub-grid-scale fluxes due to eddy transports in the whole atmospheric column, determines the flux profiles within the well-mixed boundary layer and the stable layer, and thus provide atmospheric tendencies of temperature, moisture (including clouds), and horizontal momentum in the entire atmospheric column. The YSU scheme is very suitable for massively parallel computation as there are no interactions among horizontal grid points. To accelerate the computation process of the YSU scheme, we employ Intel Many Integrated Core (MIC) Architecture as it is a multiprocessor computer structure with merits of efficient parallelization and vectorization essentials. Our results show that the MIC-based optimization improved the performance of the first version of multi-threaded code on Xeon Phi 5110P by a factor of 2.4x. Furthermore, the same CPU-based optimizations improved the performance on Intel Xeon E5-2603 by a factor of 1.6x as compared to the first version of multi-threaded code.
Fabricating High-Resolution X-Ray Collimators
NASA Technical Reports Server (NTRS)
Appleby, Michael; Atkinson, James E.; Fraser, Iain; Klinger, Jill
2008-01-01
A process and method for fabricating multi-grid, high-resolution rotating modulation collimators for arcsecond and sub-arcsecond x-ray and gamma-ray imaging involves photochemical machining and precision stack lamination. The special fixturing and etching techniques that have been developed are used for the fabrication of multiple high-resolution grids on a single array substrate. This technology has application in solar and astrophysics and in a number of medical imaging applications including mammography, computed tomography (CT), single photon emission computed tomography (SPECT), and gamma cameras used in nuclear medicine. This collimator improvement can also be used in non-destructive testing, hydrodynamic weapons testing, and microbeam radiation therapy.
NASA Astrophysics Data System (ADS)
Ajami, H.; Sharma, A.
2016-12-01
A computationally efficient, semi-distributed hydrologic modeling framework is developed to simulate water balance at a catchment scale. The Soil Moisture and Runoff simulation Toolkit (SMART) is based upon the delineation of contiguous and topologically connected Hydrologic Response Units (HRUs). In SMART, HRUs are delineated using thresholds obtained from topographic and geomorphic analysis of a catchment, and simulation elements are distributed cross sections or equivalent cross sections (ECS) delineated in first order sub-basins. ECSs are formulated by aggregating topographic and physiographic properties of the part or entire first order sub-basins to further reduce computational time in SMART. Previous investigations using SMART have shown that temporal dynamics of soil moisture are well captured at a HRU level using the ECS delineation approach. However, spatial variability of soil moisture within a given HRU is ignored. Here, we examined a number of disaggregation schemes for soil moisture distribution in each HRU. The disaggregation schemes are either based on topographic based indices or a covariance matrix obtained from distributed soil moisture simulations. To assess the performance of the disaggregation schemes, soil moisture simulations from an integrated land surface-groundwater model, ParFlow.CLM in Baldry sub-catchment, Australia are used. ParFlow is a variably saturated sub-surface flow model that is coupled to the Common Land Model (CLM). Our results illustrate that the statistical disaggregation scheme performs better than the methods based on topographic data in approximating soil moisture distribution at a 60m scale. Moreover, the statistical disaggregation scheme maintains temporal correlation of simulated daily soil moisture while preserves the mean sub-basin soil moisture. Future work is focused on assessing the performance of this scheme in catchments with various topographic and climate settings.
NASA Astrophysics Data System (ADS)
Alapaty, Kiran; Bullock, O. Russell; Herwehe, Jerold; Spero, Tanya; Nolte, Christopher; Mallard, Megan
2014-05-01
The Regional Climate Modeling Team at the U.S. Environmental Protection Agency has been improving the quality of regional climate fields generated by the Weather Research and Forecasting (WRF) model. Active areas of research include improving core physics within the WRF model and adapting the physics for regional climate applications, improving the representation of inland lakes that are unresolved by the driving fields, evaluating nudging strategies, and devising techniques to demonstrate value added by dynamical downscaling. These research efforts have been conducted using reanalysis data as driving fields, and then their results have been applied to downscale data from global climate models. The goals of this work are to equip environmental managers and policy/decision makers in the U.S. with science, tools, and data to inform decisions related to adapting to and mitigating the potential impacts of climate change on air quality, ecosystems, and human health. Our presentation will focus mainly on one area of the Team's research: Development and testing of a seamless convection parameterization scheme. For the continental U.S., one of the impediments to high-resolution (~3 to 15 km) climate modeling is related to the lack of a seamless convection parameterization that works across many scales. Since many convection schemes are not developed to work at those "gray scales", they often lead to excessive precipitation during warm periods (e.g., summer). The Kain-Fritsch (KF) convection parameterization in the WRF model has been updated such that it can be used seamlessly across spatial scales down to ~1 km grid spacing. First, we introduced subgrid-scale cloud and radiation interactions that had not been previously considered in the KF scheme. Then, a scaling parameter was developed to introduce scale-dependency in the KF scheme for use with various processes. In addition, we developed new formulations for: (1) convective adjustment timescale; (2) entrainment of environmental air; (3) impacts of convective updraft on grid-scale vertical velocity; (4) convective cloud microphysics; (5) stabilizing capacity; (6) elimination of double counting of precipitation; and (7) estimation of updraft mass flux at the lifting condensation level. Some of these scale-dependent formulations make the KF scheme operable at all scales up to about sub-kilometer grid resolution. In this presentation, regional climate simulations using the WRF model will be presented to demonstrate the effects of these changes to the KF scheme. Additionally, we briefly present results obtained from the improved representation of inland lakes, various nudging strategies, and added value of dynamical downscaling of regional climate. Requesting for a plenary talk for the session: "Regional climate modeling, including CORDEX" (session number CL6.4) at the EGU 2014 General Assembly, to be held 27 April - 2 May 2014 in Vienna, Austria.
NASA Astrophysics Data System (ADS)
Loftis, D.
2016-02-01
In the wake of Hurricane Katrina (2005), Hurricane Ike (2008) is the second most devastating tropical cyclone to make landfall in the Gulf of Mexico in recent history. The path of the eye of Hurricane Ike passing directly over the Galveston's City Center requires the finesse of a street-level hydrodynamic model to accurately resolve the spatial inundation extent observed during the storm. A version of the Holland wind model was coupled with a sub-grid hydrodynamic model to address the complexity of spatially-varying hurricane force winds on the irregular movement of fluid though the streets of the coastal cities adjacent to the Galveston Bay. Sub-grid modeling technology is useful for incorporating high-resolution lidar-derived elevation measurements into the conventional hydrodynamic modeling framework to resolve detailed topographic features for inclusion in a hydrological transport model for storm surge simulations. Buildings were mosaicked into a lidar-derived Digital Surface Model at 5m spatial resolution for the study area, and in turn, embedded within a sub-grid layer of the hydrodynamic model mesh in a cross-scale approach to address the movement of Ike's storm surge from the Gulf of Mexico through the Galveston Bay, up estuaries and onto land. Model predictions for timing and depth of flooding during Hurricane Ike were compared with 8 verified water level gauges throughout the study area to evaluate the effectiveness of the sub-grid model's partial wetting and drying scheme. Statistical comparison yielded a mean R2 of 0.914, a relative error of 4.19%, and a root-mean-squared error of 19.47cm. A rigorous point-to-point comparison between street-level model results and 217 high water mark observations collected by the USGS and FEMA at several sites after the storm revealed that the model predicted the depth of inundation comparably well with an aggregate root-mean-squared error 0.283m. Finally, sea-level rise scenarios using Hurricane Ike as a base case revealed future storm-induced inundation could extend 0.6-2.8 km inland corresponding to increases in mean sea level of 37.5-150 cm based upon IPCC climate change prediction scenarios specified in their 5th assessment report in 2013.
An Analysis of Performance Enhancement Techniques for Overset Grid Applications
NASA Technical Reports Server (NTRS)
Djomehri, J. J.; Biswas, R.; Potsdam, M.; Strawn, R. C.; Biegel, Bryan (Technical Monitor)
2002-01-01
The overset grid methodology has significantly reduced time-to-solution of high-fidelity computational fluid dynamics (CFD) simulations about complex aerospace configurations. The solution process resolves the geometrical complexity of the problem domain by using separately generated but overlapping structured discretization grids that periodically exchange information through interpolation. However, high performance computations of such large-scale realistic applications must be handled efficiently on state-of-the-art parallel supercomputers. This paper analyzes the effects of various performance enhancement techniques on the parallel efficiency of an overset grid Navier-Stokes CFD application running on an SGI Origin2000 machine. Specifically, the role of asynchronous communication, grid splitting, and grid grouping strategies are presented and discussed. Results indicate that performance depends critically on the level of latency hiding and the quality of load balancing across the processors.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lu, Yongqi; DeVries, Nicholas; Ruhter, David
A novel Hot Carbonate Absorption Process with Crystallization-Enabled High-Pressure Stripping (Hot-CAP) has been developed by the University of Illinois at Urbana-Champaign and Carbon Capture Scientific, LLC in this three-year, bench-scale project. The Hot-CAP features a concentrated carbonate solution (e.g., K{sub 2}CO{sub 3}) for CO{sub 2} absorption and a bicarbonate slurry (e.g., KHCO{sub 3}) for high-pressure CO{sub 2} stripping to overcome the energy use and other disadvantages associated with the benchmark monoethanolamine (MEA) process. The project was aimed at performing laboratory- and bench-scale experiments to prove its technical feasibility and generate process engineering and scale-up data, and conducting a techno-economic analysismore » (TEA) to demonstrate its energy use and cost competitiveness over MEA. To meet project goals and objectives, a combination of experimental, modeling, process simulation, and economic analysis studies were applied. Carefully designed and intensive experiments were conducted to measure thermodynamic and reaction engineering data relevant to four major unit operations in the Hot-CAP (i.e., CO{sub 2} absorption, CO{sub 2} stripping, bicarbonate crystallization, and sulfate reclamation). The rate promoters that could accelerate the CO{sub 2} absorption rate into the potassium carbonate/bicarbonate (PCB) solution to a level greater than that into the 5 M MEA solution were identified, and the superior performance of CO{sub 2} absorption into PCB was demonstrated in a bench-scale packed-bed column. Kinetic data on bicarbonate crystallization were developed and applied for crystallizer design and sizing. Parametric testing of high-pressure CO{sub 2} stripping with concentrated bicarbonate-dominant slurries at high temperatures ({>=}140{degrees}C) in a bench-scale stripping column demonstrated lower heat use than with MEA. The feasibility of a modified process for combining SO{sub 2} removal with CO{sub 2} capture was preliminarily demonstrated. In addition to the experimental studies, the technical challenges pertinent to fouling of slurry-handling equipment and the design of the crystallizer and stripper were addressed through consultation with vendors and engineering analyses. A process flow diagram of the Hot-CAP was then developed and a TEA was performed to compare the energy use and cost performance of a nominal 550-MWe subcritical pulverized coal (PC)-fired power plant without CO{sub 2} capture (DOE/NETL Case 9) with the benchmark MEA-based post-combustion CO{sub 2} capture (PCC; DOE/NETL Case 10) and the Hot-CAP-based PCC. The results revealed that the net power produced in the PC + Hot-CAP is 609 MWe, greater than the PC + MEA (550 MWe). The 20-year levelized cost of electricity (LCOE) for the PC + Hot-CAP, including CO{sub 2} transportation and storage, is 120.3 mills/kWh, a 60% increase over the base PC plant without CO{sub 2} capture. The LCOE increase for the Hot-CAP is 29% lower than that for MEA. TEA results demonstrated that the Hot-CAP is energy-efficient and cost-effective compared with the benchmark MEA process.« less
Dai, Ruizhi; Thomas, Ayanna K; Taylor, Holly A
2018-01-30
Research examining object identity and location processing in visuo-spatial working memory (VSWM) has yielded inconsistent results on whether age differences exist in VSWM. The present study investigated whether these inconsistencies may stem from age-related differences in VSWM sub-processes, and whether processing of component VSWM information can be facilitated. In two experiments, younger and older adults studied 5 × 5 grids containing five objects in separate locations. In a continuous recognition paradigm, participants were tested on memory for object identity, location, or identity and location information combined. Spatial and categorical relationships were manipulated within grids to provide trial-level facilitation. In Experiment 1, randomizing trial types (location, identity, combination) assured that participants could not predict the information that would be queried. In Experiment 2, blocking trials by type encouraged strategic processing. Thus, we manipulated the nature of the task through object categorical relationship and spatial organization, and trial blocking. Our findings support age-related declines in VSWM. Additionally, grid organizations (categorical and spatial relationships), and trial blocking differentially affected younger and older adults. Younger adults used spatial organizations more effectively whereas older adults demonstrated an association bias. Our finding also suggests that older adults may be less efficient than younger adults in strategically engaging information processing.
A Bayesian Analysis of Scale-Invariant Processes
2012-01-01
Earth Grid (EASE- Grid). The NED raster elevation data of one arc-second resolution (30 m) over the continental US are derived from multiple satellites ...instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send...empirical and ME distributions, yet ensuring computational efficiency. Instead of com- puting empirical histograms from large amount of data , only some
Challenges and Opportunities in Modeling of the Global Atmosphere
NASA Astrophysics Data System (ADS)
Janjic, Zavisa; Djurdjevic, Vladimir; Vasic, Ratko
2016-04-01
Modeling paradigms on global scales may need to be reconsidered in order to better utilize the power of massively parallel processing. For high computational efficiency with distributed memory, each core should work on a small subdomain of the full integration domain, and exchange only few rows of halo data with the neighbouring cores. Note that the described scenario strongly favors horizontally local discretizations. This is relatively easy to achieve in regional models. However, the spherical geometry complicates the problem. The latitude-longitude grid with local in space and explicit in time differencing has been an early choice and remained in use ever since. The problem with this method is that the grid size in the longitudinal direction tends to zero as the poles are approached. So, in addition to having unnecessarily high resolution near the poles, polar filtering has to be applied in order to use a time step of a reasonable size. However, the polar filtering requires transpositions involving extra communications as well as more computations. The spectral transform method and the semi-implicit semi-Lagrangian schemes opened the way for application of spectral representation. With some variations, such techniques are currently dominating in global models. Unfortunately, the horizontal non-locality is inherent to the spectral representation and implicit time differencing, which inhibits scaling on a large number of cores. In this respect the lat-lon grid with polar filtering is a step in the right direction, particularly at high resolutions where the Legendre transforms become increasingly expensive. Other grids with reduced variability of grid distances, such as various versions of the cubed sphere and the hexagonal/pentagonal ("soccer ball") grids, were proposed almost fifty years ago. However, on these grids, large-scale (wavenumber 4 and 5) fictitious solutions ("grid imprinting") with significant amplitudes can develop. Due to their large scales, that are comparable to the scales of the dominant Rossby waves, such fictitious solutions are hard to identify and remove. Another new challenge on the global scale is that the limit of validity of the hydrostatic approximation is rapidly being approached. Relaxing the hydrostatic approximation requieres careful reformulation of the model dynamics and more computations and communications. The unified Non-hydrostatic Multi-scale Model (NMMB) will be briefly discussed as an example. The non-hydrostatic dynamics were designed in such a way as to avoid over-specification. The global version is run on the latitude-longitude grid, and the polar filter selectively slows down the waves that would otherwise be unstable without modifying their amplitudes. The model has been successfully tested on various scales. The skill of the medium range forecasts produced by the NMMB is comparable to that of other major medium range models, and its computational efficiency on parallel computers is good.
The evaluation and development of the Met Office Unified Model using surface and space borne radar.
NASA Astrophysics Data System (ADS)
Petch, J.
2012-12-01
The Met Office Unified Model is used for the prediction of weather and climate on time scales of hours through to centuries. Therefore, the parametrizations in that model need to work on weather and climate timescale, and with grid-lengths from hundres of meters through to several hundred kilometres. Focusing on the development of the cloud and radiation schemes I will discuss how we are using ground-based remote-sensing observations from Chilbolton (England) and a combination of Cloudsat and Calipso data to evaluate and improve the performance of the model. I will show how the prediction of the clouds has improved since the AR5 version of the model and how we have developed an improved cloud generator to rebresent the sub-grid variability of clouds for radiative transfer.
A gridded hourly rainfall dataset for the UK applied to a national physically-based modelling system
NASA Astrophysics Data System (ADS)
Lewis, Elizabeth; Blenkinsop, Stephen; Quinn, Niall; Freer, Jim; Coxon, Gemma; Woods, Ross; Bates, Paul; Fowler, Hayley
2016-04-01
An hourly gridded rainfall product has great potential for use in many hydrological applications that require high temporal resolution meteorological data. One important example of this is flood risk management, with flooding in the UK highly dependent on sub-daily rainfall intensities amongst other factors. Knowledge of sub-daily rainfall intensities is therefore critical to designing hydraulic structures or flood defences to appropriate levels of service. Sub-daily rainfall rates are also essential inputs for flood forecasting, allowing for estimates of peak flows and stage for flood warning and response. In addition, an hourly gridded rainfall dataset has significant potential for practical applications such as better representation of extremes and pluvial flash flooding, validation of high resolution climate models and improving the representation of sub-daily rainfall in weather generators. A new 1km gridded hourly rainfall dataset for the UK has been created by disaggregating the daily Gridded Estimates of Areal Rainfall (CEH-GEAR) dataset using comprehensively quality-controlled hourly rain gauge data from over 1300 observation stations across the country. Quality control measures include identification of frequent tips, daily accumulations and dry spells, comparison of daily totals against the CEH-GEAR daily dataset, and nearest neighbour checks. The quality control procedure was validated against historic extreme rainfall events and the UKCP09 5km daily rainfall dataset. General use of the dataset has been demonstrated by testing the sensitivity of a physically-based hydrological modelling system for Great Britain to the distribution and rates of rainfall and potential evapotranspiration. Of the sensitivity tests undertaken, the largest improvements in model performance were seen when an hourly gridded rainfall dataset was combined with potential evapotranspiration disaggregated to hourly intervals, with 61% of catchments showing an increase in NSE between observed and simulated streamflows as a result of more realistic sub-daily meteorological forcing.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Zihan; Swantek, Andrew; Scarcelli, Riccardo
This paper focuses on detailed numerical simulations of direct injection diesel and gasoline sprays from production grade, multi-hole injectors. In a dual-fuel engine the direct injection of both the fuels can facilitate appropriate mixture preparation prior to ignition and combustion. Diesel and gasoline sprays were simulated using high-fidelity Large Eddy Simulations (LES) with the dynamic structure sub-grid scale model. Numerical predictions of liquid penetration, fuel density distribution as well as transverse integrated mass (TIM) at different axial locations versus time were compared against x-ray radiography data obtained from Argonne National Laboratory. A necessary, but often overlooked, criterion of grid-convergence ismore » ensured by using Adaptive Mesh Refinement (AMR) for both diesel and gasoline. Nine different realizations were performed and the effects of random seeds on spray behavior were investigated. Additional parametric studies under different ambient and injection conditions were performed to study their influence on global and local flow structures for gasoline sprays. It is concluded that LES can generally well capture all experimental trends and comes close to matching the x-ray data. Discrepancies between experimental and simulation results can be correlated to uncertainties in boundary and initial conditions such as rate of injection and spray and turbulent dispersion sub-model constants.« less
Improved Climate Simulations through a Stochastic Parameterization of Ocean Eddies
NASA Astrophysics Data System (ADS)
Williams, Paul; Howe, Nicola; Gregory, Jonathan; Smith, Robin; Joshi, Manoj
2016-04-01
In climate simulations, the impacts of the sub-grid scales on the resolved scales are conventionally represented using deterministic closure schemes, which assume that the impacts are uniquely determined by the resolved scales. Stochastic parameterization relaxes this assumption, by sampling the sub-grid variability in a computationally inexpensive manner. This presentation shows that the simulated climatological state of the ocean is improved in many respects by implementing a simple stochastic parameterization of ocean eddies into a coupled atmosphere-ocean general circulation model. Simulations from a high-resolution, eddy-permitting ocean model are used to calculate the eddy statistics needed to inject realistic stochastic noise into a low-resolution, non-eddy-permitting version of the same model. A suite of four stochastic experiments is then run to test the sensitivity of the simulated climate to the noise definition, by varying the noise amplitude and decorrelation time within reasonable limits. The addition of zero-mean noise to the ocean temperature tendency is found to have a non-zero effect on the mean climate. Specifically, in terms of the ocean temperature and salinity fields both at the surface and at depth, the noise reduces many of the biases in the low-resolution model and causes it to more closely resemble the high-resolution model. The variability of the strength of the global ocean thermohaline circulation is also improved. It is concluded that stochastic ocean perturbations can yield reductions in climate model error that are comparable to those obtained by refining the resolution, but without the increased computational cost. Therefore, stochastic parameterizations of ocean eddies have the potential to significantly improve climate simulations. Reference PD Williams, NJ Howe, JM Gregory, RS Smith, and MM Joshi (2016) Improved Climate Simulations through a Stochastic Parameterization of Ocean Eddies. Journal of Climate, under revision.
Improvements for retrieval of cloud droplet size by the POLDER instrument
NASA Astrophysics Data System (ADS)
Shang, H.; Husi, L.; Bréon, F. M.; Ma, R.; Chen, L.; Wang, Z.
2017-12-01
The principles of cloud droplet size retrieval via Polarization and Directionality of the Earth's Reflectance (POLDER) requires that clouds be horizontally homogeneous. The retrieval is performed by combining all measurements from an area of 150 km × 150 km to compensate for POLDER's insufficient directional sampling. Using POLDER-like data simulated with the RT3 model, we investigate the impact of cloud horizontal inhomogeneity and directional sampling on the retrieval and analyze which spatial resolution is potentially accessible from the measurements. Case studies show that the sub-grid-scale variability in droplet effective radius (CDR) can significantly reduce valid retrievals and introduce small biases to the CDR ( 1.5µm) and effective variance (EV) estimates. Nevertheless, the sub-grid-scale variations in EV and cloud optical thickness (COT) only influence the EV retrievals and not the CDR estimate. In the directional sampling cases studied, the retrieval using limited observations is accurate and is largely free of random noise. Several improvements have been made to the original POLDER droplet size retrieval. For example, measurements in the primary rainbow region (137-145°) are used to ensure retrievals of large droplet (>15 µm) and to reduce the uncertainties caused by cloud heterogeneity. A premium resoltion of 0.8° is determined by considering successful retrievals and cloud horizontal homogeneity. The improved algorithm is applied to measurements of POLDER in 2008, and we further compared our retrievals with cloud effective radii estimations of Moderate Resolution Imaging Spectroradiometer (MODIS). The results indicate that in global scale, the cloud effective radii and effective variance is larger in the central ocean than inland and coast areas. Over heavy polluted regions, the cloud droplets has small effective radii and narraw distribution due to the influence of aerosol particles.
Integrating Unified Gravity Wave Physics into the NOAA Next Generation Global Prediction System
NASA Astrophysics Data System (ADS)
Alpert, J. C.; Yudin, V.; Fuller-Rowell, T. J.; Akmaev, R. A.
2017-12-01
The Unified Gravity Wave Physics (UGWP) project for the Next Generation Global Prediction System (NGGPS) is a NOAA collaborative effort between the National Centers for Environmental Prediction (NCEP), Environemntal Modeling Center (EMC) and the University of Colorado, Cooperative Institute for Research in Environmental Sciences (CU-CIRES) to support upgrades and improvements of GW dynamics (resolved scales) and physics (sub-grid scales) in the NOAA Environmental Modeling System (NEMS)†. As envisioned the global climate, weather and space weather models of NEMS will substantially improve their predictions and forecasts with the resolution-sensitive (scale-aware) formulations planned under the UGWP framework for both orographic and non-stationary waves. In particular, the planned improvements for the Global Forecast System (GFS) model of NEMS are: calibration of model physics for higher vertical and horizontal resolution and an extended vertical range of simulations, upgrades to GW schemes, including the turbulent heating and eddy mixing due to wave dissipation and breaking, and representation of the internally-generated QBO. The main priority of the UGWP project is unified parameterization of orographic and non-orographic GW effects including momentum deposition in the middle atmosphere and turbulent heating and eddies due to wave dissipation and breaking. The latter effects are not currently represented in NOAA atmosphere models. The team has tested and evaluated four candidate GW solvers integrating the selected GW schemes into the NGGPS model. Our current work and planned activity is to implement the UGWP schemes in the first available GFS/FV3 (open FV3) configuration including adapted GFDL modification for sub-grid orography in GFS. Initial global model results will be shown for the operational and research GFS configuration for spectral and FV3 dynamical cores. †http://www.emc.ncep.noaa.gov/index.php?branch=NEMS
Comparison of AGE and Spectral Methods for the Simulation of Far-Wakes
NASA Technical Reports Server (NTRS)
Bisset, D. K.; Rogers, M. M.; Kega, Dennis (Technical Monitor)
1999-01-01
Turbulent flow simulation methods based on finite differences are attractive for their simplicity, flexibility and efficiency, but not always for accuracy or stability. This report demonstrates that a good compromise is possible with the Advected Grid Explicit (AGE) method. AGE has proven to be both efficient and accurate for simulating turbulent free-shear flows, including planar mixing layers and planar jets. Its efficiency results from its localized fully explicit finite difference formulation (Bisset 1998a,b) that is very straightforward to compute, outweighing the need for a fairly small timestep. Also, most of the successful simulations were slightly under-resolved, and therefore they were, in effect, large-eddy simulations (LES) without a sub-grid-scale (SGS) model, rather than direct numerical simulations (DNS). The principle is that the role of the smallest scales of turbulent motion (when the Reynolds number is not too low) is to dissipate turbulent energy, and therefore they do not have to be simulated when the numerical method is inherently dissipative at its resolution limits. Such simulations are termed 'auto-LES' (LES with automatic SGS modeling) in this report.
INITIAL APPL;ICATION OF THE ADAPTIVE GRID AIR POLLUTION MODEL
The paper discusses an adaptive-grid algorithm used in air pollution models. The algorithm reduces errors related to insufficient grid resolution by automatically refining the grid scales in regions of high interest. Meanwhile the grid scales are coarsened in other parts of the d...
NASA Astrophysics Data System (ADS)
Wood, M.; Neal, J. C.; Hostache, R.; Corato, G.; Chini, M.; Giustarini, L.; Matgen, P.; Wagener, T.; Bates, P. D.
2015-12-01
Synthetic Aperture Radar (SAR) satellites are capable of all-weather day and night observations that can discriminate between land and smooth open water surfaces over large scales. Because of this there has been much interest in the use of SAR satellite data to improve our understanding of water processes, in particular for fluvial flood inundation mechanisms. Past studies prove that integrating SAR derived data with hydraulic models can improve simulations of flooding. However while much of this work focusses on improving model channel roughness values or inflows in ungauged catchments, improvement of model bathymetry is often overlooked. The provision of good bathymetric data is critical to the performance of hydraulic models but there are only a small number of ways to obtain bathymetry information where no direct measurements exist. Spatially distributed river depths are also rarely available. We present a methodology for calibration of model average channel depth and roughness parameters concurrently using SAR images of flood extent and a Sub-Grid model utilising hydraulic geometry concepts. The methodology uses real data from the European Space Agency's archive of ENVISAT[1] Wide Swath Mode images of the River Severn between Worcester and Tewkesbury during flood peaks between 2007 and 2010. Historic ENVISAT WSM images are currently free and easy to access from archive but the methodology can be applied with any available SAR data. The approach makes use of the SAR image processing algorithm of Giustarini[2] et al. (2013) to generate binary flood maps. A unique feature of the calibration methodology is to also use parameter 'identifiability' to locate the parameters with higher accuracy from a pre-assigned range (adopting the DYNIA method proposed by Wagener[3] et al., 2003). [1] https://gpod.eo.esa.int/services/ [2] Giustarini. 2013. 'A Change Detection Approach to Flood Mapping in Urban Areas Using TerraSAR-X'. IEEE Transactions on Geoscience and Remote Sensing, vol. 51, no. 4. [3] Wagener. 2003. 'Towards reduced uncertainty in conceptual rainfall-runoff modelling: Dynamic identifiability analysis'. Hydrol. Process. 17, 455-476.
NASA Astrophysics Data System (ADS)
Kumar, Dheeraj; Gautam, Amar Kant; Palmate, Santosh S.; Pandey, Ashish; Suryavanshi, Shakti; Rathore, Neha; Sharma, Nayan
2017-08-01
To support the GPM mission which is homologous to its predecessor, the Tropical Rainfall Measuring Mission (TRMM), this study has been undertaken to evaluate the accuracy of Tropical Rainfall Measuring Mission multi-satellite precipitation analysis (TMPA) daily-accumulated precipitation products for 5 years (2008-2012) using the statistical methods and contingency table method. The analysis was performed on daily, monthly, seasonal and yearly basis. The TMPA precipitation estimates were also evaluated for each grid point i.e. 0.25° × 0.25° and for 18 rain gauge stations of the Betwa River basin, India. Results indicated that TMPA precipitation overestimates the daily and monthly precipitation in general, particularly for the middle sub-basin in the non-monsoon season. Furthermore, precision of TMPA precipitation estimates declines with the decrease of altitude at both grid and sub-basin scale. The study also revealed that TMPA precipitation estimates provide better accuracy in the upstream of the basin compared to downstream basin. Nevertheless, the detection capability of daily TMPA precipitation improves with increase in altitude for drizzle rain events. However, the detection capability decreases during non-monsoon and monsoon seasons when capturing moderate and heavy rain events, respectively. The veracity of TMPA precipitation estimates was improved during the rainy season than during the dry season at all scenarios investigated. The analyses suggest that there is a need for better precipitation estimation algorithm and extensive accuracy verification against terrestrial precipitation measurement to capture the different types of rain events more reliably over the sub-humid tropical regions of India.
NASA Astrophysics Data System (ADS)
Pitman, Andrew J.; Yang, Zong-Liang; Henderson-Sellers, Ann
1993-10-01
The sensitivity of a land surface scheme to the distribution of precipitation within a general circulation model's grid element is investigated. Earlier experiments which showed considerable sensitivity of the runoff and evaporation simulation to the distribution of precipitation are repeated in the light of other results which show no sensitivity of evaporation to the distribution of precipitation. Results show that while the earlier results over-estimated the sensitivity of the surface hydrology to the precipitation distribution, the general conclusion that the system is sensitive is supported. It is found that changing the distribution of precipitation from falling over 100% of the grid square to falling over 10% leads to a reduction in evaporation from 1578 mm y-1 to 1195 mm y -1 while runoff increases from 278 mm y-1 to 602 mm y-1. The sensitivity is explained in terms of evaporation being dominated by available energy when precipitation falls over nearly the entire grid square, but by moisture availability (mainly intercepted water) when it falls over little of the grid square. These results also indicate that earlier work using stand-alone forcing to drive land surface schemes ‘off-line’, and to investigate the sensitivity of land surface codes to various parameters, leads to results which are non-repeatable in single column simulations.
NASA Astrophysics Data System (ADS)
Montzka, C.; Rötzer, K.; Bogena, H. R.; Vereecken, H.
2017-12-01
Improving the coarse spatial resolution of global soil moisture products from SMOS, SMAP and ASCAT is currently an up-to-date topic. Soil texture heterogeneity is known to be one of the main sources of soil moisture spatial variability. A method has been developed that predicts the soil moisture standard deviation as a function of the mean soil moisture based on soil texture information. It is a closed-form expression using stochastic analysis of 1D unsaturated gravitational flow in an infinitely long vertical profile based on the Mualem-van Genuchten model and first-order Taylor expansions. With the recent development of high resolution maps of basic soil properties such as soil texture and bulk density, relevant information to estimate soil moisture variability within a satellite product grid cell is available. Here, we predict for each SMOS, SMAP and ASCAT grid cell the sub-grid soil moisture variability based on the SoilGrids1km data set. We provide a look-up table that indicates the soil moisture standard deviation for any given soil moisture mean. The resulting data set provides important information for downscaling coarse soil moisture observations of the SMOS, SMAP and ASCAT missions. Downscaling SMAP data by a field capacity proxy indicates adequate accuracy of the sub-grid soil moisture patterns.
NASA Astrophysics Data System (ADS)
Siegenthaler-Le Drian, C.; Spichtinger, P.; Lohmann, U.
2010-09-01
Marine stratocumulus-capped boundary layers exhibit a strong net cooling impact on the Earth-Atmosphere system. Moreover, they are highly persistent over subtropical oceans. Therefore climate models need to represent them well in order to make reliable projections of future climate. One of the reasons for the absence of stratocumuli in the general circulation model ECHAM5-HAM (Roeckner et al., 2003; Stier et al., 2005) is due to the limited vertical resolution. In the current model version, no vertical sub-grid scale variability of clouds is taken into account, such that clouds occupy the full vertical layer. Around the inversion on top of the planetary boundary layer (PBL), conserved variables often have a steep gradient, which in a GCM may produce large discretization errors (Bretherton and Park, 2009). This inversion has a large diurnal cycle and varies with location around the globe, which is difficult to represent in a classical, coarse Eulerian approach. Furthermore, Lenderink and Holtslag (2000) and Lock (2001) showed that an inconsistent numerical representation between the entrainment parametrization and the other schemes, particularly with the vertical advection can lead to the occurrence of 'numerical entrainment'. The problem can be resolved by introducing a dynamical inversion as introduced by Grenier and Bretherton (2001) and Lock (2001). As these features can be seen in our version of ECHAM5-HAM, our implementation is aimed to reduce the numerical entrainment and to better represent stratocumuli in ECHAM5-HAM. To better resolve stratocumulus clouds, their inversion and the interaction between the turbulent diffusion and the vertical advection, the vertical grid is dynamically refined. The new grid is based on the reconstruction of the profiles of variables experiencing a sharp gradient (temperature, mixing ratio) applying the method presented in Grenier and Bretherton (2001). In typical stratocumulus regions, an additional grid level is thus associated with the PBL top. In case a cloud can be formed, a new level is associated with the lifting condensation level as well. The regular grid plus the two additional levels define the new dynamical grid, which varies geographically and temporally. The physical processes are computed on this new dynamical grid, Consequently, the sharp gradients and the interaction between the different processes can be better resolved. Some results of this new parametrization will be presented. On a single column model set-up, the reconstruction method accurately finds the inversion at the PBL top for the EPIC stratocumulus case. Also, on a global scale, the occurrence of a successful reconstruction, which is restricted in typical stratocumulus regions, occurs with a high frequency. The impact of the new dynamical grid on clouds and the radiation balance will be presented in the talk. References [Bretherton and Park, 2009] Bretherton, C. S. and Park, S. (2009). A new moist turbulence parametrization in the community atmosphere model. J. Climate, 22:3422-3448. [Grenier and Bretherton, 2001] Grenier, H. and Bretherton, C. S. (2001). A moist parametrization for large-scale models and its application to subtropical cloud-topped marine boundary layers. Mon. Wea. Rev., 129:357-377. [Lenderink and Holtslag, 2000] Lenderink, G. and Holtslag, A. M. (2000). Evaluation of the kinetic energy approach for modeling turbulent fluxes in stratocumulus. Mon. Wea. Rev., 128:244-258. [Lock, 2001] Lock, A. P. (2001). The numerical representation of entrainment in parametrizations of boundary layer turbulent mixing. Mon. Wea. Rev., 129:1148-1163. [Roeckner et al., 2003] Roeckner, E., Bäuml, G., Bonaventura, L. et al. (2003). The atmospheric general circulation model echam5, part I: Model description. Technical Report 349, Max-Planck-Institute for Meteorology, Hamburg,Germany. [Stier et al., 2005] Stier, P., Feichter, J., Kinne, S. et al. (2005). The aerosol-climate model ECHAM5-HAM. Atmos. Chem. Phys., 5:1125-1156.
Using Computing and Data Grids for Large-Scale Science and Engineering
NASA Technical Reports Server (NTRS)
Johnston, William E.
2001-01-01
We use the term "Grid" to refer to a software system that provides uniform and location independent access to geographically and organizationally dispersed, heterogeneous resources that are persistent and supported. These emerging data and computing Grids promise to provide a highly capable and scalable environment for addressing large-scale science problems. We describe the requirements for science Grids, the resulting services and architecture of NASA's Information Power Grid (IPG) and DOE's Science Grid, and some of the scaling issues that have come up in their implementation.
California's Snow Gun and its implications for mass balance predictions under greenhouse warming
NASA Astrophysics Data System (ADS)
Howat, I.; Snyder, M.; Tulaczyk, S.; Sloan, L.
2003-12-01
Precipitation has received limited treatment in glacier and snowpack mass balance models, largely due to the poor resolution and confidence of precipitation predictions relative to temperature predictions derived from atmospheric models. Most snow and glacier mass balance models rely on statistical or lapse rate-based downscaling of general or regional circulation models (GCM's and RCM's), essentially decoupling sub-grid scale, orographically-driven evolution of atmospheric heat and moisture. Such models invariably predict large losses in the snow and ice volume under greenhouse warming. However, positive trends in the mass balance of glaciers in some warming maritime climates, as well as at high elevations of the Greenland Ice Sheet, suggest that increased precipitation may play an important role in snow- and glacier-climate interactions. Here, we present a half century of April snowpack data from the Sierra Nevada and Cascade mountains of California, USA. This high-density network of snow-course data indicates that a gain in winter snow accumulation at higher elevations has compensated loss in snow volume at lower elevations by over 50% and has led to glacier expansion on Mt. Shasta. These trends are concurrent with a region-wide increase in winter temperatures up to 2° C. They result from the orographic lifting and saturation of warmer, more humid air leading to increased precipitation at higher elevations. Previous studies have invoked such a "Snow Gun" effect to explain contemporaneous records of Tertiary ocean warming and rapid glacial expansion. A climatological context of the California's "snow gun" effect is elucidated by correlation between the elevation distribution of April SWE observations and the phase of the Pacific Decadal Oscillation and the El Nino Southern Oscillation, both controlling the heat and moisture delivered to the U.S. Pacific coast. The existence of a significant "Snow Gun" effect presents two challenges to snow and glacier mass balance modeling. Firstly, the link between amplification of orographic precipitation and the temporal evolution of ocean-climate oscillations indicates that prediction of future mass balance trends requires consideration of the timing and amplitude of such oscillations. Only recently have ocean-atmosphere models begun to realistically produce such temporal variability. Secondly, the steepening snow mass-balance elevation-gradient associated with the "Snow Gun" implies greater spatial variability in balance with warming. In a warming climate, orographic processes at a scale finer that the highest resolution RCM (>20km grid) become increasingly important and predictions based on lower elevations become increasingly inaccurate for higher elevations. Therefore, thermodynamic interaction between atmospheric heat, moisture and topography must be included in downscaling techniques. In order to demonstrate the importance of the thermodynamic downscaling in mass balance predictions, we nest a high-resolution (100m grid), coupled Orographic Precipitation and Surface Energy balance Model (OPSEM) into the RegC2.5 RCM (40 km grid) and compare results. We apply this nesting technique to Mt. Shasta, California, an area of high topography (~4000m) relative to its RegCM2.5 grid elevation (1289m). These models compute average April snow volume under present and doubled-present Atmospheric CO2 concentrations. While the RegCM2.5 regional model predicts an 83% decrease in April SWE, OPSEM predicts a 16% increase. These results indicate that thermodynamic interactions between the atmosphere and topography at sub- RCM grid resolution must be considered in mass balance models.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Hoyoung; Korea Institute of Materials Science, 797 Changwon-daero, Seongsan-gu, Changwon, Gyeongnam 642-831; Kang, Jun-Yun, E-mail: firice@kims.re.kr
This study aimed to present the complete history of carbide evolution in a cold-work tool steel along its full processing route for fabrication and application. A sequence of processes from cast to final hardening heat treatment was conducted on an 8% Cr-steel to reproduce a typical commercial processing route in a small scale. The carbides found at each process step were then identified by electron diffraction with energy dispersive spectroscopy in a scanning or transmission electron microscope. After solidification, MC, M{sub 7}C{sub 3} and M{sub 2}C carbides were identified and the last one dissolved during hot compression at 1180 °C.more » In a subsequent annealing at 870 °C followed by slow cooling, M{sub 6}C and M{sub 23}C{sub 6} were added, while they were dissolved in the following austenitization at 1030 °C. After the final tempering at 520 °C, fine M{sub 23}C{sub 6} precipitated again, thus the final microstructure was the tempered martensite with MC, M{sub 7}C{sub 3} and M{sub 23}C{sub 6} carbide. The transient M{sub 2}C and M{sub 6}C originated from the segregation of Mo and finally disappeared due to attenuated segregation and the consequent thermodynamic instability. - Highlights: • The full processing route of a cold-work tool steel was simulated in a small scale. • The carbides in the tool steel were identified by chemical–crystallographic analyses. • MC, M{sub 7}C{sub 3}, M{sub 2}C, M{sub 6}C and M{sub 23}C{sub 6} carbides were found during the processing of the steel. • M{sub 2}C and M{sub 6}C finally disappeared due to thermodynamic instability.« less
Silicon cells made by self-aligned selective-emitter plasma-etchback process
Ruby, Douglas S.; Schubert, William K.; Gee, James M.; Zaidi, Saleem H.
2000-01-01
Photovoltaic cells and methods for making them are disclosed wherein the metallized grids of the cells are used to mask portions of cell emitter regions to allow selective etching of phosphorus-doped emitter regions. The preferred etchant is SF.sub.6 or a combination of SF.sub.6 and O.sub.2. This self-aligned selective etching allows for enhanced blue response (versus cells with uniform heavy doping of the emitter) while preserving heavier doping in the region beneath the gridlines needed for low contact resistance. Embodiments are disclosed for making cells with or without textured surfaces. Optional steps include plasma hydrogenation and PECVD nitride deposition, each of which are suited to customized applications for requirements of given cells to be manufactured. The techniques disclosed could replace expensive and difficult alignment methodologies used to obtain selectively etched emitters, and they may be easily integrated with existing plasma processing methods and techniques of the invention may be accomplished in a single plasma-processing chamber.
Validation of Smithsonian Astrophysical Observatory's OMI Water Vapor Product
NASA Astrophysics Data System (ADS)
Wang, H.; Gonzalez Abad, G.; Liu, X.; Chance, K.
2015-12-01
We perform a comprehensive validation of SAO's OMI water vapor product. The SAO OMI water vapor slant column is retrieved using the 430 - 480 nm wavelength range. In addition to water vapor, the retrieval considers O3, NO2, liquid water, O4, C2H2O2, the Ring effect, water ring, 3rd order polynomial, common mode and under-sampling. The slant column is converted to vertical column using AMF. AMF is calculated using GEOS-Chem water vapor profile shape, OMCLDO2 cloud information and OMLER surface albedo information. We validate our product using NCAR's GPS network data over the world and RSS's gridded microwave data over the ocean. We also compare our product with the total precipitable water derived from the AERONET ground-based sun photometer data, the GlobVapour gridded product, and other datasets. We investigate the influence of sub-grid scale variability and filtering criteria on the comparison. We study the influence of clouds, aerosols and a priori profiles on the retrieval. We also assess the long-term performance and stability of our product and seek ways to improve it.
Numerical Simulation of Bow Waves and Transom-Stern Flows
NASA Astrophysics Data System (ADS)
Dommermuth, Douglas G.; Schlageter, Eric A.; Talcott, John C.; Wyatt, Donald C.; Novikov, Evgeny A.
1997-11-01
A stratified-flow formulation is used to model the breaking bow wave and the separated transom-stern flow that are generated by a ship moving with forward speed. The interface of the air with the water is identified as the zero level-set of a three-dimensional function. The ship is modeled using a body-force technique on a cartesian grid. The three-dimensional body-force is generated using a surface panelization of the entire ship, including the above-water geometry up to and including the deck. The effects of surface tension are modeled as a source term that is concentrated at the air-water interface. The effects of gravity are modeled as a volumetric force. The three-dimensional, unsteady, Navier-Stokes equations are expressed in primitive-variable form. A LES formulation with a Smagorinsky sub-grid-scale model is used to model turbulence. Numerical convergence is demonstrated using 128x64x65, 256x128x129, and 512x256x257 grid points. The numerical results compare well to whisker-probe measurements of the free-surface elevation generated by a naval combatant.
Variational estimation of process parameters in a simplified atmospheric general circulation model
NASA Astrophysics Data System (ADS)
Lv, Guokun; Koehl, Armin; Stammer, Detlef
2016-04-01
Parameterizations are used to simulate effects of unresolved sub-grid-scale processes in current state-of-the-art climate model. The values of the process parameters, which determine the model's climatology, are usually manually adjusted to reduce the difference of model mean state to the observed climatology. This process requires detailed knowledge of the model and its parameterizations. In this work, a variational method was used to estimate process parameters in the Planet Simulator (PlaSim). The adjoint code was generated using automatic differentiation of the source code. Some hydrological processes were switched off to remove the influence of zero-order discontinuities. In addition, the nonlinearity of the model limits the feasible assimilation window to about 1day, which is too short to tune the model's climatology. To extend the feasible assimilation window, nudging terms for all state variables were added to the model's equations, which essentially suppress all unstable directions. In identical twin experiments, we found that the feasible assimilation window could be extended to over 1-year and accurate parameters could be retrieved. Although the nudging terms transform to a damping of the adjoint variables and therefore tend to erases the information of the data over time, assimilating climatological information is shown to provide sufficient information on the parameters. Moreover, the mechanism of this regularization is discussed.
LLMapReduce: Multi-Level Map-Reduce for High Performance Data Analysis
2016-05-23
LLMapReduce works with several schedulers such as SLURM, Grid Engine and LSF. Keywords—LLMapReduce; map-reduce; performance; scheduler; Grid Engine ...SLURM; LSF I. INTRODUCTION Large scale computing is currently dominated by four ecosystems: supercomputing, database, enterprise , and big data [1...interconnects [6]), High performance math libraries (e.g., BLAS [7, 8], LAPACK [9], ScaLAPACK [10]) designed to exploit special processing hardware, High
Genomic analysis of regulatory network dynamics reveals large topological changes
NASA Astrophysics Data System (ADS)
Luscombe, Nicholas M.; Madan Babu, M.; Yu, Haiyuan; Snyder, Michael; Teichmann, Sarah A.; Gerstein, Mark
2004-09-01
Network analysis has been applied widely, providing a unifying language to describe disparate systems ranging from social interactions to power grids. It has recently been used in molecular biology, but so far the resulting networks have only been analysed statically. Here we present the dynamics of a biological network on a genomic scale, by integrating transcriptional regulatory information and gene-expression data for multiple conditions in Saccharomyces cerevisiae. We develop an approach for the statistical analysis of network dynamics, called SANDY, combining well-known global topological measures, local motifs and newly derived statistics. We uncover large changes in underlying network architecture that are unexpected given current viewpoints and random simulations. In response to diverse stimuli, transcription factors alter their interactions to varying degrees, thereby rewiring the network. A few transcription factors serve as permanent hubs, but most act transiently only during certain conditions. By studying sub-network structures, we show that environmental responses facilitate fast signal propagation (for example, with short regulatory cascades), whereas the cell cycle and sporulation direct temporal progression through multiple stages (for example, with highly inter-connected transcription factors). Indeed, to drive the latter processes forward, phase-specific transcription factors inter-regulate serially, and ubiquitously active transcription factors layer above them in a two-tiered hierarchy. We anticipate that many of the concepts presented here-particularly the large-scale topological changes and hub transience-will apply to other biological networks, including complex sub-systems in higher eukaryotes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Myers, A.M.; Ruzic, D.N.; Powell, R.C.
An analyzer capable of determining the mass as well as the energy (5--200 eV) of neutral and ion species has been developed from a quadrupole mass spectrometer (QMS). The system, which is similar to a retarding grid energy analyzer (RGEA), functions by biasing the rods of a QMS and monitoring the analyzer signal as a function of bias potential. Modulation of the pole bias greatly increases the minimum detectable signal level. Experiments were performed using species generated in a single-grid Kaufman ion gun operated with N{sub 2} or Ar. Results show that the pole bias techniques can provide energy resolutionmore » of 1--2 eV. Ion species from the gun were found to have an energy equal to the sum of the beam and the plasma potentials, with an energy spread between 1 and 3 eV. Fast N{sub 2} and Ar neutral species were measured as a function of discharge voltage (30--80 V), beam acceleration voltage (50--100 V), grid voltage ({minus}20 to +5 V), and pressure (0.5 and 1.5 mTorr). The energy of the fast neutral species was always less than that of the ions. This was consistent with the fast neutrals being formed by a charge-exchange process.« less
Towards retrieving critical relative humidity from ground-based remote sensing observations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Van Weverberg, Kwinten; Boutle, Ian; Morcrette, Cyril J.
2016-08-22
Nearly all parameterisations of large-scale cloud require the specification of the critical relative humidity (RHcrit). This is the gridbox-mean relative humidity at which the subgrid fluctuations in temperature and water vapour become so large that part of a subsaturated gridbox becomes saturated and cloud starts to form. Until recently, the lack of high-resolution observations of temperature and moisture variability has hindered a reasonable estimate of the RHcrit from observations. However, with the advent of ground-based measurements from Raman lidar, it becomes possible to obtain long records of temperature and moisture (co-)variances with sub-minute sample rates. Lidar observations are inherently noisymore » and any analysis of higher-order moments will be very dependent on the ability to quantify and remove this noise. We present an exporatory study aimed at understanding whether current noise levels of lidar-retrieved temperature and water vapour are sufficient to obtain a reasonable estimate of the RHcrit. We show that vertical profiles of RHcrit can be derived for a gridbox length of up to about 30 km (120) with an uncertainty of about 4 % (2 %). RHcrit tends to be smallest near the scale height and seems to be fairly insensitive to the horizontal grid spacing at the scales investigated here (30 - 120 km). However, larger sensitivity was found to the vertical grid spacing. As the grid spacing decreases from 400 to 100 m, RHcrit is observed to increase by about 6 %, which is more than the uncertainty in the RHcrit retrievals.« less
Modeling and Grid Generation of Iced Airfoils
NASA Technical Reports Server (NTRS)
Vickerman, Mary B.; Baez, Marivell; Braun, Donald C.; Hackenberg, Anthony W.; Pennline, James A.; Schilling, Herbert W.
2007-01-01
SmaggIce Version 2.0 is a software toolkit for geometric modeling and grid generation for two-dimensional, singleand multi-element, clean and iced airfoils. A previous version of SmaggIce was described in Preparing and Analyzing Iced Airfoils, NASA Tech Briefs, Vol. 28, No. 8 (August 2004), page 32. To recapitulate: Ice shapes make it difficult to generate quality grids around airfoils, yet these grids are essential for predicting ice-induced complex flow. This software efficiently creates high-quality structured grids with tools that are uniquely tailored for various ice shapes. SmaggIce Version 2.0 significantly enhances the previous version primarily by adding the capability to generate grids for multi-element airfoils. This version of the software is an important step in streamlining the aeronautical analysis of ice airfoils using computational fluid dynamics (CFD) tools. The user may prepare the ice shape, define the flow domain, decompose it into blocks, generate grids, modify/divide/merge blocks, and control grid density and smoothness. All these steps may be performed efficiently even for the difficult glaze and rime ice shapes. Providing the means to generate highly controlled grids near rough ice, the software includes the creation of a wrap-around block (called the "viscous sublayer block"), which is a thin, C-type block around the wake line and iced airfoil. For multi-element airfoils, the software makes use of grids that wrap around and fill in the areas between the viscous sub-layer blocks for all elements that make up the airfoil. A scripting feature records the history of interactive steps, which can be edited and replayed later to produce other grids. Using this version of SmaggIce, ice shape handling and grid generation can become a practical engineering process, rather than a laborious research effort.
Z-Earth: 4D topography from space combining short-baseline stereo and lidar
NASA Astrophysics Data System (ADS)
Dewez, T. J.; Akkari, H.; Kaab, A. M.; Lamare, M. L.; Doyon, G.; Costeraste, J.
2013-12-01
The advent of free-of-charge global topographic data sets SRTM and Aster GDEM have enabled testing a host of geoscience hypotheses. Availability of such data is now considered standard, and though resolved at 30-m to 90-m pixel size, they are today regarded as obsolete and inappropriate given the regularly updated sub-meter imagery coming through web services like Google Earth. Two features will thus help meet the current topographic data needs of the Geoscience communities: field-scale-compatible elevation datasets (i.e. meter-scale digital models and sub-meter elevation precision) and provision for regularly updated topography to tackle earth surface changes in 4D, while retaining the key for success: data availability at no charge. A new space borne instrumental concept called Z-Earth has undergone phase 0 study at CNES, the French space agency to fulfill these aims. The scientific communities backing this proposal are that of natural hazards, glaciology and biomass. The system under study combines a short-baseline native stereo imager and a lidar profiler. This combination provides spatially resolved elevation swaths together with absolute along-track elevation control point profiles. Acquisition is designed for revisit time better than a year. Intended products not only target single pass digital surface models, color orthoimages and small footprint full-wave-form lidar profiles to update existing topographic coverage, but also time series of them. 3D change detection targets centimetre-scale horizontal precision and metric vertical precision, in complement of -now traditional- spectral change detection. To assess the actual concept value, two real-size experiments were carried out. We used sub-meter-scale Pleiades panchromatic stereo-images to generate digital surface models and check them against dense airborne lidar coverages, one heliborne set purposely flown in Corsica (50-100pts/sq.m) and a second one retrieved from OpenTopography.org (~10pts/sq.m.). In Corsica, over a challenging 45-degree-grade tree-covered mountain side, the Pleiades 2-m-grid-posting digital surface model described the topography with a median error of -4.75m +/-2.59m (NMAD). A planimetric bias between both datasets was found to be about 7m to the South. This planimetric misregistration, though well within Pleiades specifications, partly explains the dramatic effect on elevation difference. In the Redmond area (eastern Oregon), a very gentle desert landscape, elevation differences also contained a vertical median bias of -4.02m+/-1.22m (NMAD). Though here, sub-pixel planimetric registration between stereo DSM and lidar coverage was enforced. This real-size experiment hints that sub-meter accuracy for 2-m-grid-posting DSM is an achievable goal when combining stereoimaging and lidar.
Small-Grid Dithers for the JWST Coronagraphs
NASA Technical Reports Server (NTRS)
Lajoie, Charles-Philippe; Soummer, Remi; Pueyo, Laurent; Hines, Dean C.; Nelan, Edmund P.; Perrin, Marshall; Clampin, Mark; Isaacs, John C.
2016-01-01
We discuss new results of coronagraphic simulations demonstrating a novel mode for JWST that utilizes sub-pixel dithered reference images, called Small-Grid Dithers, to optimize coronagraphic PSF subtraction. These sub-pixel dithers are executed with the Fine Steering Mirror under fine guidance, are accurate to approx.2-3 milliarcseconds (1-s/axis), and provide ample speckle diversity to reconstruct an optimized synthetic reference PSF using LOCI or KLIP. We also discuss the performance gains of Small-Grid Dithers compared to the standard undithered scenario, and show potential contrast gain factors for the NIRCam and MIRI coronagraphs ranging from 2 to more than 10, respectively.
Wind-sea surface temperature-sea ice relationship in the Chukchi-Beaufort Seas during autumn
NASA Astrophysics Data System (ADS)
Zhang, Jing; Stegall, Steve T.; Zhang, Xiangdong
2018-03-01
Dramatic climate changes, especially the largest sea ice retreat during September and October, in the Chukchi-Beaufort Seas could be a consequence of, and further enhance, complex air-ice-sea interactions. To detect these interaction signals, statistical relationships between surface wind speed, sea surface temperature (SST), and sea ice concentration (SIC) were analyzed. The results show a negative correlation between wind speed and SIC. The relationships between wind speed and SST are complicated by the presence of sea ice, with a negative correlation over open water but a positive correlation in sea ice dominated areas. The examination of spatial structures indicates that wind speed tends to increase when approaching the ice edge from open water and the area fully covered by sea ice. The anomalous downward radiation and thermal advection, as well as their regional distribution, play important roles in shaping these relationships, though wind-driven sub-grid scale boundary layer processes may also have contributions. Considering the feedback loop involved in the wind-SST-SIC relationships, climate model experiments would be required to further untangle the underlying complex physical processes.
A principle of economy predicts the functional architecture of grid cells.
Wei, Xue-Xin; Prentice, Jason; Balasubramanian, Vijay
2015-09-03
Grid cells in the brain respond when an animal occupies a periodic lattice of 'grid fields' during navigation. Grids are organized in modules with different periodicity. We propose that the grid system implements a hierarchical code for space that economizes the number of neurons required to encode location with a given resolution across a range equal to the largest period. This theory predicts that (i) grid fields should lie on a triangular lattice, (ii) grid scales should follow a geometric progression, (iii) the ratio between adjacent grid scales should be √e for idealized neurons, and lie between 1.4 and 1.7 for realistic neurons, (iv) the scale ratio should vary modestly within and between animals. These results explain the measured grid structure in rodents. We also predict optimal organization in one and three dimensions, the number of modules, and, with added assumptions, the ratio between grid periods and field widths.
Quality Assurance Framework for Mini-Grids
DOE Office of Scientific and Technical Information (OSTI.GOV)
Baring-Gould, Ian; Burman, Kari; Singh, Mohit
Providing clean and affordable energy services to the more than 1 billion people globally who lack access to electricity is a critical driver for poverty reduction, economic development, improved health, and social outcomes. More than 84% of populations without electricity are located in rural areas where traditional grid extension may not be cost-effective; therefore, distributed energy solutions such as mini-grids are critical. To address some of the root challenges of providing safe, quality, and financially viable mini-grid power systems to remote customers, the U.S. Department of Energy (DOE) teamed with the National Renewable Energy Laboratory (NREL) to develop a Qualitymore » Assurance Framework (QAF) for isolated mini-grids. The QAF for mini-grids aims to address some root challenges of providing safe, quality, and affordable power to remote customers via financially viable mini-grids through two key components: (1) Levels of service: Defines a standard set of tiers of end-user service and links them to technical parameters of power quality, power availability, and power reliability. These levels of service span the entire energy ladder, from basic energy service to high-quality, high-reliability, and high-availability service (often considered 'grid parity'); (2) Accountability and performance reporting framework: Provides a clear process of validating power delivery by providing trusted information to customers, funders, and/or regulators. The performance reporting protocol can also serve as a robust monitoring and evaluation tool for mini-grid operators and funding organizations. The QAF will provide a flexible alternative to rigid top-down standards for mini-grids in energy access contexts, outlining tiers of end-user service and linking them to relevant technical parameters. In addition, data generated through implementation of the QAF will provide the foundation for comparisons across projects, assessment of impacts, and greater confidence that will drive investment and scale-up in this sector. The QAF implementation process also defines a set of implementation guidelines that help the deployment of mini-grids on a regional or national scale, helping to insure successful rapid deployment of these relatively new remote energy options. Note that the QAF is technology agnostic, addressing both alternating current (AC) and direct current (DC) mini-grids, and is also applicable to renewable, fossil-fuel, and hybrid systems.« less
NASA Astrophysics Data System (ADS)
Voisin, N.; Kintner-Meyer, M.; Skaggs, R.; Xie, Y.; Wu, D.; Nguyen, T. B.; Fu, T.; Zhou, T.
2016-12-01
Heat waves and droughts are projected to be more frequent and intense. We have seen in the past the effects of each of those extreme climate events on electricity demand and constrained electricity generation, challenging power system operations. Our aim here is to understand the compounding effects under historical conditions. We present a benchmark of Western US grid performance under 55 years of historical climate, and including droughts, using 2010-level of water demand and water management infrastructure, and 2010-level of electricity grid infrastructure and operations. We leverage CMIP5 historical hydrology simulations and force a large scale river routing- reservoir model with 2010-level sectoral water demands. The regulated flow at each water-dependent generating plants is processed to adjust water-dependent electricity generation parameterization in a production cost model, that represents 2010-level power system operations with hourly energy demand of 2010. The resulting benchmark includes a risk distribution of several grid performance metrics (unserved energy, production cost, carbon emission) as a function of inter-annual variability in regional water availability and predictability using large scale climate oscillations. In the second part of the presentation, we describe an approach to map historical heat waves onto this benchmark grid performance using a building energy demand model. The impact of the heat waves, combined with the impact of droughts, is explored at multiple scales to understand the compounding effects. Vulnerabilities of the power generation and transmission systems are highlighted to guide future adaptation.
Gridded uncertainty in fossil fuel carbon dioxide emission maps, a CDIAC example
Andres, Robert J.; Boden, Thomas A.; Higdon, David M.
2016-12-05
Due to a current lack of physical measurements at appropriate spatial and temporal scales, all current global maps and distributions of fossil fuel carbon dioxide (FFCO2) emissions use one or more proxies to distribute those emissions. These proxies and distribution schemes introduce additional uncertainty into these maps. This paper examines the uncertainty associated with the magnitude of gridded FFCO2 emissions. This uncertainty is gridded at the same spatial and temporal scales as the mass magnitude maps. This gridded uncertainty includes uncertainty contributions from the spatial, temporal, proxy, and magnitude components used to create the magnitude map of FFCO2 emissions. Throughoutmore » this process, when assumptions had to be made or expert judgment employed, the general tendency in most cases was toward overestimating or increasing the magnitude of uncertainty. The results of the uncertainty analysis reveal a range of 4–190 %, with an average of 120 % (2 σ) for populated and FFCO2-emitting grid spaces over annual timescales. This paper also describes a methodological change specific to the creation of the Carbon Dioxide Information Analysis Center (CDIAC) FFCO2 emission maps: the change from a temporally fixed population proxy to a temporally varying population proxy.« less
Gridded uncertainty in fossil fuel carbon dioxide emission maps, a CDIAC example
DOE Office of Scientific and Technical Information (OSTI.GOV)
Andres, Robert J.; Boden, Thomas A.; Higdon, David M.
Due to a current lack of physical measurements at appropriate spatial and temporal scales, all current global maps and distributions of fossil fuel carbon dioxide (FFCO2) emissions use one or more proxies to distribute those emissions. These proxies and distribution schemes introduce additional uncertainty into these maps. This paper examines the uncertainty associated with the magnitude of gridded FFCO2 emissions. This uncertainty is gridded at the same spatial and temporal scales as the mass magnitude maps. This gridded uncertainty includes uncertainty contributions from the spatial, temporal, proxy, and magnitude components used to create the magnitude map of FFCO2 emissions. Throughoutmore » this process, when assumptions had to be made or expert judgment employed, the general tendency in most cases was toward overestimating or increasing the magnitude of uncertainty. The results of the uncertainty analysis reveal a range of 4–190 %, with an average of 120 % (2 σ) for populated and FFCO2-emitting grid spaces over annual timescales. This paper also describes a methodological change specific to the creation of the Carbon Dioxide Information Analysis Center (CDIAC) FFCO2 emission maps: the change from a temporally fixed population proxy to a temporally varying population proxy.« less
Gridded uncertainty in fossil fuel carbon dioxide emission maps, a CDIAC example
NASA Astrophysics Data System (ADS)
Andres, Robert J.; Boden, Thomas A.; Higdon, David M.
2016-12-01
Due to a current lack of physical measurements at appropriate spatial and temporal scales, all current global maps and distributions of fossil fuel carbon dioxide (FFCO2) emissions use one or more proxies to distribute those emissions. These proxies and distribution schemes introduce additional uncertainty into these maps. This paper examines the uncertainty associated with the magnitude of gridded FFCO2 emissions. This uncertainty is gridded at the same spatial and temporal scales as the mass magnitude maps. This gridded uncertainty includes uncertainty contributions from the spatial, temporal, proxy, and magnitude components used to create the magnitude map of FFCO2 emissions. Throughout this process, when assumptions had to be made or expert judgment employed, the general tendency in most cases was toward overestimating or increasing the magnitude of uncertainty. The results of the uncertainty analysis reveal a range of 4-190 %, with an average of 120 % (2σ) for populated and FFCO2-emitting grid spaces over annual timescales. This paper also describes a methodological change specific to the creation of the Carbon Dioxide Information Analysis Center (CDIAC) FFCO2 emission maps: the change from a temporally fixed population proxy to a temporally varying population proxy.
NASA Astrophysics Data System (ADS)
Liu, Jia; Liu, Longli; Xue, Yong; Dong, Jing; Hu, Yingcui; Hill, Richard; Guang, Jie; Li, Chi
2017-01-01
Workflow for remote sensing quantitative retrieval is the ;bridge; between Grid services and Grid-enabled application of remote sensing quantitative retrieval. Workflow averts low-level implementation details of the Grid and hence enables users to focus on higher levels of application. The workflow for remote sensing quantitative retrieval plays an important role in remote sensing Grid and Cloud computing services, which can support the modelling, construction and implementation of large-scale complicated applications of remote sensing science. The validation of workflow is important in order to support the large-scale sophisticated scientific computation processes with enhanced performance and to minimize potential waste of time and resources. To research the semantic correctness of user-defined workflows, in this paper, we propose a workflow validation method based on tacit knowledge research in the remote sensing domain. We first discuss the remote sensing model and metadata. Through detailed analysis, we then discuss the method of extracting the domain tacit knowledge and expressing the knowledge with ontology. Additionally, we construct the domain ontology with Protégé. Through our experimental study, we verify the validity of this method in two ways, namely data source consistency error validation and parameters matching error validation.
NASA Astrophysics Data System (ADS)
Hazenberg, P.; Broxton, P. D.; Brunke, M.; Gochis, D.; Niu, G. Y.; Pelletier, J. D.; Troch, P. A. A.; Zeng, X.
2015-12-01
The terrestrial hydrological system, including surface and subsurface water, is an essential component of the Earth's climate system. Over the past few decades, land surface modelers have built one-dimensional (1D) models resolving the vertical flow of water through the soil column for use in Earth system models (ESMs). These models generally have a relatively coarse model grid size (~25-100 km) and only account for sub-grid lateral hydrological variations using simple parameterization schemes. At the same time, hydrologists have developed detailed high-resolution (~0.1-10 km grid size) three dimensional (3D) models and showed the importance of accounting for the vertical and lateral redistribution of surface and subsurface water on soil moisture, the surface energy balance and ecosystem dynamics on these smaller scales. However, computational constraints have limited the implementation of the high-resolution models for continental and global scale applications. The current work presents a hybrid-3D hydrological approach is presented, where the 1D vertical soil column model (available in many ESMs) is coupled with a high-resolution lateral flow model (h2D) to simulate subsurface flow and overland flow. H2D accounts for both local-scale hillslope and regional-scale unconfined aquifer responses (i.e. riparian zone and wetlands). This approach was shown to give comparable results as those obtained by an explicit 3D Richards model for the subsurface, but improves runtime efficiency considerably. The h3D approach is implemented for the Delaware river basin, where Noah-MP land surface model (LSM) is used to calculated vertical energy and water exchanges with the atmosphere using a 10km grid resolution. Noah-MP was coupled within the WRF-Hydro infrastructure with the lateral 1km grid resolution h2D model, for which the average depth-to-bedrock, hillslope width function and soil parameters were estimated from digital datasets. The ability of this h3D approach to simulate the hydrological dynamics of the Delaware River basin will be assessed by comparing the model results (both hydrological performance and numerical efficiency) with the standard setup of the NOAH-MP model and a high-resolution (1km) version of NOAH-MP, which also explicitly accounts for lateral subsurface and overland flow.
Simulating multi-scale oceanic processes around Taiwan on unstructured grids
NASA Astrophysics Data System (ADS)
Yu, Hao-Cheng; Zhang, Yinglong J.; Yu, Jason C. S.; Terng, C.; Sun, Weiling; Ye, Fei; Wang, Harry V.; Wang, Zhengui; Huang, Hai
2017-11-01
We validate a 3D unstructured-grid (UG) model for simulating multi-scale processes as occurred in Northwestern Pacific around Taiwan using recently developed new techniques (Zhang et al., Ocean Modeling, 102, 64-81, 2016) that require no bathymetry smoothing even for this region with prevalent steep bottom slopes and many islands. The focus is on short-term forecast for several months instead of long-term variability. Compared with satellite products, the errors for the simulated Sea-surface Height (SSH) and Sea-surface Temperature (SST) are similar to a reference data-assimilated global model. In the nearshore region, comparison with 34 tide gauges located around Taiwan indicates an average RMSE of 13 cm for the tidal elevation. The average RMSE for SST at 6 coastal buoys is 1.2 °C. The mean transport and eddy kinetic energy compare reasonably with previously published values and the reference model used to provide boundary and initial conditions. The model suggests ∼2-day interruption of Kuroshio east of Taiwan during a typhoon period. The effect of tidal mixing is shown to be significant nearshore. The multi-scale model is easily extendable to target regions of interest due to its UG framework and a flexible vertical gridding system, which is shown to be superior to terrain-following coordinates.
NASA Astrophysics Data System (ADS)
Tran, Quoc Quan; Willems, Patrick; Pannemans, Bart; Blanckaert, Joris; Pereira, Fernando; Nossent, Jiri; Cauwenberghs, Kris; Vansteenkiste, Thomas
2015-04-01
Based on an international literature review on model structures of existing rainfall-runoff and hydrological models, a generalized model structure is proposed. It consists of different types of meteorological components, storage components, splitting components and routing components. They can be spatially organized in a lumped way, or on a grid, spatially interlinked by source-to-sink or grid-to-grid (cell-to-cell) routing. The grid size of the model can be chosen depending on the application. The user can select/change the spatial resolution depending on the needs and/or the evaluation of the accuracy of the model results, or use different spatial resolutions in parallel for different applications. Major research questions addressed during the study are: How can we assure consistent results of the model at any spatial detail? How can we avoid strong or sudden changes in model parameters and corresponding simulation results, when one moves from one level of spatial detail to another? How can we limit the problem of overparameterization/equifinality when we move from the lumped model to the spatially distributed model? The proposed approach is a step-wise one, where first the lumped conceptual model is calibrated using a systematic, data-based approach, followed by a disaggregation step where the lumped parameters are disaggregated based on spatial catchment characteristics (topography, land use, soil characteristics). In this way, disaggregation can be done down to any spatial scale, and consistently among scales. Only few additional calibration parameters are introduced to scale the absolute spatial differences in model parameters, but keeping the relative differences as obtained from the spatial catchment characteristics. After calibration of the spatial model, the accuracies of the lumped and spatial models were compared for peak, low and cumulative runoff total and sub-flows (at downstream and internal gauging stations). For the distributed models, additional validation on spatial results was done for the groundwater head values at observation wells. To ensure that the lumped model can produce results as accurate as the spatially distributed models or close regardless to the number of parameters and implemented physical processes, it was checked whether the structure of the lumped models had to be adjusted. The concept has been implemented in a PCRaster - Python platform and tested for two Belgian case studies (catchments of the rivers Dijle and Grote Nete). So far, use is made of existing model structures (NAM, PDM, VHM and HBV). Acknowledgement: These results were obtained within the scope of research activities for the Flemish Environment Agency (VMM) - division Operational Water Management on "Next Generation hydrological modeling", in cooperation with IMDC consultants, and for Flanders Hydraulics Research (Waterbouwkundig Laboratorium) on "Effect of climate change on the hydrological regime of navigable watercourses in Belgium".
Tariff Considerations for Micro-Grids in Sub-Saharan Africa
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reber, Timothy J.; Booth, Samuel S.; Cutler, Dylan S.
This report examines some of the key drivers and considerations policymakers and decision makers face when deciding if and how to regulate electricity tariffs for micro-grids. Presenting a range of tariff options, from mandating some variety of national (uniform) tariff to allowing micro-grid developers and operators to set fully cost-reflective tariffs, it examines various benefits and drawbacks of each. In addition, the report and explores various types of cross-subsidies and other transitional forms of regulation that may offer a regulatory middle ground that can help balance the often competing goals of providing price control on electricity service in the namemore » of social good while still providing a means for investors to ensure high enough returns on their investment to attract the necessary capital financing to the market. Using the REopt tool developed by the U.S. Department of Energy's National Renewable Energy Laboratory to inform their study, the authors modeled a few representative micro-grid systems and the resultant levelized cost of electricity, lending context and scale to the consideration of these tariff questions. This simple analysis provides an estimate of the gap between current tariff regimes and the tariffs that would be necessary for developers to recover costs and attract investment, offering further insight into the potential scale of subsidies or other grants that may be required to enable micro-grid development under current regulatory structures. It explores potential options for addressing this gap while trying to balance This report examines some of the key drivers and considerations policymakers and decision makers face when deciding if and how to regulate electricity tariffs for micro-grids. Presenting a range of tariff options, from mandating some variety of national (uniform) tariff to allowing micro-grid developers and operators to set fully cost-reflective tariffs, it examines various benefits and drawbacks of each. In addition, the report and explores various types of cross-subsidies and other transitional forms of regulation that may offer a regulatory middle ground that can help balance the often competing goals of providing price control on electricity service in the name of social good while still providing a means for investors to ensure high enough returns on their investment to attract the necessary capital financing to the market. Using the REopt tool developed by the U.S. Department of Energy's National Renewable Energy Laboratory to inform their study, the authors modeled a few representative micro-grid systems and the resultant levelized cost of electricity, lending context and scale to the consideration of these tariff questions. This simple analysis provides an estimate of the gap between current tariff regimes and the tariffs that would be necessary for developers to recover costs and attract investment, offering further insight into the potential scale of subsidies or other grants that may be required to enable micro-grid development under current regulatory structures. It explores potential options for addressing this gap while trying to balance stakeholder needs, from subsidized national tariffs to lightly regulated cost-reflective tariffs to more of a compromise approach, such as different standards of regulation based on the size of a micro-grid.takeholder needs, from subsidized national tariffs to lightly regulated cost-reflective tariffs to more of a compromise approach, such as different standards of regulation based on the size of a micro-grid.« less
NASA Astrophysics Data System (ADS)
Reid, T. D.; Essery, R.; Rutter, N.; Huntley, B.; Baxter, R.; Holden, R.; King, M.; Hancock, S.; Carle, J.
2012-12-01
Boreal forests exert a strong influence on weather and climate by modifying the surface energy and radiation balance. However, global climate and numerical weather prediction models use forest parameter values from simple look-up tables or maps that are derived from limited satellite data, on large grid scales. In reality, Arctic landscapes are inherently heterogeneous, with highly variable land cover types and structures on a variety of spatial scales. There is value in collecting detailed field data for different areas of vegetation cover, to assess the accuracy of large-scale assumptions. To address these issues, a consortium of researchers funded by the UK's Natural Environment Research Council have collected extensive data on radiation, meteorology, snow cover and canopy structure at two contrasting Arctic forest sites. The chosen study sites were an area of boreal birch forest near Abisko, Sweden in March/April 2011 and mixed conifer forest at Sodankylä, Finland in March/April 2012. At both sites, arrays comprising ten shortwave pyranometers and four longwave pyrgeometers were deployed for periods of up to 50 days, under forest plots of varying canopy structures and densities. In addition, downwelling longwave irradiance and global and diffuse shortwave irradiances were recorded at nearby open sites representing the top-of-canopy conditions. Meteorological data were recorded at all sub-canopy and open sites using automatic weather stations. Over the same periods, tree skin temperatures were measured on selected trees using contact thermocouples, infrared thermocouples and thermal imagery. Canopy structure was accurately quantified through manual surveys, extensive hemispherical photography and terrestrial laser scans of every study plot. Sub-canopy snow depth and snow water equivalent were measured on fine-scale grids at each study plot. Regular site maintenance ensured a high quality dataset covering the important Arctic spring period. The data have several applications, for example in forest ecology, canopy radiative transfer models, snow hydrological modelling, and land surface schemes, for a variety of canopy types from sparse, leafless birch to dense pine and spruce. The work also allows the comparison of modern, highly detailed methods such as laser scanning and thermal imagery with older, well-established data collection methods. By combining these data with airborne and satellite remote sensing data, snow-vegetation-atmosphere interactions could be estimated over a wide area of the heterogeneous boreal landscape. This could improve estimates of crucial parameters such as land surface albedo on the grid scales required for global or regional weather and climate models.
Golze, Dorothea; Iannuzzi, Marcella; Hutter, Jürg
2017-05-09
A local resolution-of-the-identity (LRI) approach is introduced in combination with the Gaussian and plane waves (GPW) scheme to enable large-scale Kohn-Sham density functional theory calculations. In GPW, the computational bottleneck is typically the description of the total charge density on real-space grids. Introducing the LRI approximation, the linear scaling of the GPW approach with respect to system size is retained, while the prefactor for the grid operations is reduced. The density fitting is an O(N) scaling process implemented by approximating the atomic pair densities by an expansion in one-center fit functions. The computational cost for the grid-based operations becomes negligible in LRIGPW. The self-consistent field iteration is up to 30 times faster for periodic systems dependent on the symmetry of the simulation cell and on the density of grid points. However, due to the overhead introduced by the local density fitting, single point calculations and complete molecular dynamics steps, including the calculation of the forces, are effectively accelerated by up to a factor of ∼10. The accuracy of LRIGPW is assessed for different systems and properties, showing that total energies, reaction energies, intramolecular and intermolecular structure parameters are well reproduced. LRIGPW yields also high quality results for extended condensed phase systems such as liquid water, ice XV, and molecular crystals.
NASA Astrophysics Data System (ADS)
Chapelier, Jean-Baptiste; Wasistho, Bono; Scalo, Carlo
2017-11-01
A new approach to Large-Eddy Simulation (LES) is introduced, where subgrid-scale (SGS) dissipation is applied proportionally to the degree of local spectral broadening, hence mitigated in regions dominated by large-scale vortical motion. The proposed CvP-LES methodology is based on the evaluation of the ratio of the test-filtered to resolved (or grid-filtered) enstrophy: σ = ξ ∧ / ξ . Values of σ = 1 indicate low sub-test-filter turbulent activity, justifying local deactivation of any subgrid-scale model. Values of σ < 1 span conditions ranging from incipient spectral broadening σ <= 1 , to equilibrium turbulence σ =σeq < 1 , where σeq is solely as a function of the test-to-grid filter-width ratio Δ ∧ / Δ , derived assuming a Kolmogorov's spectrum. Eddy viscosity is fully restored for σ <=σeq . The proposed approach removes unnecessary SGS dissipation, can be applied to any eddy-viscosity model, is algorithmically simple and computationally inexpensive. A CvP-LES of a pair of unstable helical vortices, representative of rotor-blade wake dynamics, show the ability of the method to sort the coherent motion from the small-scale dynamics. This work is funded by subcontract KSC-17-001 between Purdue University and Kord Technologies, Inc (Huntsville), under the US Navy Contract N68335-17-C-0159 STTR-Phase II, Purdue Proposal No. 00065007, Topic N15A-T002.
NASA Astrophysics Data System (ADS)
Castiglioni, Giacomo
Flows over airfoils and blades in rotating machinery, for unmanned and micro-aerial vehicles, wind turbines, and propellers consist of a laminar boundary layer near the leading edge that is often followed by a laminar separation bubble and transition to turbulence further downstream. Typical Reynolds averaged Navier-Stokes turbulence models are inadequate for such flows. Direct numerical simulation is the most reliable, but is also the most computationally expensive alternative. This work assesses the capability of immersed boundary methods and large eddy simulations to reduce the computational requirements for such flows and still provide high quality results. Two-dimensional and three-dimensional simulations of a laminar separation bubble on a NACA-0012 airfoil at Rec = 5x104 and at 5° of incidence have been performed with an immersed boundary code and a commercial code using body fitted grids. Several sub-grid scale models have been implemented in both codes and their performance evaluated. For the two-dimensional simulations with the immersed boundary method the results show good agreement with the direct numerical simulation benchmark data for the pressure coefficient Cp and the friction coefficient Cf, but only when using dissipative numerical schemes. There is evidence that this behavior can be attributed to the ability of dissipative schemes to damp numerical noise coming from the immersed boundary. For the three-dimensional simulations the results show a good prediction of the separation point, but an inaccurate prediction of the reattachment point unless full direct numerical simulation resolution is used. The commercial code shows good agreement with the direct numerical simulation benchmark data in both two and three-dimensional simulations, but the presence of significant, unquantified numerical dissipation prevents a conclusive assessment of the actual prediction capabilities of very coarse large eddy simulations with low order schemes in general cases. Additionally, a two-dimensional sweep of angles of attack from 0° to 5° is performed showing a qualitative prediction of the jump in lift and drag coefficients due to the appearance of the laminar separation bubble. The numerical dissipation inhibits the predictive capabilities of large eddy simulations whenever it is of the same order of magnitude or larger than the sub-grid scale dissipation. The need to estimate the numerical dissipation is most pressing for low-order methods employed by commercial computational fluid dynamics codes. Following the recent work of Schranner et al., the equations and procedure for estimating the numerical dissipation rate and the numerical viscosity in a commercial code are presented. The method allows for the computation of the numerical dissipation rate and numerical viscosity in the physical space for arbitrary sub-domains in a self-consistent way, using only information provided by the code in question. The method is first tested for a three-dimensional Taylor-Green vortex flow in a simple cubic domain and compared with benchmark results obtained using an accurate, incompressible spectral solver. Afterwards the same procedure is applied for the first time to a realistic flow configuration, specifically to the above discussed laminar separation bubble flow over a NACA 0012 airfoil. The method appears to be quite robust and its application reveals that for the code and the flow in question the numerical dissipation can be significantly larger than the viscous dissipation or the dissipation of the classical Smagorinsky sub-grid scale model, confirming the previously qualitative finding.
Thermo-Gas-Dynamic Model of Afterburning in Explosions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuhl, A L; Ferguson, R E; Bell, J B
2003-07-27
A theoretical model of afterburning in explosions created by turbulent mixing of the detonation products from fuel-rich charges with air is described. It contains three key elements: (i) a thermodynamic-equilibrium description of the fluids (fuel, air, and products), (ii) a multi-component gas-dynamic treatment of the flow field, and (iii) a sub-grid model of molecular processes of mixing, combustion and equilibration.
Organic light-emitting diodes using novel embedded al gird transparent electrodes
NASA Astrophysics Data System (ADS)
Peng, Cuiyun; Chen, Changbo; Guo, Kunping; Tian, Zhenghao; Zhu, Wenqing; Xu, Tao; Wei, Bin
2017-03-01
This work demonstrates a novel transparent electrode using embedded Al grids fabricated by a simple and cost-effective approach using photolithography and wet etching. The optical and electrical properties of Al grids versus grid geometry have been systematically investigated, it was found that Al grids exhibited a low sheet resistance of 70 Ω □-1 and a light transmission of 69% at 550 nm with advantages in terms of processing conditions and material cost as well as potential to large scale fabrication. Indium Tin Oxide-free green organic light-emitting diodes (OLED) based on Al grids transparent electrodes was demonstrated, yielding a power efficiency >15 lm W-1 and current efficiency >39 cd A-1 at a brightness of 2396 cd m-2. Furthermore, a reduced efficiency roll-off and higher brightness have been achieved compared with ITO-base device.
Andres, R. J. [CDIAC; Boden, T. A. [CDIAC
2016-01-01
The annual, gridded fossil-fuel CO2 emissions uncertainty estimates from 1950-2013 provided in this database are derived from time series of global, regional, and national fossil-fuel CO2 emissions (Boden et al. 2016). Andres et al. (2016) describes the basic methodology in estimating the uncertainty in the (gridded fossil fuel data product ). This uncertainty is gridded at the same spatial and temporal scales as the mass magnitude maps. This gridded uncertainty includes uncertainty contributions from the spatial, temporal, proxy, and magnitude components used to create the magnitude map of FFCO2 emissions. Throughout this process, when assumptions had to be made or expert judgment employed, the general tendency in most cases was toward overestimating or increasing the magnitude of uncertainty.
Andres, J.A. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Boden, T.A. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)
2016-01-01
The monthly, gridded fossil-fuel CO2 emissions uncertainty estimates from 1950-2013 provided in this database are derived from time series of global, regional, and national fossil-fuel CO2 emissions (Boden et al. 2016). Andres et al. (2016) describes the basic methodology in estimating the uncertainty in the (gridded fossil fuel data product ). This uncertainty is gridded at the same spatial and temporal scales as the mass magnitude maps. This gridded uncertainty includes uncertainty contributions from the spatial, temporal, proxy, and magnitude components used to create the magnitude map of FFCO2 emissions. Throughout this process, when assumptions had to be made or expert judgment employed, the general tendency in most cases was toward overestimating or increasing the magnitude of uncertainty.
Global hydrodynamic modelling of flood inundation in continental rivers: How can we achieve it?
NASA Astrophysics Data System (ADS)
Yamazaki, D.
2016-12-01
Global-scale modelling of river hydrodynamics is essential for understanding global hydrological cycle, and is also required in interdisciplinary research fields . Global river models have been developed continuously for more than two decades, but modelling river flow at a global scale is still a challenging topic because surface water movement in continental rivers is a multi-spatial-scale phenomena. We have to consider the basin-wide water balance (>1000km scale), while hydrodynamics in river channels and floodplains is regulated by much smaller-scale topography (<100m scale). For example, heavy precipitation in upstream regions may later cause flooding in farthest downstream reaches. In order to realistically simulate the timing and amplitude of flood wave propagation for a long distance, consideration of detailed local topography is unavoidable. I have developed the global hydrodynamic model CaMa-Flood to overcome this scale-discrepancy of continental river flow. The CaMa-Flood divides river basins into multiple "unit-catchments", and assumes the water level is uniform within each unit-catchment. One unit-catchment is assigned to each grid-box defined at the typical spatial resolution of global climate models (10 100 km scale). Adopting a uniform water level in a >10km river segment seems to be a big assumption, but it is actually a good approximation for hydrodynamic modelling of continental rivers. The number of grid points required for global hydrodynamic simulations is largely reduced by this "unit-catchment assumption". Alternative to calculating 2-dimensional floodplain flows as in regional flood models, the CaMa-Flood treats floodplain inundation in a unit-catchment as a sub-grid physics. The water level and inundated area in each unit-catchment are diagnosed from water volume using topography parameters derived from high-resolution digital elevation models. Thus, the CaMa-Flood is at least 1000 times computationally more efficient compared to regional flood inundation models while the reality of simulated flood dynamics is kept. I will explain in detail how the CaMa-Flood model has been constructed from high-resolution topography datasets, and how the model can be used for various interdisciplinary applications.
A first large-scale flood inundation forecasting model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schumann, Guy J-P; Neal, Jeffrey C.; Voisin, Nathalie
2013-11-04
At present continental to global scale flood forecasting focusses on predicting at a point discharge, with little attention to the detail and accuracy of local scale inundation predictions. Yet, inundation is actually the variable of interest and all flood impacts are inherently local in nature. This paper proposes a first large scale flood inundation ensemble forecasting model that uses best available data and modeling approaches in data scarce areas and at continental scales. The model was built for the Lower Zambezi River in southeast Africa to demonstrate current flood inundation forecasting capabilities in large data-scarce regions. The inundation model domainmore » has a surface area of approximately 170k km2. ECMWF meteorological data were used to force the VIC (Variable Infiltration Capacity) macro-scale hydrological model which simulated and routed daily flows to the input boundary locations of the 2-D hydrodynamic model. Efficient hydrodynamic modeling over large areas still requires model grid resolutions that are typically larger than the width of many river channels that play a key a role in flood wave propagation. We therefore employed a novel sub-grid channel scheme to describe the river network in detail whilst at the same time representing the floodplain at an appropriate and efficient scale. The modeling system was first calibrated using water levels on the main channel from the ICESat (Ice, Cloud, and land Elevation Satellite) laser altimeter and then applied to predict the February 2007 Mozambique floods. Model evaluation showed that simulated flood edge cells were within a distance of about 1 km (one model resolution) compared to an observed flood edge of the event. Our study highlights that physically plausible parameter values and satisfactory performance can be achieved at spatial scales ranging from tens to several hundreds of thousands of km2 and at model grid resolutions up to several km2. However, initial model test runs in forecast mode revealed that it is crucial to account for basin-wide hydrological response time when assessing lead time performances notwithstanding structural limitations in the hydrological model and possibly large inaccuracies in precipitation data.« less
Some effects of horizontal discretization on linear baroclinic and symmetric instabilities
NASA Astrophysics Data System (ADS)
Barham, William; Bachman, Scott; Grooms, Ian
2018-05-01
The effects of horizontal discretization on linear baroclinic and symmetric instabilities are investigated by analyzing the behavior of the hydrostatic Eady problem in ocean models on the B and C grids. On the C grid a spurious baroclinic instability appears at small wavelengths. This instability does not disappear as the grid scale decreases; instead, it simply moves to smaller horizontal scales. The peak growth rate of the spurious instability is independent of the grid scale as the latter decreases. It is equal to cf /√{Ri} where Ri is the balanced Richardson number, f is the Coriolis parameter, and c is a nondimensional constant that depends on the Richardson number. As the Richardson number increases c increases towards an upper bound of approximately 1/2; for large Richardson numbers the spurious instability is faster than the Eady instability. To suppress the spurious instability it is recommended to use fourth-order centered tracer advection along with biharmonic viscosity and diffusion with coefficients (Δx) 4 f /(32√{Ri}) or larger where Δx is the grid scale. On the B grid, the growth rates of baroclinic and symmetric instabilities are too small, and converge upwards towards the correct values as the grid scale decreases; no spurious instabilities are observed. In B grid models at eddy-permitting resolution, the reduced growth rate of baroclinic instability may contribute to partially-resolved eddies being too weak. On the C grid the growth rate of symmetric instability is better (larger) than on the B grid, and converges upwards towards the correct value as the grid scale decreases.
Scaling of Energy Deposition in Fast Ignition Targets
NASA Astrophysics Data System (ADS)
Campbell, R. B.; Welch, Dale
2005-10-01
We examine the scaling to ignition of the energy deposition of laser generated electrons in compressed fast ignition cores. Relevant cores have densities of several hundred g/cm^3, with a few keV initial temperature. As the laser intensities increase approaching ignition systems, on the order of a few 10^21W/cm^2, the hot electron energies expected to approach 100MeV[1]. Most certainly anomalous processes must play a role in the energy transfer, but the exact nature of these processes, as well as a practical way to model them, remain open issues. Traditional PIC explicit methods are limited to low densities on current and anticipated computing platforms, so the study of relevant parameter ranges has received so far little attention. We use LSP[2] to examine a relativistic electron beam (presumed generated from a laser plasma interaction) of legislated energy and angular distribution is injected into a 3D block of compressed DT. Collective effects will determine the stopping, most likely driven by magnetic field filamentation. The scaling of the stopping as a function of block density and temperature, as well as hot electron current and laser intensity is presented. Sub-grid models may be profitably used and degenerate effects included in the solution of this problem. Sandia is operated by Sandia Corporation, for the USDOE. [1] A. Pukhov, et. al., Phys. Plas. 6, p2847 (1999) [2] D. R. Welch et al., Comput. Phys.Commun. 164, p183 (2004).
Cloud microphysics modification with an online coupled COSMO-MUSCAT regional model
NASA Astrophysics Data System (ADS)
Sudhakar, D.; Quaas, J.; Wolke, R.; Stoll, J.; Muehlbauer, A. D.; Tegen, I.
2015-12-01
Abstract: The quantification of clouds, aerosols, and aerosol-cloud interactions in models, continues to be a challenge (IPCC, 2013). In this scenario two-moment bulk microphysical scheme is used to understand the aerosol-cloud interactions in the regional model COSMO (Consortium for Small Scale Modeling). The two-moment scheme in COSMO has been especially designed to represent aerosol effects on the microphysics of mixed-phase clouds (Seifert et al., 2006). To improve the model predictability, the radiation scheme has been coupled with two-moment microphysical scheme. Further, the cloud microphysics parameterization has been modified via coupling COSMO with MUSCAT (MultiScale Chemistry Aerosol Transport model, Wolke et al., 2004). In this study, we will be discussing the initial result from the online-coupled COSMO-MUSCAT model system with modified two-moment parameterization scheme along with COSP (CFMIP Observational Simulator Package) satellite simulator. This online coupled model system aims to improve the sub-grid scale process in the regional weather prediction scenario. The constant aerosol concentration used in the Seifert and Beheng, (2006) parameterizations in COSMO model has been replaced by aerosol concentration derived from MUSCAT model. The cloud microphysical process from the modified two-moment scheme is compared with stand-alone COSMO model. To validate the robustness of the model simulation, the coupled model system is integrated with COSP satellite simulator (Muhlbauer et al., 2012). Further, the simulations are compared with MODIS (Moderate Resolution Imaging Spectroradiometer) and ISCCP (International Satellite Cloud Climatology Project) satellite products.
A Molecular Dynamics Simulation of the Turbulent Couette Minimal Flow Unit
NASA Astrophysics Data System (ADS)
Smith, Edward
2016-11-01
What happens to turbulent motions below the Kolmogorov length scale? In order to explore this question, a 300 million molecule Molecular Dynamics (MD) simulation is presented for the minimal Couette channel in which turbulence can be sustained. The regeneration cycle and turbulent statistics show excellent agreement to continuum based computational fluid dynamics (CFD) at Re=400. As MD requires only Newton's laws and a form of inter-molecular potential, it captures a much greater range of phenomena without requiring the assumptions of Newton's law of viscosity, thermodynamic equilibrium, fluid isotropy or the limitation of grid resolution. The fundamental nature of MD means it is uniquely placed to explore the nature of turbulent transport. A number of unique insights from MD are presented, including energy budgets, sub-grid turbulent energy spectra, probability density functions, Lagrangian statistics and fluid wall interactions. EPSRC Post Doctoral Prize Fellowship.
A NEW THREE-DIMENSIONAL SOLAR WIND MODEL IN SPHERICAL COORDINATES WITH A SIX-COMPONENT GRID
DOE Office of Scientific and Technical Information (OSTI.GOV)
Feng, Xueshang; Zhang, Man; Zhou, Yufen, E-mail: fengx@spaceweather.ac.cn
In this paper, we introduce a new three-dimensional magnetohydrodynamics numerical model to simulate the steady state ambient solar wind from the solar surface to 215 R {sub s} or beyond, and the model adopts a splitting finite-volume scheme based on a six-component grid system in spherical coordinates. By splitting the magnetohydrodynamics equations into a fluid part and a magnetic part, a finite volume method can be used for the fluid part and a constrained-transport method able to maintain the divergence-free constraint on the magnetic field can be used for the magnetic induction part. This new second-order model in space andmore » time is validated when modeling the large-scale structure of the solar wind. The numerical results for Carrington rotation 2064 show its ability to produce structured solar wind in agreement with observations.« less
Wall modeled LES of wind turbine wakes with geometrical effects
NASA Astrophysics Data System (ADS)
Bricteux, Laurent; Benard, Pierre; Zeoli, Stephanie; Moureau, Vincent; Lartigue, Ghislain; Vire, Axelle
2017-11-01
This study focuses on prediction of wind turbine wakes when geometrical effects such as nacelle, tower, and built environment, are taken into account. The aim is to demonstrate the ability of a high order unstructured solver called YALES2 to perform wall modeled LES of wind turbine wake turbulence. The wind turbine rotor is modeled using an Actuator Line Model (ALM) while the geometrical details are explicitly meshed thanks to the use of an unstructured grid. As high Reynolds number flows are considered, sub-grid scale models as well as wall modeling are required. The first test case investigated concerns a wind turbine flow located in a wind tunnel that allows to validate the proposed methodology using experimental data. The second test case concerns the simulation of a wind turbine wake in a complex environment (e.g. a Building) using realistic turbulent inflow conditions.
AgMIP: Next Generation Models and Assessments
NASA Astrophysics Data System (ADS)
Rosenzweig, C.
2014-12-01
Next steps in developing next-generation crop models fall into several categories: significant improvements in simulation of important crop processes and responses to stress; extension from simplified crop models to complex cropping systems models; and scaling up from site-based models to landscape, national, continental, and global scales. Crop processes that require major leaps in understanding and simulation in order to narrow uncertainties around how crops will respond to changing atmospheric conditions include genetics; carbon, temperature, water, and nitrogen; ozone; and nutrition. The field of crop modeling has been built on a single crop-by-crop approach. It is now time to create a new paradigm, moving from 'crop' to 'cropping system.' A first step is to set up the simulation technology so that modelers can rapidly incorporate multiple crops within fields, and multiple crops over time. Then the response of these more complex cropping systems can be tested under different sustainable intensification management strategies utilizing the updated simulation environments. Model improvements for diseases, pests, and weeds include developing process-based models for important diseases, frameworks for coupling air-borne diseases to crop models, gathering significantly more data on crop impacts, and enabling the evaluation of pest management strategies. Most smallholder farming in the world involves integrated crop-livestock systems that cannot be represented by crop modeling alone. Thus, next-generation cropping system models need to include key linkages to livestock. Livestock linkages to be incorporated include growth and productivity models for grasslands and rangelands as well as the usual annual crops. There are several approaches for scaling up, including use of gridded models and development of simpler quasi-empirical models for landscape-scale analysis. On the assessment side, AgMIP is leading a community process for coordinated contributions to IPCC AR6 that involves the key modeling groups from around the world including North America, Europe, South America, Sub-Saharan Africa, South Asia, East Asia, and Australia and Oceania. This community process will lead to mutually agreed protocols for coordinated global and regional assessments.
Microphysics in the Multi-Scale Modeling Systems with Unified Physics
NASA Technical Reports Server (NTRS)
Tao, Wei-Kuo; Chern, J.; Lamg, S.; Matsui, T.; Shen, B.; Zeng, X.; Shi, R.
2011-01-01
In recent years, exponentially increasing computer power has extended Cloud Resolving Model (CRM) integrations from hours to months, the number of computational grid points from less than a thousand to close to ten million. Three-dimensional models are now more prevalent. Much attention is devoted to precipitating cloud systems where the crucial 1-km scales are resolved in horizontal domains as large as 10,000 km in two-dimensions, and 1,000 x 1,000 km2 in three-dimensions. Cloud resolving models now provide statistical information useful for developing more realistic physically based parameterizations for climate models and numerical weather prediction models. It is also expected that NWP and mesoscale model can be run in grid size similar to cloud resolving model through nesting technique. Recently, a multi-scale modeling system with unified physics was developed at NASA Goddard. It consists of (l) a cloud-resolving model (Goddard Cumulus Ensemble model, GCE model), (2) a regional scale model (a NASA unified weather research and forecast, WRF), (3) a coupled CRM and global model (Goddard Multi-scale Modeling Framework, MMF), and (4) a land modeling system. The same microphysical processes, long and short wave radiative transfer and land processes and the explicit cloud-radiation, and cloud-surface interactive processes are applied in this multi-scale modeling system. This modeling system has been coupled with a multi-satellite simulator to use NASA high-resolution satellite data to identify the strengths and weaknesses of cloud and precipitation processes simulated by the model. In this talk, the microphysics developments of the multi-scale modeling system will be presented. In particular, the results from using multi-scale modeling system to study the heavy precipitation processes will be presented.
User's Guide - WRF Lightning Assimilation
This document describes how to run WRF with the lightning assimilation technique described in Heath et al. (2016). The assimilation method uses gridded lightning data to trigger and suppress sub-grid deep convection in Kain-Fritsch.
NASA Technical Reports Server (NTRS)
Montesano, P. M.; Cook, B. D.; Sun, G.; Simard, M.; Zhang, Z.; Nelson, R. F.; Ranson, K. J.; Lutchke, S.; Blair, J. B.
2012-01-01
The synergistic use of active and passive remote sensing (i.e., data fusion) demonstrates the ability of spaceborne light detection and ranging (LiDAR), synthetic aperture radar (SAR) and multispectral imagery for achieving the accuracy requirements of a global forest biomass mapping mission. This data fusion approach also provides a means to extend 3D information from discrete spaceborne LiDAR measurements of forest structure across scales much larger than that of the LiDAR footprint. For estimating biomass, these measurements mix a number of errors including those associated with LiDAR footprint sampling over regional - global extents. A general framework for mapping above ground live forest biomass (AGB) with a data fusion approach is presented and verified using data from NASA field campaigns near Howland, ME, USA, to assess AGB and LiDAR sampling errors across a regionally representative landscape. We combined SAR and Landsat-derived optical (passive optical) image data to identify forest patches, and used image and simulated spaceborne LiDAR data to compute AGB and estimate LiDAR sampling error for forest patches and 100m, 250m, 500m, and 1km grid cells. Forest patches were delineated with Landsat-derived data and airborne SAR imagery, and simulated spaceborne LiDAR (SSL) data were derived from orbit and cloud cover simulations and airborne data from NASA's Laser Vegetation Imaging Sensor (L VIS). At both the patch and grid scales, we evaluated differences in AGB estimation and sampling error from the combined use of LiDAR with both SAR and passive optical and with either SAR or passive optical alone. This data fusion approach demonstrates that incorporating forest patches into the AGB mapping framework can provide sub-grid forest information for coarser grid-level AGB reporting, and that combining simulated spaceborne LiDAR with SAR and passive optical data are most useful for estimating AGB when measurements from LiDAR are limited because they minimized forest AGB sampling errors by 15 - 38%. Furthermore, spaceborne global scale accuracy requirements were achieved. At least 80% of the grid cells at 100m, 250m, 500m, and 1km grid levels met AGB density accuracy requirements using a combination of passive optical and SAR along with machine learning methods to predict vegetation structure metrics for forested areas without LiDAR samples. Finally, using either passive optical or SAR, accuracy requirements were met at the 500m and 250m grid level, respectively.
NASA Astrophysics Data System (ADS)
Roman, C. N.; Reves-Sohn, R.; Singh, H.; Humphris, S.
2005-12-01
The spatial resolution of microbathymetry maps created using robotic vehicles such as ROVs, AUVs and manned submersibles in the deep ocean is currently limited by the accuracy of the vehicle navigation data. Errors in the vehicle position estimate commonly exceed the ranging errors of the acoustic mapping sensor itself, which creates inconsistency in the map making process and produces artifacts that lower resolution and distort map integrity. We present a methodology for producing self-consistent maps and improving vehicle position estimation by exploiting accurate local navigation and utilizing terrain relative measurements. The complete map is broken down into individual "sub-maps'', which are generated using short term Doppler based navigation. The sub-maps are pairwise registered to constrain the vehicle position estimates by matching terrain that has been imaged multiple times. This procedure is implemented using a delayed state Kalman filter to incorporate the sub-map registrations as relative position measurements between previously visited vehicle locations. Archiving of previous positions in a filter state vector allows for continual adjustment of the sub-map locations. The terrain registration is accomplished using a two dimensional correlation and a six degree of freedom point cloud alignment method tailored to bathymetric data. This registration procedure is applicable to fully 3 dimensional complex underwater environments. The complete bathymetric map is then created from the union of all sub-maps that have been aligned in a consistent manner. The method is applied to an SM2000 multibeam survey of the TAG hydrothermal structure on the Mid-Atlantic Ridge at 26(°)N using the Jason II ROV. The survey included numerous crossing tracklines designed to test this algorithm, and the final gridded bathymetry data is sub-meter accurate. The high-resolution map has allowed for the identification of previously unrecognized fracture patterns associated with flow focusing at TAG, as well as imaging of fine-scale features such as individual sulfide talus blocks and ODP re-entry cones.
Synergies Between Grace and Regional Atmospheric Modeling Efforts
NASA Astrophysics Data System (ADS)
Kusche, J.; Springer, A.; Ohlwein, C.; Hartung, K.; Longuevergne, L.; Kollet, S. J.; Keune, J.; Dobslaw, H.; Forootan, E.; Eicker, A.
2014-12-01
In the meteorological community, efforts converge towards implementation of high-resolution (< 12km) data-assimilating regional climate modelling/monitoring systems based on numerical weather prediction (NWP) cores. This is driven by requirements of improving process understanding, better representation of land surface interactions, atmospheric convection, orographic effects, and better forecasting on shorter timescales. This is relevant for the GRACE community since (1) these models may provide improved atmospheric mass separation / de-aliasing and smaller topography-induced errors, compared to global (ECMWF-Op, ERA-Interim) data, (2) they inherit high temporal resolution from NWP models, (3) parallel efforts towards improving the land surface component and coupling groundwater models; this may provide realistic hydrological mass estimates with sub-diurnal resolution, (4) parallel efforts towards re-analyses, with the aim of providing consistent time series. (5) On the other hand, GRACE can help validating models and aids in the identification of processes needing improvement. A coupled atmosphere - land surface - groundwater modelling system is currently being implemented for the European CORDEX region at 12.5 km resolution, based on the TerrSysMP platform (COSMO-EU NWP, CLM land surface and ParFlow groundwater models). We report results from Springer et al. (J. Hydromet., accept.) on validating the water cycle in COSMO-EU using GRACE and precipitation, evapotranspiration and runoff data; confirming that the model does favorably at representing observations. We show that after GRACE-derived bias correction, basin-average hydrological conditions prior to 2002 can be reconstructed better than before. Next, comparing GRACE with CLM forced by EURO-CORDEX simulations allows identifying processes needing improvement in the model. Finally, we compare COSMO-EU atmospheric pressure, a proxy for mass corrections in satellite gravimetry, with ERA-Interim over Europe at timescales shorter/longer than 1 month, and spatial scales below/above ERA resolution. We find differences between regional and global model more pronounced at high frequencies, with magnitude at sub-grid scale and larger scale corresponding to 1-3 hPa (1-3 cm EWH); relevant for the assessment of post-GRACE concepts.
DEM Based Modeling: Grid or TIN? The Answer Depends
NASA Astrophysics Data System (ADS)
Ogden, F. L.; Moreno, H. A.
2015-12-01
The availability of petascale supercomputing power has enabled process-based hydrological simulations on large watersheds and two-way coupling with mesoscale atmospheric models. Of course with increasing watershed scale come corresponding increases in watershed complexity, including wide ranging water management infrastructure and objectives, and ever increasing demands for forcing data. Simulations of large watersheds using grid-based models apply a fixed resolution over the entire watershed. In large watersheds, this means an enormous number of grids, or coarsening of the grid resolution to reduce memory requirements. One alternative to grid-based methods is the triangular irregular network (TIN) approach. TINs provide the flexibility of variable resolution, which allows optimization of computational resources by providing high resolution where necessary and low resolution elsewhere. TINs also increase required effort in model setup, parameter estimation, and coupling with forcing data which are often gridded. This presentation discusses the costs and benefits of the use of TINs compared to grid-based methods, in the context of large watershed simulations within the traditional gridded WRF-HYDRO framework and the new TIN-based ADHydro high performance computing watershed simulator.
Schnek: A C++ library for the development of parallel simulation codes on regular grids
NASA Astrophysics Data System (ADS)
Schmitz, Holger
2018-05-01
A large number of algorithms across the field of computational physics are formulated on grids with a regular topology. We present Schnek, a library that enables fast development of parallel simulations on regular grids. Schnek contains a number of easy-to-use modules that greatly reduce the amount of administrative code for large-scale simulation codes. The library provides an interface for reading simulation setup files with a hierarchical structure. The structure of the setup file is translated into a hierarchy of simulation modules that the developer can specify. The reader parses and evaluates mathematical expressions and initialises variables or grid data. This enables developers to write modular and flexible simulation codes with minimal effort. Regular grids of arbitrary dimension are defined as well as mechanisms for defining physical domain sizes, grid staggering, and ghost cells on these grids. Ghost cells can be exchanged between neighbouring processes using MPI with a simple interface. The grid data can easily be written into HDF5 files using serial or parallel I/O.
INTRINSIC COLORS, TEMPERATURES, AND BOLOMETRIC CORRECTIONS OF PRE-MAIN-SEQUENCE STARS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pecaut, Mark J.; Mamajek, Eric E.
2013-09-01
We present an analysis of the intrinsic colors and temperatures of 5-30 Myr old pre-main-sequence (pre-MS) stars using the F0- through M9-type members of nearby, negligibly reddened groups: the η Cha cluster, the TW Hydra Association, the β Pic Moving Group, and the Tucana-Horologium Association. To check the consistency of spectral types from the literature, we estimate new spectral types for 52 nearby pre-MS stars with spectral types F3 through M4 using optical spectra taken with the SMARTS 1.5 m telescope. Combining these new types with published spectral types and photometry from the literature (Johnson-Cousins BVI{sub C} , 2MASS JHK{submore » S} and WISE W1, W2, W3, and W4), we derive a new empirical spectral type-color sequence for 5-30 Myr old pre-MS stars. Colors for pre-MS stars match dwarf colors for some spectral types and colors, but for other spectral types and colors, deviations can exceed 0.3 mag. We estimate effective temperatures (T {sub eff}) and bolometric corrections (BCs) for our pre-MS star sample through comparing their photometry to synthetic photometry generated using the BT-Settl grid of model atmosphere spectra. We derive a new T {sub eff} and BC scale for pre-MS stars, which should be a more appropriate match for T Tauri stars than often-adopted dwarf star scales. While our new T {sub eff} scale for pre-MS stars is within ≅100 K of dwarfs at a given spectral type for stars« less
NASA Astrophysics Data System (ADS)
Machalek, P.; Kim, S. M.; Berry, R. D.; Liang, A.; Small, T.; Brevdo, E.; Kuznetsova, A.
2012-12-01
We describe how the Climate Corporation uses Python and Clojure, a language impleneted on top of Java, to generate climatological forecasts for precipitation based on the Advanced Hydrologic Prediction Service (AHPS) radar based daily precipitation measurements. A 2-year-long forecasts is generated on each of the ~650,000 CONUS land based 4-km AHPS grids by constructing 10,000 ensembles sampled from a 30-year reconstructed AHPS history for each grid. The spatial and temporal correlations between neighboring AHPS grids and the sampling of the analogues are handled by Python. The parallelization for all the 650,000 CONUS stations is further achieved by utilizing the MAP-REDUCE framework (http://code.google.com/edu/parallel/mapreduce-tutorial.html). Each full scale computational run requires hundreds of nodes with up to 8 processors each on the Amazon Elastic MapReduce (http://aws.amazon.com/elasticmapreduce/) distributed computing service resulting in 3 terabyte datasets. We further describe how we have productionalized a monthly run of the simulations process at full scale of the 4km AHPS grids and how the resultant terabyte sized datasets are handled.
Ensuring Safety of Navigation: A Three-Tiered Approach
NASA Astrophysics Data System (ADS)
Johnson, S. D.; Thompson, M.; Brazier, D.
2014-12-01
The primary responsibility of the Hydrographic Department at the Naval Oceanographic Office (NAVOCEANO) is to support US Navy surface and sub-surface Safety of Navigation (SoN) requirements. These requirements are interpreted, surveys are conducted, and accurate products are compiled and archived for future exploitation. For a number of years NAVOCEANO has employed a two-tiered data-basing structure to support SoN. The first tier (Data Warehouse, or DWH) provides access to the full-resolution sonar and lidar data. DWH preserves the original data such that any scale product can be built. The second tier (Digital Bathymetric Database - Variable resolution, or DBDB-V) served as the final archive for SoN chart scale, gridded products compiled from source bathymetry. DBDB-V has been incorporated into numerous DoD tactical decision aids and serves as the foundation bathymetry for ocean modeling. With the evolution of higher density survey systems and the addition of high-resolution gridded bathymetry product requirements, a two-tiered model did not provide an efficient solution for SoN. The two-tiered approach required scientists to exploit full-resolution data in order to build any higher resolution product. A new perspective on the archival and exploitation of source data was required. This new perspective has taken the form of a third tier, the Navigation Surface Database (NSDB). NSDB is an SQLite relational database populated with International Hydrographic Organization (IHO), S-102 compliant Bathymetric Attributed Grids (BAGs). BAGs archived within NSDB are developed at the highest resolution that the collection sensor system can support and contain nodal estimates for depth, uncertainty, separation values and metadata. Gridded surface analysis efforts culminate in the generation of the source resolution BAG files and their storage within NSDB. Exploitation of these resources eliminates the time and effort needed to re-grid and re-analyze native source file formats.
Emissions of methane in Europe inferred by total column measurements
NASA Astrophysics Data System (ADS)
Wunch, D.; Deutscher, N. M.; Hase, F.; Notholt, J.; Sussmann, R.; Toon, G. C.; Warneke, T.
2017-12-01
Atmospheric total column measurements have been used to infer emissions of methane in urban centres around the world. These measurements have been shown to be useful for verifying city-scale bottom-up inventories, and they can provide both timely and sub-annual emission information. We will present our analysis of atmospheric total column measurements of methane and carbon monoxide to infer annual and seasonal regional emissions of methane within Europe using five long-running atmospheric observatories. These observatories are part of the Total Carbon Column Observing Network, part of a global network that has been carefully designed to measure these gases on a consistent scale. Our inferred emissions will then be used to evaluate gridded emissions inventories in the region.
NASA Astrophysics Data System (ADS)
Beckmann, R. S.; Slyz, A.; Devriendt, J.
2018-07-01
Whilst in galaxy-size simulations, supermassive black holes (SMBHs) are entirely handled by sub-grid algorithms, computational power now allows the accretion radius of such objects to be resolved in smaller scale simulations. In this paper, we investigate the impact of resolution on two commonly used SMBH sub-grid algorithms; the Bondi-Hoyle-Lyttleton (BHL) formula for accretion on to a point mass, and the related estimate of the drag force exerted on to a point mass by a gaseous medium. We find that when the accretion region around the black hole scales with resolution, and the BHL formula is evaluated using local mass-averaged quantities, the accretion algorithm smoothly transitions from the analytic BHL formula (at low resolution) to a supply-limited accretion scheme (at high resolution). However, when a similar procedure is employed to estimate the drag force, it can lead to significant errors in its magnitude, and/or apply this force in the wrong direction in highly resolved simulations. At high Mach numbers and for small accretors, we also find evidence of the advective-acoustic instability operating in the adiabatic case, and of an instability developing around the wake's stagnation point in the quasi-isothermal case. Moreover, at very high resolution, and Mach numbers above M_∞ ≥ 3, the flow behind the accretion bow shock becomes entirely dominated by these instabilities. As a result, accretion rates on to the black hole drop by about an order of magnitude in the adiabatic case, compared to the analytic BHL formula.
NASA Astrophysics Data System (ADS)
Beckmann, R. S.; Slyz, A.; Devriendt, J.
2018-04-01
Whilst in galaxy-size simulations, supermassive black holes (SMBH) are entirely handled by sub-grid algorithms, computational power now allows the accretion radius of such objects to be resolved in smaller scale simulations. In this paper, we investigate the impact of resolution on two commonly used SMBH sub-grid algorithms; the Bondi-Hoyle-Lyttleton (BHL) formula for accretion onto a point mass, and the related estimate of the drag force exerted onto a point mass by a gaseous medium. We find that when the accretion region around the black hole scales with resolution, and the BHL formula is evaluated using local mass-averaged quantities, the accretion algorithm smoothly transitions from the analytic BHL formula (at low resolution) to a supply limited accretion (SLA) scheme (at high resolution). However, when a similar procedure is employed to estimate the drag force it can lead to significant errors in its magnitude, and/or apply this force in the wrong direction in highly resolved simulations. At high Mach numbers and for small accretors, we also find evidence of the advective-acoustic instability operating in the adiabatic case, and of an instability developing around the wake's stagnation point in the quasi-isothermal case. Moreover, at very high resolution, and Mach numbers above M_∞ ≥ 3, the flow behind the accretion bow shock becomes entirely dominated by these instabilities. As a result, accretion rates onto the black hole drop by about an order of magnitude in the adiabatic case, compared to the analytic BHL formula.
NASA Astrophysics Data System (ADS)
Kandel, D. D.; Western, A. W.; Grayson, R. B.
2004-12-01
Mismatches in scale between the fundamental processes, the model and supporting data are a major limitation in hydrologic modelling. Surface runoff generation via infiltration excess and the process of soil erosion are fundamentally short time-scale phenomena and their average behaviour is mostly determined by the short time-scale peak intensities of rainfall. Ideally, these processes should be simulated using time-steps of the order of minutes to appropriately resolve the effect of rainfall intensity variations. However, sub-daily data support is often inadequate and the processes are usually simulated by calibrating daily (or even coarser) time-step models. Generally process descriptions are not modified but rather effective parameter values are used to account for the effect of temporal lumping, assuming that the effect of the scale mismatch can be counterbalanced by tuning the parameter values at the model time-step of interest. Often this results in parameter values that are difficult to interpret physically. A similar approach is often taken spatially. This is problematic as these processes generally operate or interact non-linearly. This indicates a need for better techniques to simulate sub-daily processes using daily time-step models while still using widely available daily information. A new method applicable to many rainfall-runoff-erosion models is presented. The method is based on temporal scaling using statistical distributions of rainfall intensity to represent sub-daily intensity variations in a daily time-step model. This allows the effect of short time-scale nonlinear processes to be captured while modelling at a daily time-step, which is often attractive due to the wide availability of daily forcing data. The approach relies on characterising the rainfall intensity variation within a day using a cumulative distribution function (cdf). This cdf is then modified by various linear and nonlinear processes typically represented in hydrological and erosion models. The statistical description of sub-daily variability is thus propagated through the model, allowing the effects of variability to be captured in the simulations. This results in cdfs of various fluxes, the integration of which over a day gives respective daily totals. Using 42-plot-years of surface runoff and soil erosion data from field studies in different environments from Australia and Nepal, simulation results from this cdf approach are compared with the sub-hourly (2-minute for Nepal and 6-minute for Australia) and daily models having similar process descriptions. Significant improvements in the simulation of surface runoff and erosion are achieved, compared with a daily model that uses average daily rainfall intensities. The cdf model compares well with a sub-hourly time-step model. This suggests that the approach captures the important effects of sub-daily variability while utilizing commonly available daily information. It is also found that the model parameters are more robustly defined using the cdf approach compared with the effective values obtained at the daily scale. This suggests that the cdf approach may offer improved model transferability spatially (to other areas) and temporally (to other periods).
A principle of economy predicts the functional architecture of grid cells
Wei, Xue-Xin; Prentice, Jason; Balasubramanian, Vijay
2015-01-01
Grid cells in the brain respond when an animal occupies a periodic lattice of ‘grid fields’ during navigation. Grids are organized in modules with different periodicity. We propose that the grid system implements a hierarchical code for space that economizes the number of neurons required to encode location with a given resolution across a range equal to the largest period. This theory predicts that (i) grid fields should lie on a triangular lattice, (ii) grid scales should follow a geometric progression, (iii) the ratio between adjacent grid scales should be √e for idealized neurons, and lie between 1.4 and 1.7 for realistic neurons, (iv) the scale ratio should vary modestly within and between animals. These results explain the measured grid structure in rodents. We also predict optimal organization in one and three dimensions, the number of modules, and, with added assumptions, the ratio between grid periods and field widths. DOI: http://dx.doi.org/10.7554/eLife.08362.001 PMID:26335200
Metadata Creation Tool Content Template For Data Stewards
A space-time Bayesian fusion model (McMillan, Holland, Morara, and Feng, 2009) is used to provide daily, gridded predictive PM2.5 (daily average) and O3 (daily 8-hr maximum) surfaces for 2001-2005. The fusion model uses both air quality monitoring data from ...
Benefits of using enhanced air quality information in human health studies
The ability of four (4) enhancements of gridded PM2.5 concentrations derived from observations and air quality models to detect the relative risk of long-term exposure to PM2.5 are evaluated with a simulation study. The four enhancements include nearest-nei...
NASA Technical Reports Server (NTRS)
Pawson, Steven; Zhu, Z.; Ott, L. E.; Molod, A.; Duncan, B. N.; Nielsen, J. E.
2009-01-01
Sub-grid transport, by convection and turbulence, is known to play an important role in lofting pollutants from their source regions. Consequently, the long-range transport and climatology of simulated atmospheric composition are impacted. This study uses the Goddard Earth Observing System, Version 5 (GEOS-5) atmospheric model to study pollutant transport. The baseline model uses a Relaxed Arakawa-Schubert (RAS) scheme that represents convection through a sequence of linearly entraining cloud plumes characterized by unique detrainment levels. Thermodynamics, moisture and trace gases are transported in the same manner. Various approximate forms of trace-gas transport are implemented, in which the box-averaged cloud mass fluxes from RAS are used with different numerical approaches. Substantial impacts on forward-model simulations of CO (using a linearized chemistry) are evident. In particular, some aspects of simulations using a diffusive form of sub-grid transport bear more resemblance to space-biased CO observations than do the baseline simulations with RAS transport. Implications for transport in the real atmosphere will be discussed. Another issue of importance is that many adjoint/inversion computations use simplified representations of sub-grid transport that may be inconsistent with the forward models: implications will be discussed. Finally, simulations using a complex chemistry model in GEOS-5 (in place of the linearized CO model) are underway: noteworthy results from this simulation will be mentioned.
NASA Technical Reports Server (NTRS)
Elliott, Joshua; Muller, Christoff
2015-01-01
Climate change is a significant risk for agricultural production. Even under optimistic scenarios for climate mitigation action, present-day agricultural areas are likely to face significant increases in temperatures in the coming decades, in addition to changes in precipitation, cloud cover, and the frequency and duration of extreme heat, drought, and flood events (IPCC, 2013). These factors will affect the agricultural system at the global scale by impacting cultivation regimes, prices, trade, and food security (Nelson et al., 2014a). Global-scale evaluation of crop productivity is a major challenge for climate impact and adaptation assessment. Rigorous global assessments that are able to inform planning and policy will benefit from consistent use of models, input data, and assumptions across regions and time that use mutually agreed protocols designed by the modeling community. To ensure this consistency, large-scale assessments are typically performed on uniform spatial grids, with spatial resolution of typically 10 to 50 km, over specified time-periods. Many distinct crop models and model types have been applied on the global scale to assess productivity and climate impacts, often with very different results (Rosenzweig et al., 2014). These models are based to a large extent on field-scale crop process or ecosystems models and they typically require resolved data on weather, environmental, and farm management conditions that are lacking in many regions (Bondeau et al., 2007; Drewniak et al., 2013; Elliott et al., 2014b; Gueneau et al., 2012; Jones et al., 2003; Liu et al., 2007; M¨uller and Robertson, 2014; Van den Hoof et al., 2011;Waha et al., 2012; Xiong et al., 2014). Due to data limitations, the requirements of consistency, and the computational and practical limitations of running models on a large scale, a variety of simplifying assumptions must generally be made regarding prevailing management strategies on the grid scale in both the baseline and future periods. Implementation differences in these and other modeling choices contribute to significant variation among global-scale crop model assessments in addition to differences in crop model implementations that also cause large differences in site-specific crop modeling (Asseng et al., 2013; Bassu et al., 2014).
NASA Astrophysics Data System (ADS)
Huang, D.; Liu, Y.
2014-12-01
The effects of subgrid cloud variability on grid-average microphysical rates and radiative fluxes are examined by use of long-term retrieval products at the Tropical West Pacific (TWP), Southern Great Plains (SGP), and North Slope of Alaska (NSA) sites of the Department of Energy's Atmospheric Radiation Measurement (ARM) Program. Four commonly used distribution functions, the truncated Gaussian, Gamma, lognormal, and Weibull distributions, are constrained to have the same mean and standard deviation as observed cloud liquid water content. The PDFs are then used to upscale relevant physical processes to obtain grid-average process rates. It is found that the truncated Gaussian representation results in up to 30% mean bias in autoconversion rate whereas the mean bias for the lognormal representation is about 10%. The Gamma and Weibull distribution function performs the best for the grid-average autoconversion rate with the mean relative bias less than 5%. For radiative fluxes, the lognormal and truncated Gaussian representations perform better than the Gamma and Weibull representations. The results show that the optimal choice of subgrid cloud distribution function depends on the nonlinearity of the process of interest and thus there is no single distribution function that works best for all parameterizations. Examination of the scale (window size) dependence of the mean bias indicates that the bias in grid-average process rates monotonically increases with increasing window sizes, suggesting the increasing importance of subgrid variability with increasing grid sizes.
Ebner, Marc; Hameroff, Stuart
2011-01-01
Cognitive brain functions, for example, sensory perception, motor control and learning, are understood as computation by axonal-dendritic chemical synapses in networks of integrate-and-fire neurons. Cognitive brain functions may occur either consciously or nonconsciously (on “autopilot”). Conscious cognition is marked by gamma synchrony EEG, mediated largely by dendritic-dendritic gap junctions, sideways connections in input/integration layers. Gap-junction-connected neurons define a sub-network within a larger neural network. A theoretical model (the “conscious pilot”) suggests that as gap junctions open and close, a gamma-synchronized subnetwork, or zone moves through the brain as an executive agent, converting nonconscious “auto-pilot” cognition to consciousness, and enhancing computation by coherent processing and collective integration. In this study we implemented sideways “gap junctions” in a single-layer artificial neural network to perform figure/ground separation. The set of neurons connected through gap junctions form a reconfigurable resistive grid or sub-network zone. In the model, outgoing spikes are temporally integrated and spatially averaged using the fixed resistive grid set up by neurons of similar function which are connected through gap-junctions. This spatial average, essentially a feedback signal from the neuron's output, determines whether particular gap junctions between neurons will open or close. Neurons connected through open gap junctions synchronize their output spikes. We have tested our gap-junction-defined sub-network in a one-layer neural network on artificial retinal inputs using real-world images. Our system is able to perform figure/ground separation where the laterally connected sub-network of neurons represents a perceived object. Even though we only show results for visual stimuli, our approach should generalize to other modalities. The system demonstrates a moving sub-network zone of synchrony, within which the contents of perception are represented and contained. This mobile zone can be viewed as a model of the neural correlate of consciousness in the brain. PMID:22046178
Ebner, Marc; Hameroff, Stuart
2011-01-01
Cognitive brain functions, for example, sensory perception, motor control and learning, are understood as computation by axonal-dendritic chemical synapses in networks of integrate-and-fire neurons. Cognitive brain functions may occur either consciously or nonconsciously (on "autopilot"). Conscious cognition is marked by gamma synchrony EEG, mediated largely by dendritic-dendritic gap junctions, sideways connections in input/integration layers. Gap-junction-connected neurons define a sub-network within a larger neural network. A theoretical model (the "conscious pilot") suggests that as gap junctions open and close, a gamma-synchronized subnetwork, or zone moves through the brain as an executive agent, converting nonconscious "auto-pilot" cognition to consciousness, and enhancing computation by coherent processing and collective integration. In this study we implemented sideways "gap junctions" in a single-layer artificial neural network to perform figure/ground separation. The set of neurons connected through gap junctions form a reconfigurable resistive grid or sub-network zone. In the model, outgoing spikes are temporally integrated and spatially averaged using the fixed resistive grid set up by neurons of similar function which are connected through gap-junctions. This spatial average, essentially a feedback signal from the neuron's output, determines whether particular gap junctions between neurons will open or close. Neurons connected through open gap junctions synchronize their output spikes. We have tested our gap-junction-defined sub-network in a one-layer neural network on artificial retinal inputs using real-world images. Our system is able to perform figure/ground separation where the laterally connected sub-network of neurons represents a perceived object. Even though we only show results for visual stimuli, our approach should generalize to other modalities. The system demonstrates a moving sub-network zone of synchrony, within which the contents of perception are represented and contained. This mobile zone can be viewed as a model of the neural correlate of consciousness in the brain.
CIELO-A GIS integrated model for climatic and water balance simulation in islands environments
NASA Astrophysics Data System (ADS)
Azevedo, E. B.; Pereira, L. S.
2003-04-01
The model CIELO (acronym for "Clima Insular à Escala Local") is a physically based model that simulates the climatic variables in an island using data from a single synoptic reference meteorological station. The reference station "knows" its position in the orographic and dynamic regime context. The domain of computation is a GIS raster grid parameterised with a digital elevation model (DEM). The grid is oriented following the direction of the air masses circulation through a specific algorithm named rotational terrain model (RTM). The model consists of two main sub-models. One, relative to the advective component simulation, assumes the Foehn effect to reproduce the dynamic and thermodynamic processes occurring when an air mass moves through the island orographic obstacle. This makes possible to simulate the air temperature, air humidity, cloudiness and precipitation as influenced by the orography along the air displacement. The second concerns the radiative component as affected by the clouds of orographic origin and by the shadow produced by the relief. The initial state parameters are computed starting from the reference meteorological station across the DEM transept until the sea level at the windward side. Then, starting from the sea level, the model computes the local scale meteorological parameters according to the direction of the air displacement, which is adjusted with the RTM. The air pressure, temperature and humidity are directly calculated for each cell in the computational grid, while several algorithms are used to compute the cloudiness, net radiation, evapotranspiration, and precipitation. The model presented in this paper has been calibrated and validated using data from some meteorological stations and a larger number of rainfall stations located at various elevations in the Azores Islands.
Consequence modeling using the fire dynamics simulator.
Ryder, Noah L; Sutula, Jason A; Schemel, Christopher F; Hamer, Andrew J; Van Brunt, Vincent
2004-11-11
The use of Computational Fluid Dynamics (CFD) and in particular Large Eddy Simulation (LES) codes to model fires provides an efficient tool for the prediction of large-scale effects that include plume characteristics, combustion product dispersion, and heat effects to adjacent objects. This paper illustrates the strengths of the Fire Dynamics Simulator (FDS), an LES code developed by the National Institute of Standards and Technology (NIST), through several small and large-scale validation runs and process safety applications. The paper presents two fire experiments--a small room fire and a large (15 m diameter) pool fire. The model results are compared to experimental data and demonstrate good agreement between the models and data. The validation work is then extended to demonstrate applicability to process safety concerns by detailing a model of a tank farm fire and a model of the ignition of a gaseous fuel in a confined space. In this simulation, a room was filled with propane, given time to disperse, and was then ignited. The model yields accurate results of the dispersion of the gas throughout the space. This information can be used to determine flammability and explosive limits in a space and can be used in subsequent models to determine the pressure and temperature waves that would result from an explosion. The model dispersion results were compared to an experiment performed by Factory Mutual. Using the above examples, this paper will demonstrate that FDS is ideally suited to build realistic models of process geometries in which large scale explosion and fire failure risks can be evaluated with several distinct advantages over more traditional CFD codes. Namely transient solutions to fire and explosion growth can be produced with less sophisticated hardware (lower cost) than needed for traditional CFD codes (PC type computer verses UNIX workstation) and can be solved for longer time histories (on the order of hundreds of seconds of computed time) with minimal computer resources and length of model run. Additionally results that are produced can be analyzed, viewed, and tabulated during and following a model run within a PC environment. There are some tradeoffs, however, as rapid computations in PC's may require a sacrifice in the grid resolution or in the sub-grid modeling, depending on the size of the geometry modeled.
A Multi-scale Modeling System with Unified Physics to Study Precipitation Processes
NASA Astrophysics Data System (ADS)
Tao, W. K.
2017-12-01
In recent years, exponentially increasing computer power has extended Cloud Resolving Model (CRM) integrations from hours to months, the number of computational grid points from less than a thousand to close to ten million. Three-dimensional models are now more prevalent. Much attention is devoted to precipitating cloud systems where the crucial 1-km scales are resolved in horizontal domains as large as 10,000 km in two-dimensions, and 1,000 x 1,000 km2 in three-dimensions. Cloud resolving models now provide statistical information useful for developing more realistic physically based parameterizations for climate models and numerical weather prediction models. It is also expected that NWP and mesoscale model can be run in grid size similar to cloud resolving model through nesting technique. Recently, a multi-scale modeling system with unified physics was developed at NASA Goddard. It consists of (1) a cloud-resolving model (Goddard Cumulus Ensemble model, GCE model), (2) a regional scale model (a NASA unified weather research and forecast, WRF), and (3) a coupled CRM and global model (Goddard Multi-scale Modeling Framework, MMF). The same microphysical processes, long and short wave radiative transfer and land processes and the explicit cloud-radiation, and cloud-land surface interactive processes are applied in this multi-scale modeling system. This modeling system has been coupled with a multi-satellite simulator to use NASA high-resolution satellite data to identify the strengths and weaknesses of cloud and precipitation processes simulated by the model. In this talk, a review of developments and applications of the multi-scale modeling system will be presented. In particular, the results from using multi-scale modeling system to study the precipitation, processes and their sensitivity on model resolution and microphysics schemes will be presented. Also how to use of the multi-satellite simulator to improve precipitation processes will be discussed.
Using Multi-Scale Modeling Systems and Satellite Data to Study the Precipitation Processes
NASA Technical Reports Server (NTRS)
Tao, Wei-Kuo; Chern, J.; Lamg, S.; Matsui, T.; Shen, B.; Zeng, X.; Shi, R.
2011-01-01
In recent years, exponentially increasing computer power has extended Cloud Resolving Model (CRM) integrations from hours to months, the number of computational grid points from less than a thousand to close to ten million. Three-dimensional models are now more prevalent. Much attention is devoted to precipitating cloud systems where the crucial 1-km scales are resolved in horizontal domains as large as 10,000 km in two-dimensions, and 1,000 x 1,000 km2 in three-dimensions. Cloud resolving models now provide statistical information useful for developing more realistic physically based parameterizations for climate models and numerical weather prediction models. It is also expected that NWP and mesoscale model can be run in grid size similar to cloud resolving model through nesting technique. Recently, a multi-scale modeling system with unified physics was developed at NASA Goddard. It consists of (l) a cloud-resolving model (Goddard Cumulus Ensemble model, GCE model), (2) a regional scale model (a NASA unified weather research and forecast, WRF), (3) a coupled CRM and global model (Goddard Multi-scale Modeling Framework, MMF), and (4) a land modeling system. The same microphysical processes, long and short wave radiative transfer and land processes and the explicit cloud-radiation, and cloud-land surface interactive processes are applied in this multi-scale modeling system. This modeling system has been coupled with a multi-satellite simulator to use NASA high-resolution satellite data to identify the strengths and weaknesses of cloud and precipitation processes simulated by the model. In this talk, the recent developments and applications of the multi-scale modeling system will be presented. In particular, the results from using multi-scale modeling system to study the precipitating systems and hurricanes/typhoons will be presented. The high-resolution spatial and temporal visualization will be utilized to show the evolution of precipitation processes. Also how to use of the multi-satellite simulator tqimproy precipitation processes will be discussed.
IR-drop analysis for validating power grids and standard cell architectures in sub-10nm node designs
NASA Astrophysics Data System (ADS)
Ban, Yongchan; Wang, Chenchen; Zeng, Jia; Kye, Jongwook
2017-03-01
Since chip performance and power are highly dependent on the operating voltage, the robust power distribution network (PDN) is of utmost importance in designs to provide with the reliable voltage without voltage (IR)-drop. However, rapid increase of parasitic resistance and capacitance (RC) in interconnects makes IR-drop much worse with technology scaling. This paper shows various IR-drop analyses in sub 10nm designs. The major objectives are to validate standard cell architectures, where different sizes of power/ground and metal tracks are validated, and to validate PDN architecture, where types of power hook-up approaches are evaluated with IR-drop calculation. To estimate IR-drops in 10nm and below technologies, we first prepare physically routed designs given standard cell libraries, where we use open RISC RTL, synthesize the CPU, and apply placement & routing with process-design kits (PDK). Then, static and dynamic IR-drop flows are set up with commercial tools. Using the IR-drop flow, we compare standard cell architectures, and analysis impacts on performance, power, and area (PPA) with the previous technology-node designs. With this IR-drop flow, we can optimize the best PDN structure against IR-drops as well as types of standard cell library.
Basin-Scale Hydrologic Impacts of CO2 Storage: Regulatory and Capacity Implications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Birkholzer, J.T.; Zhou, Q.
Industrial-scale injection of CO{sub 2} into saline sedimentary basins will cause large-scale fluid pressurization and migration of native brines, which may affect valuable groundwater resources overlying the deep sequestration reservoirs. In this paper, we discuss how such basin-scale hydrologic impacts can (1) affect regulation of CO{sub 2} storage projects and (2) may reduce current storage capacity estimates. Our assessment arises from a hypothetical future carbon sequestration scenario in the Illinois Basin, which involves twenty individual CO{sub 2} storage projects in a core injection area suitable for long-term storage. Each project is assumed to inject five million tonnes of CO{sub 2}more » per year for 50 years. A regional-scale three-dimensional simulation model was developed for the Illinois Basin that captures both the local-scale CO{sub 2}-brine flow processes and the large-scale groundwater flow patterns in response to CO{sub 2} storage. The far-field pressure buildup predicted for this selected sequestration scenario suggests that (1) the area that needs to be characterized in a permitting process may comprise a very large region within the basin if reservoir pressurization is considered, and (2) permits cannot be granted on a single-site basis alone because the near- and far-field hydrologic response may be affected by interference between individual sites. Our results also support recent studies in that environmental concerns related to near-field and far-field pressure buildup may be a limiting factor on CO{sub 2} storage capacity. In other words, estimates of storage capacity, if solely based on the effective pore volume available for safe trapping of CO{sub 2}, may have to be revised based on assessments of pressure perturbations and their potential impact on caprock integrity and groundwater resources, respectively. We finally discuss some of the challenges in making reliable predictions of large-scale hydrologic impacts related to CO{sub 2} sequestration projects.« less
A hierarchical wavefront reconstruction algorithm for gradient sensors
NASA Astrophysics Data System (ADS)
Bharmal, Nazim; Bitenc, Urban; Basden, Alastair; Myers, Richard
2013-12-01
ELT-scale extreme adaptive optics systems will require new approaches tocompute the wavefront suitably quickly, when the computational burden ofapplying a MVM is no longer practical. An approach is demonstrated here whichis hierarchical in transforming wavefront slopes from a WFS into a wavefront,and then to actuator values. First, simple integration in 1D is used to create1D-wavefront estimates with unknown starting points at the edges of independentspatial domains. Second, these starting points are estimated globally. By thesestarting points are a sub-set of the overall grid where wavefront values are tobe estimated, sparse representations are produced and numerical complexity canbe chosen by the spacing of the starting point grid relative to the overallgrid. Using a combination of algebraic expressions, sparse representation, anda conjugate gradient solver, the number of non-parallelized operations forreconstruction on a 100x100 sub-aperture sized problem is ~600,000 or O(N^3/2),which is approximately the same as for each thread of a MVM solutionparallelized over 100 threads. To reduce the effects of noise propagationwithin each domain, a noise reduction algorithm can be applied which ensuresthe continuity of the wavefront. To apply this additional step has a cost of~1,200,000 operations. We conclude by briefly discussing how the final step ofconverting from wavefront to actuator values can be achieved.
GridPix detectors: Production and beam test results
NASA Astrophysics Data System (ADS)
Koppert, W. J. C.; van Bakel, N.; Bilevych, Y.; Colas, P.; Desch, K.; Fransen, M.; van der Graaf, H.; Hartjes, F.; Hessey, N. P.; Kaminski, J.; Schmitz, J.; Schön, R.; Zappon, F.
2013-12-01
The innovative GridPix detector is a Time Projection Chamber (TPC) that is read out with a Timepix-1 pixel chip. By using wafer post-processing techniques an aluminium grid is placed on top of the chip. When operated, the electric field between the grid and the chip is sufficient to create electron induced avalanches which are detected by the pixels. The time-to-digital converter (TDC) records the drift time enabling the reconstruction of high precision 3D track segments. Recently GridPixes were produced on full wafer scale, to meet the demand for more reliable and cheaper devices in large quantities. In a recent beam test the contribution of both diffusion and time walk to the spatial and angular resolutions of a GridPix detector with a 1.2 mm drift gap are studied in detail. In addition long term tests show that in a significant fraction of the chips the protection layer successfully quenches discharges, preventing harm to the chip.
Computational Investigations of Noise Suppression in Subsonic Round Jets
NASA Technical Reports Server (NTRS)
Pruett, C. David
1997-01-01
NASA Grant NAG1-1802, originally submitted in June 1996 as a two-year proposal, was awarded one-year's funding by NASA LaRC for the period 5 Oct., 1996, through 4 Oct., 1997. Because of the inavailability (from IT at NASA ARC) of sufficient supercomputer time in fiscal 1998 to complete the computational goals of the second year of the original proposal (estimated to be at least 400 Cray C-90 CPU hours), those goals have been appropriately amended, and a new proposal has been submitted to LaRC as a follow-on to NAG1-1802. The current report documents the activities and accomplishments on NAG1-1802 during the one-year period from 5 Oct., 1996, through 4 Oct., 1997. NASA Grant NAG1-1802, and its predecessor, NAG1-1772, have been directed toward adapting the numerical tool of Large-Eddy Simulation (LES) to aeroacoustic applications, with particular focus on noise suppression in subsonic round jets. In LES, the filtered Navier-Stokes equations are solved numerically on a relatively coarse computational grid. Residual stresses, generated by scales of motion too small to be resolved on the coarse grid, are modeled. Although most LES incorporate spatial filtering, time-domain filtering affords certain conceptual and computational advantages, particularly for aeroacoustic applications. Consequently, this work has focused on the development of SubGrid-Scale (SGS) models that incorporate time- domain filters. The author is unaware of any previous attempt at purely time-filtered LES; however, Aldama and Dakhoul and Bedford have considered approaches that combine both spatial and temporal filtering. In our view, filtering in both space and time is redundant, because removal of high frequencies effects the removal of small spatial scales and vice versa.
Aerosol-cloud interactions in a multi-scale modeling framework
NASA Astrophysics Data System (ADS)
Lin, G.; Ghan, S. J.
2017-12-01
Atmospheric aerosols play an important role in changing the Earth's climate through scattering/absorbing solar and terrestrial radiation and interacting with clouds. However, quantification of the aerosol effects remains one of the most uncertain aspects of current and future climate projection. Much of the uncertainty results from the multi-scale nature of aerosol-cloud interactions, which is very challenging to represent in traditional global climate models (GCMs). In contrast, the multi-scale modeling framework (MMF) provides a viable solution, which explicitly resolves the cloud/precipitation in the cloud resolved model (CRM) embedded in the GCM grid column. In the MMF version of community atmospheric model version 5 (CAM5), aerosol processes are treated with a parameterization, called the Explicit Clouds Parameterized Pollutants (ECPP). It uses the cloud/precipitation statistics derived from the CRM to treat the cloud processing of aerosols on the GCM grid. However, this treatment treats clouds on the CRM grid but aerosols on the GCM grid, which is inconsistent with the reality that cloud-aerosol interactions occur on the cloud scale. To overcome the limitation, here, we propose a new aerosol treatment in the MMF: Explicit Clouds Explicit Aerosols (ECEP), in which we resolve both clouds and aerosols explicitly on the CRM grid. We first applied the MMF with ECPP to the Accelerated Climate Modeling for Energy (ACME) model to have an MMF version of ACME. Further, we also developed an alternative version of ACME-MMF with ECEP. Based on these two models, we have conducted two simulations: one with the ECPP and the other with ECEP. Preliminary results showed that the ECEP simulations tend to predict higher aerosol concentrations than ECPP simulations, because of the more efficient vertical transport from the surface to the higher atmosphere but the less efficient wet removal. We also found that the cloud droplet number concentrations are also different between the two simulations due to the difference in the cloud droplet lifetime. Next, we will explore how the ECEP treatment affects the anthropogenic aerosol forcing, particularly the aerosol indirect forcing, by comparing present-day and pre-industrial simulations.
Considerations for fine hole patterning for the 7nm node
NASA Astrophysics Data System (ADS)
Yaegashi, Hidetami; Oyama, Kenichi; Hara, Arisa; Natori, Sakurako; Yamauchi, Shohei; Yamato, Masatoshi; Koike, Kyohei
2016-03-01
One of the practical candidates to produce 7nm node logic devices is to use the multiple patterning with 193-immersion exposure. For the multiple patterning, it is important to evaluate the relation between the number of mask layer and the minimum pitch systematically to judge the device manufacturability. Although the number of the time of patterning, namely LE(Litho-Etch) ^ x-time, and overlay steps have to be reduced, there are some challenges in miniaturization of hole size below 20nm. Various process fluctuations on contact hole have a direct impact on device performance. According to the technical trend, 12nm diameter hole on 30nm-pitch hole will be needed on 7nm node. Extreme ultraviolet lithography (EUV) and Directed self-assembly (DSA) are attracting considerable attention to obtain small feature size pattern, however, 193-immersion still has the potential to extend optical lithography cost-effectively for sub-7nm node. The objective of this work is to study the process variation challenges and resolution in post-processing for the CD-bias control to meet sub-20nm diameter contact hole. Another pattern modulation is also demonstrated during post-processing step for hole shrink. With the realization that pattern fidelity and pattern placement management will limit scaling long before devices and interconnects fail to perform intrinsically, the talk will also outline how circle edge roughness (CER) and Local-CD uniformity can correct efficiency. On the other hand, 1D Gridded-Design-Rules layout (1D layout) has simple rectangular shapes. Also, we have demonstrated CD-bias modification on short trench pattern to cut grating line for its fabrication.
Short-term Time Step Convergence in a Climate Model
Wan, Hui; Rasch, Philip J.; Taylor, Mark; ...
2015-02-11
A testing procedure is designed to assess the convergence property of a global climate model with respect to time step size, based on evaluation of the root-mean-square temperature difference at the end of very short (1 h) simulations with time step sizes ranging from 1 s to 1800 s. A set of validation tests conducted without sub-grid scale parameterizations confirmed that the method was able to correctly assess the convergence rate of the dynamical core under various configurations. The testing procedure was then applied to the full model, and revealed a slow convergence of order 0.4 in contrast to themore » expected first-order convergence. Sensitivity experiments showed without ambiguity that the time stepping errors in the model were dominated by those from the stratiform cloud parameterizations, in particular the cloud microphysics. This provides a clear guidance for future work on the design of more accurate numerical methods for time stepping and process coupling in the model.« less
Challenges in Modeling of the Global Atmosphere
NASA Astrophysics Data System (ADS)
Janjic, Zavisa; Djurdjevic, Vladimir; Vasic, Ratko; Black, Tom
2015-04-01
The massively parallel computer architectures require that some widely adopted modeling paradigms be reconsidered in order to utilize more productively the power of parallel processing. For high computational efficiency with distributed memory, each core should work on a small subdomain of the full integration domain, and exchange only few rows of halo data with the neighbouring cores. However, the described scenario implies that the discretization used in the model is horizontally local. The spherical geometry further complicates the problem. Various grid topologies will be discussed and examples will be shown. The latitude-longitude grid with local in space and explicit in time differencing has been an early choice and remained in use ever since. The problem with this method is that the grid size in the longitudinal direction tends to zero as the poles are approached. So, in addition to having unnecessarily high resolution near the poles, polar filtering has to be applied in order to use a time step of decent size. However, the polar filtering requires transpositions involving extra communications. The spectral transform method and the semi-implicit semi-Lagrangian schemes opened the way for a wide application of the spectral representation. With some variations, these techniques are used in most major centers. However, the horizontal non-locality is inherent to the spectral representation and implicit time differencing, which inhibits scaling on a large number of cores. In this respect the lat-lon grid with a fast Fourier transform represents a significant step in the right direction, particularly at high resolutions where the Legendre transforms become increasingly expensive. Other grids with reduced variability of grid distances such as various versions of the cubed sphere and the hexagonal/pentagonal ("soccer ball") grids were proposed almost fifty years ago. However, on these grids, large-scale (wavenumber 4 and 5) fictitious solutions ("grid imprinting") with significant amplitudes can develop. Due to their large scales, that are comparable to the scales of the dominant Rossby waves, such fictitious solutions are hard to identify and remove. Another new challenge on the global scale is that the limit of validity of the hydrostatic approximation is rapidly being approached. Having in mind the sensitivity of extended deterministic forecasts to small disturbances, we may need global non-hydrostatic models sooner than we think. The unified Non-hydrostatic Multi-scale Model (NMMB) that is being developed at the National Centers for Environmental Prediction (NCEP) as a part of the new NOAA Environmental Modeling System (NEMS) will be discussed as an example. The non-hydrostatic dynamics were designed in such a way as to avoid over-specification. The global version is run on the latitude-longitude grid, and the polar filter selectively slows down the waves that would otherwise be unstable. The model formulation has been successfully tested on various scales. A global forecasting system based on the NMMB has been run in order to test and tune the model. The skill of the medium range forecasts produced by the NMMB is comparable to that of other major medium range models. The computational efficiency of the global NMMB on parallel computers is good.
NASA Astrophysics Data System (ADS)
Abderrahim, Iheb
Wind power generation has grown strongly in the last decade. This results in the development of Wind Energy Conversion System WECS at the levels of modeling and electrical control. Modern WECS operate at varying wind speeds and are equipped with synchronous and asynchronous generators. Among these generators, the Doubly-Fed Induction Generator (DFIG) offers several advantages and capabilities of active and reactive power in four quadrants. WECS based DFIG also causes less conversion costs and minimum energy losses compared with a WECS based on a synchronous generator powered entirely by full scale of power converters. The connection of such a system to the electrical distribution network involves bidirectional operation of networks. This is clearly established in sub and super synchronous operating modes of DFIG. The grid provides the active power to the rotor of DFIG in sub synchronous operating mode and receives the active power of the rotor in super synchronous operating mode of DFIG. Energy quality is thus of major importance during the integration of wind power to the grid. Poor wave quality can affect network stability and could even cause major problems and consequences. This is even more critical where non-linear loads such as the switching power supplies and variable speed drives, are connected to the grid. The idea of this research work is how to mitigate the problems associated with the wave quality while ensuring better implementation of DFIG so that the whole of WECS remains insensitive to external disturbances and parametric variations. The Grid Side Converter (GSC) must be able to compensate harmonics, current unbalance and reactive power injected by a nonlinear three-phase unbalanced load connected to the grid. In addition to these innovative features to improve the conditions of operation of the grid, it provides also the power flow during different modes of operation of the DFIG. It is considered a simple, efficient and cost competitive solution by saving the use of other power equipment. At the same time, the energy efficiency of wind power conversion chain should be improved by extracting the MPPT. Searching allows us to select vector control and control in synchronous reference to achieve these objectives. WECS based DFIG is simulated in MATLAB SIMULINK in the presence of a non-linear balanced and unbalanced three-phase load.
NASA Astrophysics Data System (ADS)
Vionnet, Vincent; Six, Delphine; Auger, Ludovic; Lafaysse, Matthieu; Quéno, Louis; Réveillet, Marion; Dombrowski-Etchevers, Ingrid; Thibert, Emmanuel; Dumont, Marie
2017-04-01
Capturing spatial and temporal variabilities of meteorological conditions at fine scale is necessary for modelling snowpack and glacier winter mass balance in alpine terrain. In particular, precipitation amount and phase are strongly influenced by the complex topography. In this study, we assess the impact of three sub-kilometer precipitation datasets (rainfall and snowfall) on distributed simulations of snowpack and glacier winter mass balance with the detailed snowpack model Crocus for winter 2011-2012. The different precipitation datasets at 500-m grid spacing over part of the French Alps (200*200 km2 area) are coming either from (i) the SAFRAN precipitation analysis specially developed for alpine terrain, or from (ii) operational outputs of the atmospheric model AROME at 2.5-km grid spacing downscaled to 500 m with fixed lapse rate or from (iii) a version of the atmospheric model AROME at 500-m grid spacing. Others atmospherics forcings (air temperature and humidity, incoming longwave and shortwave radiation, wind speed) are taken from the AROME simulations at 500-m grid spacing. These atmospheric forcings are firstly compared against a network of automatic weather stations. Results are analysed with respect to station location (valley, mid- and high-altitude). The spatial pattern of seasonal snowfall and its dependency with elevation is then analysed for the different precipitation datasets. Large differences between SAFRAN and the two versions of AROME are found at high-altitude. Finally, results of Crocus snowpack simulations are evaluated against (i) punctual in-situ measurements of snow depth and snow water equivalent, and (ii) maps of snow covered areas retrieved from optical satellite data (MODIS). Measurements of winter accumulation of six glaciers of the French Alps are also used and provide very valuable information on precipitation at high-altitude where the conventional observation network is scarce. This study illustrates the potential and limitations of high-resolution atmospheric models to drive simulations of snowpack and glacier winter mass balance in alpine terrain.
NASA Astrophysics Data System (ADS)
Stern, M.; West, R.; Fourer, G.; Whalen, W.; Van Loo, M.; Duran, G.
1997-02-01
Utility Power Group has achieved a significant reduction in the installed cost of grid-connected PV systems. The two part technical approach focused on 1) The utilization of a large area factory assembled PV panel, and 2) The integration and packaging of all sub-array power conversion and control functions within a single factory produced enclosure. Eight engineering prototype 15kW ac single axis solar tracking sub-arrays were designed, fabricated, and installed at the Sacramento Municipal Utility District's Hedge Substation site in 1996 and are being evaluated for performance and reliability. A number of design enhancements will be implemented in 1997 and demonstrated by the field deployment and operation of over twenty advanced sub-array PV power systems.
A high resolution WRF model for wind energy forecasting
NASA Astrophysics Data System (ADS)
Vincent, Claire Louise; Liu, Yubao
2010-05-01
The increasing penetration of wind energy into national electricity markets has increased the demand for accurate surface layer wind forecasts. There has recently been a focus on forecasting the wind at wind farm sites using both statistical models and numerical weather prediction (NWP) models. Recent advances in computing capacity and non-hydrostatic NWP models means that it is possible to nest mesoscale models down to Large Eddy Simulation (LES) scales over the spatial area of a typical wind farm. For example, the WRF model (Skamarock 2008) has been run at a resolution of 123 m over a wind farm site in complex terrain in Colorado (Liu et al. 2009). Although these modelling attempts indicate a great hope for applying such models for detailed wind forecasts over wind farms, one of the obvious challenges of running the model at this resolution is that while some boundary layer structures are expected to be modelled explicitly, boundary layer eddies into the inertial sub-range can only be partly captured. Therefore, the amount and nature of sub-grid-scale mixing that is required is uncertain. Analysis of Liu et al. (2009) modelling results in comparison to wind farm observations indicates that unrealistic wind speed fluctuations with a period of around 1 hour occasionally occurred during the two day modelling period. The problem was addressed by re-running the same modelling system with a) a modified diffusion constant and b) two-way nesting between the high resolution model and its parent domain. The model, which was run with horizontal grid spacing of 370 m, had dimensions of 505 grid points in the east-west direction and 490 points in the north-south direction. It received boundary conditions from a mesoscale model of resolution 1111 m. Both models had 37 levels in the vertical. The mesoscale model was run with a non-local-mixing planetary boundary layer scheme, while the 370 m model was run with no planetary boundary layer scheme. It was found that increasing the diffusion constant caused damping of the unrealistic fluctuations, but did not completely solve the problem. Using two-way nesting also mitigated the unrealistic fluctuations significantly. It can be concluded that for real case LES modelling of wind farm circulations, care should be taken to ensure the consistency between the mesoscale weather forcing and LES models to avoid exciting spurious noise along the forcing boundary. The development of algorithms that adequately model the sub-grid-scale mixing that cannot be resolved by LES models is an important area for further research. References Liu, Y. Y._W. Liu, W. Y.Y. Cheng, W. Wu, T. T. Warner and K. Parks, 2009: Simulating intra-farm wind variations with the WRF-RTFDDA-LES modeling system. 10th WRF Users' Workshop, Boulder, C, USA. June 23 - 26, 2009. Skamarock, W., J. Dudhia, D.O. Gill, D.M. Barker, M.G.Duda, X-Y. Huang, W. Wang and J.G. Powers, A Description of the Advanced Research WRF version 3, NCAR Technical Note TN-475+STR, NCAR, Boulder, Colorado, 2008.
Nonlinear plasma wave models in 3D fluid simulations of laser-plasma interaction
NASA Astrophysics Data System (ADS)
Chapman, Thomas; Berger, Richard; Arrighi, Bill; Langer, Steve; Banks, Jeffrey; Brunner, Stephan
2017-10-01
Simulations of laser-plasma interaction (LPI) in inertial confinement fusion (ICF) conditions require multi-mm spatial scales due to the typical laser beam size and durations of order 100 ps in order for numerical laser reflectivities to converge. To be computationally achievable, these scales necessitate a fluid-like treatment of light and plasma waves with a spatial grid size on the order of the light wave length. Plasma waves experience many nonlinear phenomena not naturally described by a fluid treatment, such as frequency shifts induced by trapping, a nonlinear (typically suppressed) Landau damping, and mode couplings leading to instabilities that can cause the plasma wave to decay rapidly. These processes affect the onset and saturation of stimulated Raman and Brillouin scattering, and are of direct interest to the modeling and prediction of deleterious LPI in ICF. It is not currently computationally feasible to simulate these Debye length-scale phenomena in 3D across experimental scales. Analytically-derived and/or numerically benchmarked models of processes occurring at scales finer than the fluid simulation grid offer a path forward. We demonstrate the impact of a range of kinetic processes on plasma reflectivity via models included in the LPI simulation code pF3D. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Scale Issues in Air Quality Modeling
This presentation reviews past model evaluation studies investigating the impact of horizontal grid spacing on model performance. It also presents several examples of using a spectral decomposition technique to separate the forcings from processes operating on different time scal...
2012-02-02
flight hours to one significant atmospheric phenomena. OBJECTIVES The P-3 Doppler Wind Lidar (P3DWL) uses the latest version of a coherent ... Doppler transceiver developed at Lockheed Martin Coherent Technologies. The lidar , with the exception of the scanner, is shown on the top in Figure 1...Processes Observed by the P-3 Doppler Wind Lidar in Support of the Western Pacific Tropical Cyclone Structure 2008 Experiment Ralph Foster Applied
NASA Technical Reports Server (NTRS)
Chen, J. H.; Chong, M. S.; Soria, J.; Sondergaard, R.; Perry, A. E.; Rogers, M.; Moser, R.; Cantwell, B. J.
1990-01-01
A preliminary investigation of the geometry of flow patterns in numerically simulated compressible and incompressible mixing layers was carried out using 3-D critical point methodology. Motions characterized by high rates of kinetic energy dissipation and/or high enstrophy were of particular interest. In the approach the partial derivatives of the velocity field are determined at every point in the flow. These are used to construct the invariants of the velocity gradient tensor and the rate-of-strain tensor (P, Q, R, and P(sub s), Q(sub s), R(sub s) respectively). For incompressible flow the first invariant is zero. For the conditions of the compressible simulation, the first invariant is found to be everywhere small, relative to the second and third invariants, and so in both cases the local topology at a point is mainly determined by the second and third invariants. The data at every grid point is used to construct scatter plots of Q versus R and Q(sub s) versus R(sub s). Most points map to a cluster near the origin in Q-R space. However, fine scale motions, that is motions which are characterized by velocity derivatives which scale with the square root of R(sub delta), tend to map to regions which lie far from the origin. Definite trends are observed for motions characterized by high enstrophy and/or high dissipation. The observed trends suggest that, for these motions, the second and third invariants of the velocity gradient and rate-of-strain tensors are strongly correlated. The second and third invariants of the rate-of-strain tensor are related by K(-Q(sub s))(exp 3/2), which is consistent with the above scaling of velocity derivatives. The quantity K appears to depend on Reynolds number with an upper limit K = 2(the square root of 3)/9 corresponding to locally axisymmetric flow. For both the compressible and incompressible mixing layer, regions corresponding to high rates of dissipation are found to be characterized by comparable magnitudes of R(sub ij)R(sub ij) and S(sub ij)S(sub ij). For the incompressible mixing layer, regions characterized by the highest values of enstrophy are found to have relatively low strain rates.
Inoue, Kentaro; Shimozono, Shinichi; Yoshida, Hideaki; Kurata, Hiroyuki
2012-01-01
Background For visualizing large-scale biochemical network maps, it is important to calculate the coordinates of molecular nodes quickly and to enhance the understanding or traceability of them. The grid layout is effective in drawing compact, orderly, balanced network maps with node label spaces, but existing grid layout algorithms often require a high computational cost because they have to consider complicated positional constraints through the entire optimization process. Results We propose a hybrid grid layout algorithm that consists of a non-grid, fast layout (preprocessor) algorithm and an approximate pattern matching algorithm that distributes the resultant preprocessed nodes on square grid points. To demonstrate the feasibility of the hybrid layout algorithm, it is characterized in terms of the calculation time, numbers of edge-edge and node-edge crossings, relative edge lengths, and F-measures. The proposed algorithm achieves outstanding performances compared with other existing grid layouts. Conclusions Use of an approximate pattern matching algorithm quickly redistributes the laid-out nodes by fast, non-grid algorithms on the square grid points, while preserving the topological relationships among the nodes. The proposed algorithm is a novel use of the pattern matching, thereby providing a breakthrough for grid layout. This application program can be freely downloaded from http://www.cadlive.jp/hybridlayout/hybridlayout.html. PMID:22679486
NASA Astrophysics Data System (ADS)
Bower, Ward
2011-09-01
An overview of the activities and progress made during the US DOE Solar Energy Grid Integration Systems (SEGIS) solicitation, while maintaining reliability and economics is provided. The SEGIS R&D opened pathways for interconnecting PV systems to intelligent utility grids and micro-grids of the future. In addition to new capabilities are "value added" features. The new hardware designs resulted in smaller, less material-intensive products that are being viewed by utilities as enabling dispatchable generation and not just unpredictable negative loads. The technical solutions enable "advanced integrated system" concepts and "smart grid" processes to move forward in a faster and focused manner. The advanced integrated inverters/controllers can now incorporate energy management functionality, intelligent electrical grid support features and a multiplicity of communication technologies. Portals for energy flow and two-way communications have been implemented. SEGIS hardware was developed for the utility grid of today, which was designed for one-way power flow, for intermediate grid scenarios, AND for the grid of tomorrow, which will seamlessly accommodate managed two-way power flows as required by large-scale deployment of solar and other distributed generation. The SEGIS hardware and control developed for today meets existing standards and codes AND provides for future connections to a "smart grid" mode that enables utility control and optimized performance.
Inoue, Kentaro; Shimozono, Shinichi; Yoshida, Hideaki; Kurata, Hiroyuki
2012-01-01
For visualizing large-scale biochemical network maps, it is important to calculate the coordinates of molecular nodes quickly and to enhance the understanding or traceability of them. The grid layout is effective in drawing compact, orderly, balanced network maps with node label spaces, but existing grid layout algorithms often require a high computational cost because they have to consider complicated positional constraints through the entire optimization process. We propose a hybrid grid layout algorithm that consists of a non-grid, fast layout (preprocessor) algorithm and an approximate pattern matching algorithm that distributes the resultant preprocessed nodes on square grid points. To demonstrate the feasibility of the hybrid layout algorithm, it is characterized in terms of the calculation time, numbers of edge-edge and node-edge crossings, relative edge lengths, and F-measures. The proposed algorithm achieves outstanding performances compared with other existing grid layouts. Use of an approximate pattern matching algorithm quickly redistributes the laid-out nodes by fast, non-grid algorithms on the square grid points, while preserving the topological relationships among the nodes. The proposed algorithm is a novel use of the pattern matching, thereby providing a breakthrough for grid layout. This application program can be freely downloaded from http://www.cadlive.jp/hybridlayout/hybridlayout.html.
The value of residential photovoltaic systems: A comprehensive assessment
NASA Technical Reports Server (NTRS)
Borden, C. S.
1983-01-01
Utility-interactive photovoltaic (PV) arrays on residential rooftops appear to be a potentially attractive, large-scale application of PV technology. Results of a comprehensive assessment of the value (i.e., break-even cost) of utility-grid connected residential photovoltaic power systems under a variety of technological and economic assumptions are presented. A wide range of allowable PV system costs are calculated for small (4.34 kW (p) sub ac) residential PV systems in various locales across the United States. Primary factor in this variation are differences in local weather conditions, utility-specific electric generation capacity, fuel types, and customer-load profiles that effect purchase and sell-back rates, and non-uniform state tax considerations. Additional results from this analysis are: locations having the highest insolation values are not necessary the most economically attractive sites; residential PV systems connected in parallel to the utility demonstrate high percentages of energy sold back to the grid, and owner financial and tax assumptions cause large variations in break-even costs. Significant cost reduction and aggressive resolution of potential institutional impediments (e.g., liability, standards, metering, and technical integration) are required for a residential PV marker to become a major electric-grid-connected energy-generation source.
The value of residential photovoltaic systems: A comprehensive assessment
NASA Astrophysics Data System (ADS)
Borden, C. S.
1983-09-01
Utility-interactive photovoltaic (PV) arrays on residential rooftops appear to be a potentially attractive, large-scale application of PV technology. Results of a comprehensive assessment of the value (i.e., break-even cost) of utility-grid connected residential photovoltaic power systems under a variety of technological and economic assumptions are presented. A wide range of allowable PV system costs are calculated for small (4.34 kW (p) sub ac) residential PV systems in various locales across the United States. Primary factor in this variation are differences in local weather conditions, utility-specific electric generation capacity, fuel types, and customer-load profiles that effect purchase and sell-back rates, and non-uniform state tax considerations. Additional results from this analysis are: locations having the highest insolation values are not necessary the most economically attractive sites; residential PV systems connected in parallel to the utility demonstrate high percentages of energy sold back to the grid, and owner financial and tax assumptions cause large variations in break-even costs. Significant cost reduction and aggressive resolution of potential institutional impediments (e.g., liability, standards, metering, and technical integration) are required for a residential PV marker to become a major electric-grid-connected energy-generation source.
A practical approach to virtualization in HEP
NASA Astrophysics Data System (ADS)
Buncic, P.; Aguado Sánchez, C.; Blomer, J.; Harutyunyan, A.; Mudrinic, M.
2011-01-01
In the attempt to solve the problem of processing data coming from LHC experiments at CERN at a rate of 15PB per year, for almost a decade the High Enery Physics (HEP) community has focused its efforts on the development of the Worldwide LHC Computing Grid. This generated large interest and expectations promising to revolutionize computing. Meanwhile, having initially taken part in the Grid standardization process, industry has moved in a different direction and started promoting the Cloud Computing paradigm which aims to solve problems on a similar scale and in equally seamless way as it was expected in the idealized Grid approach. A key enabling technology behind Cloud computing is server virtualization. In early 2008, an R&D project was established in the PH-SFT group at CERN to investigate how virtualization technology could be used to improve and simplify the daily interaction of physicists with experiment software frameworks and the Grid infrastructure. In this article we shall first briefly compare Grid and Cloud computing paradigms and then summarize the results of the R&D activity pointing out where and how virtualization technology could be effectively used in our field in order to maximize practical benefits whilst avoiding potential pitfalls.
NASA Astrophysics Data System (ADS)
Loikith, P. C.; Broccoli, A. J.; Waliser, D. E.; Lintner, B. R.; Neelin, J. D.
2015-12-01
Anomalous large-scale circulation patterns often play a key role in the occurrence of temperature extremes. For example, large-scale circulation can drive horizontal temperature advection or influence local processes that lead to extreme temperatures, such as by inhibiting moderating sea breezes, promoting downslope adiabatic warming, and affecting the development of cloud cover. Additionally, large-scale circulation can influence the shape of temperature distribution tails, with important implications for the magnitude of future changes in extremes. As a result of the prominent role these patterns play in the occurrence and character of extremes, the way in which temperature extremes change in the future will be highly influenced by if and how these patterns change. It is therefore critical to identify and understand the key patterns associated with extremes at local to regional scales in the current climate and to use this foundation as a target for climate model validation. This presentation provides an overview of recent and ongoing work aimed at developing and applying novel approaches to identifying and describing the large-scale circulation patterns associated with temperature extremes in observations and using this foundation to evaluate state-of-the-art global and regional climate models. Emphasis is given to anomalies in sea level pressure and 500 hPa geopotential height over North America using several methods to identify circulation patterns, including self-organizing maps and composite analysis. Overall, evaluation results suggest that models are able to reproduce observed patterns associated with temperature extremes with reasonable fidelity in many cases. Model skill is often highest when and where synoptic-scale processes are the dominant mechanisms for extremes, and lower where sub-grid scale processes (such as those related to topography) are important. Where model skill in reproducing these patterns is high, it can be inferred that extremes are being simulated for plausible physical reasons, boosting confidence in future projections of temperature extremes. Conversely, where model skill is identified to be lower, caution should be exercised in interpreting future projections.
Pindrik, Jonathan; Hoang, Nguyen; Tubbs, R Shane; Rocque, Brandon J; Rozzelle, Curtis J
2017-08-01
Phase II monitoring with intracranial electroencephalography (ICEEG) occasionally requires bilateral placement of subdural (SD) strips, grids, and/or depth electrodes. While phase I monitoring often demonstrates a preponderance of unilateral findings, individual studies (video EEG, single photon emission computed tomography [SPECT], and positron emission tomography [PET]) can suggest or fail to exclude a contralateral epileptogenic onset zone. This study describes previously unreported techniques of trans-falcine and sub-frontal insertion of contralateral SD grids and depth electrodes for phase II monitoring in pediatric epilepsy surgery patients when concern about bilateral abnormalities has been elicited during phase I monitoring. Pediatric patients with medically refractory epilepsy undergoing stage I surgery for phase II monitoring involving sub-frontal and/or trans-falcine insertion of SD grids and/or depth electrodes at the senior author's institution were retrospectively reviewed. Intra-operative technical details of sub-frontal and trans-falcine approaches were studied, while intra-operative complications or events were noted. Operative techniques included gentle subfrontal retraction and elevation of the olfactory tracts (while preserving the relationship between the olfactory bulb and cribriform plate) to insert SD grids across the midline for coverage of the contralateral orbito-frontal regions. Trans-falcine approaches involved accessing the inter-hemispheric space, bipolar cauterization of the anterior falx cerebri below the superior sagittal sinus, and sharp dissection using a blunt elevator and small blade scalpel. The falcine window allowed contralateral SD strip, grid, and depth electrodes to be inserted for coverage of the contralateral frontal regions. The study cohort included seven patients undergoing sub-frontal and/or trans-falcine insertion of contralateral SD strip, grid, and/or depth electrodes from February 2012 through June 2015. Five patients (71%) experienced no intra-operative events related to contralateral ICEEG electrode insertion. Intra-operative events of frontal territory venous engorgement (1/7, 14%) due to sacrifice of anterior bridging veins draining into the SSS and avulsion of a contralateral bridging vein (1/7, 14%), probably due to prior anterior corpus callosotomy, each occurred in one patient. There were no intra-operative or peri-operative complications in any of the patients studied. Two patients required additional surgery for supplemental SD strip and/or depth electrodes via burr hole craniectomy to enhance phase II monitoring. All patients proceeded to stage II surgery for resection of ipsilateral epileptogenic onset zones without adverse events. Trans-falcine and sub-frontal insertion of contralateral SD strip, grid, and depth electrodes are previously unreported techniques for achieving bilateral frontal coverage in phase II monitoring in pediatric epilepsy surgery. This technique obviates the need for contralateral craniotomy and parenchymal exposure with limited, remediable risks. Larger case series using the method described herein are now necessary.
Using Multi-Scale Modeling Systems and Satellite Data to Study the Precipitation Processes
NASA Technical Reports Server (NTRS)
Tao, Wei--Kuo; Chern, J.; Lamg, S.; Matsui, T.; Shen, B.; Zeng, X.; Shi, R.
2010-01-01
In recent years, exponentially increasing computer power extended Cloud Resolving Model (CRM) integrations from hours to months, the number of computational grid points from less than a thousand to close to ten million. Three-dimensional models are now more prevalent. Much attention is devoted to precipitating cloud systems where the crucial 1-km scales are resolved in horizontal domains as large as 10,000 km in two-dimensions, and 1,000 x 1,000 sq km in three-dimensions. Cloud resolving models now provide statistical information useful for developing more realistic physically based parameterizations for climate models and numerical weather prediction models. It is also expected that NWP and mesoscale models can be run in grid size similar to cloud resolving models through nesting technique. Recently, a multi-scale modeling system with unified physics was developed at NASA Goddard. It consists of (1) a cloud-resolving model (Goddard Cumulus Ensemble model, GCE model). (2) a regional scale model (a NASA unified weather research and forecast, W8F). (3) a coupled CRM and global model (Goddard Multi-scale Modeling Framework, MMF), and (4) a land modeling system. The same microphysical processes, long and short wave radiative transfer and land processes and the explicit cloud-radiation and cloud-land surface interactive processes are applied in this multi-scale modeling system. This modeling system has been coupled with a multi-satellite simulator to use NASA high-resolution satellite data to identify the strengths and weaknesses of cloud and precipitation processes simulated by the model. In this talk, a review of developments and applications of the multi-scale modeling system will be presented. In particular, the results from using multi-scale modeling systems to study the interactions between clouds, precipitation, and aerosols will be presented. Also how to use the multi-satellite simulator to improve precipitation processes will be discussed.
Using Multi-Scale Modeling Systems to Study the Precipitation Processes
NASA Technical Reports Server (NTRS)
Tao, Wei-Kuo
2010-01-01
In recent years, exponentially increasing computer power has extended Cloud Resolving Model (CRM) integrations from hours to months, the number of computational grid points from less than a thousand to close to ten million. Three-dimensional models are now more prevalent. Much attention is devoted to precipitating cloud systems where the crucial 1-km scales are resolved in horizontal domains as large as 10,000 km in two-dimensions, and 1,000 x 1,000 km2 in three-dimensions. Cloud resolving models now provide statistical information useful for developing more realistic physically based parameterizations for climate models and numerical weather prediction models. It is also expected that NWP and mesoscale model can be run in grid size similar to cloud resolving model through nesting technique. Recently, a multi-scale modeling system with unified physics was developed at NASA Goddard. It consists of (1) a cloud-resolving model (Goddard Cumulus Ensemble model, GCE model), (2) a regional scale model (a NASA unified weather research and forecast, WRF), (3) a coupled CRM and global model (Goddard Multi-scale Modeling Framework, MMF), and (4) a land modeling system. The same microphysical processes, long and short wave radiative transfer and land processes and the explicit cloud-radiation, and cloud-land surface interactive processes are applied in this multi-scale modeling system. This modeling system has been coupled with a multi-satellite simulator to use NASA high-resolution satellite data to identify the strengths and weaknesses of cloud and precipitation processes simulated by the model. In this talk, a review of developments and applications of the multi-scale modeling system will be presented. In particular, the results from using multi-scale modeling system to study the interactions between clouds, precipitation, and aerosols will be presented. Also how to use of the multi-satellite simulator to improve precipitation processes will be discussed.
Laser wafering for silicon solar.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friedmann, Thomas Aquinas; Sweatt, William C.; Jared, Bradley Howell
2011-03-01
Current technology cuts solar Si wafers by a wire saw process, resulting in 50% 'kerf' loss when machining silicon from a boule or brick into a wafer. We want to develop a kerf-free laser wafering technology that promises to eliminate such wasteful wire saw processes and achieve up to a ten-fold decrease in the g/W{sub p} (grams/peak watt) polysilicon usage from the starting polysilicon material. Compared to today's technology, this will also reduce costs ({approx}20%), embodied energy, and green-house gas GHG emissions ({approx}50%). We will use short pulse laser illumination sharply focused by a solid immersion lens to produce subsurfacemore » damage in silicon such that wafers can be mechanically cleaved from a boule or brick. For this concept to succeed, we will need to develop optics, lasers, cleaving, and high throughput processing technologies capable of producing wafers with thicknesses < 50 {micro}m with high throughput (< 10 sec./wafer). Wafer thickness scaling is the 'Moore's Law' of silicon solar. Our concept will allow solar manufacturers to skip entire generations of scaling and achieve grid parity with commercial electricity rates. Yet, this idea is largely untested and a simple demonstration is needed to provide credibility for a larger scale research and development program. The purpose of this project is to lay the groundwork to demonstrate the feasibility of laser wafering. First, to design and procure on optic train suitable for producing subsurface damage in silicon with the required damage and stress profile to promote lateral cleavage of silicon. Second, to use an existing laser to produce subsurface damage in silicon, and third, to characterize the damage using scanning electron microscopy and confocal Raman spectroscopy mapping.« less
Importance of Grid Center Arrangement
NASA Astrophysics Data System (ADS)
Pasaogullari, O.; Usul, N.
2012-12-01
In Digital Elevation Modeling, grid size is accepted to be the most important parameter. Despite the point density and/or scale of the source data, it is freely decided by the user. Most of the time, arrangement of the grid centers are ignored, even most GIS packages omit the choice of grid center coordinate selection. In our study; importance of the arrangement of grid centers is investigated. Using the analogy between "Raster Grid DEM" and "Bitmap Image", importance of placement of grid centers in DEMs are measured. The study has been conducted on four different grid DEMs obtained from a half ellipsoid. These grid DEMs are obtained in such a way that they are half grid size apart from each other. Resulting grid DEMs are investigated through similarity measures. Image processing scientists use different measures to investigate the dis/similarity between the images and the amount of different information they carry. Grid DEMs are projected to a finer grid in order to co-center. Similarity measures are then applied to each grid DEM pairs. These similarity measures are adapted to DEM with band reduction and real number operation. One of the measures gives function graph and the others give measure matrices. Application of similarity measures to six grid DEM pairs shows interesting results. These four different grid DEMs are created with the same method for the same area, surprisingly; thirteen out of 14 measures state that, the half grid size apart grid DEMs are different from each other. The results indicated that although grid DEMs carry mutual information, they have also additional individual information. In other words, half grid size apart constructed grid DEMs have non-redundant information.; Joint Probability Distributions Function Graphs
NASA Astrophysics Data System (ADS)
Alvarado, M. J.; Lonsdale, C. R.; Yokelson, R. J.; Travis, K.; Fischer, E. V.; Lin, J. C.
2014-12-01
Forecasting the impacts of biomass burning (BB) plumes on air quality is difficult due to the complex photochemistry that takes place in the concentrated young BB plumes. The spatial grid of global and regional scale Eulerian models is generally too large to resolve BB photochemistry, which can lead to errors in predicting the formation of secondary organic aerosol (SOA) and O3, as well as the partitioning of NOyspecies. AER's Aerosol Simulation Program (ASP v2.1) can be used within plume-scale Lagrangian models to simulate this complex photochemistry. We will present results of validation studies of the ASP model against aircraft observations of young BB smoke plumes. We will also present initial results from the coupling of ASP v2.1 into the Lagrangian particle dispersion model STILT-Chem in order to better examine the interactions between BB plume chemistry and dispersion. In addition, we have used ASP to develop a sub-grid scale parameterization of the near-source chemistry of BB plumes for use in regional and global air quality models. The parameterization takes inputs from the host model, such as solar zenith angle, temperature, and fire fuel type, and calculates enhancement ratios of O3, NOx, PAN, aerosol nitrate, and other NOy species, as well as organic aerosol (OA). We will present results from the ASP-based BB parameterization as well as its implementation into the global atmospheric composition model GEOS-Chem for the SEAC4RS campaign.
Numerical modeling and analysis of the effect of Greek complex topography on tornado genesis
NASA Astrophysics Data System (ADS)
Matsangouras, I. T.; Pytharoulis, I.; Nastos, P. T.
2014-02-01
Tornadoes have been reported in Greece over the last decades in specific sub-geographical areas and have been associated with strong synoptic forcing. It is well known that meteorological conditions over Greece are affected at various scales by the significant variability of topography, the Ionian Sea at the west and the Aegean Sea at the east. However, there is still uncertainty regarding topography's importance on tornadic generation and development. The aim of this study is to investigate the role of topography in significant tornado genesis events that were triggered under strong synoptic scale forcing over Greece. Three tornado events that occurred over the last years in Thiva (Boeotia, 17 November 2007), Vrastema (Chalkidiki, 12 February 2010) and Vlychos (Lefkada, 20 September 2011) have been selected for numerical experiments. These events were associated with synoptic scale forcing, while their intensity was T4-T5 (Torro scale) and caused significant damage. The simulations were performed using the non-hydrostatic Weather Research and Forecasting model (WRF), initialized with ECMWF gridded analyses, with telescoping nested grids that allow the representation of atmospheric circulations ranging from the synoptic scale down to the meso scale. In the experiments the topography of the inner grid was modified by: (a) 0% (actual topography) and (b) -100% (without topography). The aim was to determine whether the occurrence of tornadoes - mainly identified by various severe weather instability indices - could be indicated by modifying topography. The main utilized instability variables concerned the Bulk Richardson number shear (BRN), the energy helicity index (EHI), the storm-relative environmental helicity (SRH) and the maximum convective available potential energy (MCAPE, for parcel with maximum theta-e). Additional a verification of model was conducted for every sensitivity experiment accompanied with analysis absolute vorticity budget. Numerical simulations revealed that the complex topography was denoted as an important factor during 17 November 2007 and 12 February 2010 events, based on EHI and BRN analyses. Topography around 20 September 2011 event was characterized as the least factor based on EHI, SRH, BRN analyses.
NASA Astrophysics Data System (ADS)
Jang, W.; Engda, T. A.; Neff, J. C.; Herrick, J.
2017-12-01
Many crop models are increasingly used to evaluate crop yields at regional and global scales. However, implementation of these models across large areas using fine-scale grids is limited by computational time requirements. In order to facilitate global gridded crop modeling with various scenarios (i.e., different crop, management schedule, fertilizer, and irrigation) using the Environmental Policy Integrated Climate (EPIC) model, we developed a distributed parallel computing framework in Python. Our local desktop with 14 cores (28 threads) was used to test the distributed parallel computing framework in Iringa, Tanzania which has 406,839 grid cells. High-resolution soil data, SoilGrids (250 x 250 m), and climate data, AgMERRA (0.25 x 0.25 deg) were also used as input data for the gridded EPIC model. The framework includes a master file for parallel computing, input database, input data formatters, EPIC model execution, and output analyzers. Through the master file for parallel computing, the user-defined number of threads of CPU divides the EPIC simulation into jobs. Then, Using EPIC input data formatters, the raw database is formatted for EPIC input data and the formatted data moves into EPIC simulation jobs. Then, 28 EPIC jobs run simultaneously and only interesting results files are parsed and moved into output analyzers. We applied various scenarios with seven different slopes and twenty-four fertilizer ranges. Parallelized input generators create different scenarios as a list for distributed parallel computing. After all simulations are completed, parallelized output analyzers are used to analyze all outputs according to the different scenarios. This saves significant computing time and resources, making it possible to conduct gridded modeling at regional to global scales with high-resolution data. For example, serial processing for the Iringa test case would require 113 hours, while using the framework developed in this study requires only approximately 6 hours, a nearly 95% reduction in computing time.
Eisele, Thomas P; Keating, Joseph; Swalm, Chris; Mbogo, Charles M; Githeko, Andrew K; Regens, James L; Githure, John I; Andrews, Linda; Beier, John C
2003-12-10
BACKGROUND: Remote sensing technology provides detailed spectral and thermal images of the earth's surface from which surrogate ecological indicators of complex processes can be measured. METHODS: Remote sensing data were overlaid onto georeferenced entomological and human ecological data randomly sampled during April and May 2001 in the cities of Kisumu (population asymptotically equal to 320,000) and Malindi (population asymptotically equal to 81,000), Kenya. Grid cells of 270 meters x 270 meters were used to generate spatial sampling units for each city for the collection of entomological and human ecological field-based data. Multispectral Thermal Imager (MTI) satellite data in the visible spectrum at five meter resolution were acquired for Kisumu and Malindi during February and March 2001, respectively. The MTI data were fit and aggregated to the 270 meter x 270 meter grid cells used in field-based sampling using a geographic information system. The normalized difference vegetation index (NDVI) was calculated and scaled from MTI data for selected grid cells. Regression analysis was used to assess associations between NDVI values and entomological and human ecological variables at the grid cell level. RESULTS: Multivariate linear regression showed that as household density increased, mean grid cell NDVI decreased (global F-test = 9.81, df 3,72, P-value = <0.01; adjusted R2 = 0.26). Given household density, the number of potential anopheline larval habitats per grid cell also increased with increasing values of mean grid cell NDVI (global F-test = 14.29, df 3,36, P-value = <0.01; adjusted R2 = 0.51). CONCLUSIONS: NDVI values obtained from MTI data were successfully overlaid onto georeferenced entomological and human ecological data spatially sampled at a scale of 270 meters x 270 meters. Results demonstrate that NDVI at such a scale was sufficient to describe variations in entomological and human ecological parameters across both cities.
Integration of a neuroimaging processing pipeline into a pan-canadian computing grid
NASA Astrophysics Data System (ADS)
Lavoie-Courchesne, S.; Rioux, P.; Chouinard-Decorte, F.; Sherif, T.; Rousseau, M.-E.; Das, S.; Adalat, R.; Doyon, J.; Craddock, C.; Margulies, D.; Chu, C.; Lyttelton, O.; Evans, A. C.; Bellec, P.
2012-02-01
The ethos of the neuroimaging field is quickly moving towards the open sharing of resources, including both imaging databases and processing tools. As a neuroimaging database represents a large volume of datasets and as neuroimaging processing pipelines are composed of heterogeneous, computationally intensive tools, such open sharing raises specific computational challenges. This motivates the design of novel dedicated computing infrastructures. This paper describes an interface between PSOM, a code-oriented pipeline development framework, and CBRAIN, a web-oriented platform for grid computing. This interface was used to integrate a PSOM-compliant pipeline for preprocessing of structural and functional magnetic resonance imaging into CBRAIN. We further tested the capacity of our infrastructure to handle a real large-scale project. A neuroimaging database including close to 1000 subjects was preprocessed using our interface and publicly released to help the participants of the ADHD-200 international competition. This successful experiment demonstrated that our integrated grid-computing platform is a powerful solution for high-throughput pipeline analysis in the field of neuroimaging.
Knoch, Tobias A; Baumgärtner, Volkmar; de Zeeuw, Luc V; Grosveld, Frank G; Egger, Kurt
2009-01-01
With ever-new technologies emerging also the amount of information to be stored and processed is growing exponentially and is believed to be always at the limit. In contrast, however, huge resources are available in the IT sector alike e.g. the renewable energy sector, which are often even not at all used. This under-usage bares any rational especially in the IT sector where e.g. virtualisation and grid approaches could be fast implemented due to the great technical and fast turnover opportunities. Here, we describe this obvious paradox for the first time as the Inverse Tragedy of the Commons, in contrast to the Classical Tragedy of the Commons where resources are overexploited. From this perspective the grid IT sector attempting to share resources for better efficiency, reveals two challenges leading to the heart of the paradox: i) From a macro perspective all grid infrastructures involve not only mere technical solutions but also dominantly all of the autopoietic social sub-systems ranging from religion to policy. ii) On the micro level the individual players and their psychology and risk behaviour are of major importance for acting within the macro autopoietic framework. Thus, the challenges of grid implementation are similar to those of e.g. climate protection. This is well described by the classic Human Ecology triangle and our extension to a rectangle: invironment-individual-society-environment. Extension of this classical interdisciplinary field of basic and applied research to an e-Human Grid Ecology rational, allows the Inverse Tragedy of the Commons of the grid sector to be understood and approached better and implies obvious guidelines in the day-to-day management for grid and other (networked) resources, which is of importance for many fields with similar paradoxes as in (e-)society.
NASA Astrophysics Data System (ADS)
Lopez Bobeda, J. R.
2017-12-01
The increasing use of groundwater for irrigation of crops has exacerbated groundwater sustainability issues faced by water limited regions. Gridded, process-based crop models have the potential to help farmers and policymakers asses the effects water shortages on yield and devise new strategies for sustainable water use. Gridded crop models are typically calibrated and evaluated using county-level survey data of yield, planting dates, and maturity dates. However, little is known about the ability of these models to reproduce observed crop evapotranspiration and water use at regional scales. The aim of this work is to evaluate a gridded version of the Decision Support System for Agrotechnology Transfer (DSSAT) crop model over the continental United States. We evaluated crop seasonal evapotranspiration over 5 arc-minute grids, and irrigation water use at the county level. Evapotranspiration was assessed only for rainfed agriculture to test the model evapotranspiration equations separate from the irrigation algorithm. Model evapotranspiration was evaluated against the Atmospheric Land Exchange Inverse (ALEXI) modeling product. Using a combination of the USDA crop land data layer (CDL) and the USGS Moderate Resolution Imaging Spectroradiometer Irrigated Agriculture Dataset for the United States (MIrAD-US), we selected only grids with more than 60% of their area planted with the simulated crops (corn, cotton, and soybean), and less than 20% of their area irrigated. Irrigation water use was compared against the USGS county level irrigated agriculture water use survey data. Simulated gridded data were aggregated to county level using USDA CDL and USGS MIrAD-US. Only counties where 70% or more of the irrigated land was corn, cotton, or soybean were selected for the evaluation. Our results suggest that gridded crop models can reasonably reproduce crop evapotranspiration at the country scale (RRMSE = 10%).
NASA Technical Reports Server (NTRS)
Mohr, Karen Irene; Tao, Wei-Kuo; Chern, Jiun-Dar; Kumar, Sujay V.; Peters-Lidard, Christa D.
2013-01-01
The present generation of general circulation models (GCM) use parameterized cumulus schemes and run at hydrostatic grid resolutions. To improve the representation of cloud-scale moist processes and landeatmosphere interactions, a global, Multi-scale Modeling Framework (MMF) coupled to the Land Information System (LIS) has been developed at NASA-Goddard Space Flight Center. The MMFeLIS has three components, a finite-volume (fv) GCM (Goddard Earth Observing System Ver. 4, GEOS-4), a 2D cloud-resolving model (Goddard Cumulus Ensemble, GCE), and the LIS, representing the large-scale atmospheric circulation, cloud processes, and land surface processes, respectively. The non-hydrostatic GCE model replaces the single-column cumulus parameterization of fvGCM. The model grid is composed of an array of fvGCM gridcells each with a series of embedded GCE models. A horizontal coupling strategy, GCE4fvGCM4Coupler4LIS, offered significant computational efficiency, with the scalability and I/O capabilities of LIS permitting landeatmosphere interactions at cloud-scale. Global simulations of 2007e2008 and comparisons to observations and reanalysis products were conducted. Using two different versions of the same land surface model but the same initial conditions, divergence in regional, synoptic-scale surface pressure patterns emerged within two weeks. The sensitivity of largescale circulations to land surface model physics revealed significant functional value to using a scalable, multi-model land surface modeling system in global weather and climate prediction.
Modeling basin- and plume-scale processes of CO2 storage for full-scale deployment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhou, Q.; Birkholzer, J.T.; Mehnert, E.
Integrated modeling of basin- and plume-scale processes induced by full-scale deployment of CO{sub 2} storage was applied to the Mt. Simon Aquifer in the Illinois Basin. A three-dimensional mesh was generated with local refinement around 20 injection sites, with approximately 30 km spacing. A total annual injection rate of 100 Mt CO{sub 2} over 50 years was used. The CO{sub 2}-brine flow at the plume scale and the single-phase flow at the basin scale were simulated. Simulation results show the overall shape of a CO{sub 2} plume consisting of a typical gravity-override subplume in the bottom injection zone of highmore » injectivity and a pyramid-shaped subplume in the overlying multilayered Mt. Simon, indicating the important role of a secondary seal with relatively low-permeability and high-entry capillary pressure. The secondary-seal effect is manifested by retarded upward CO{sub 2} migration as a result of multiple secondary seals, coupled with lateral preferential CO{sub 2} viscous fingering through high-permeability layers. The plume width varies from 9.0 to 13.5 km at 200 years, indicating the slow CO{sub 2} migration and no plume interference between storage sites. On the basin scale, pressure perturbations propagate quickly away from injection centers, interfere after less than 1 year, and eventually reach basin margins. The simulated pressure buildup of 35 bar in the injection area is not expected to affect caprock geomechanical integrity. Moderate pressure buildup is observed in Mt. Simon in northern Illinois. However, its impact on groundwater resources is less than the hydraulic drawdown induced by long-term extensive pumping from overlying freshwater aquifers.« less
Method of determining the x-ray limit of an ion gauge
Edwards, Jr., David; Lanni, Christopher P.
1981-01-01
An ion gauge having a reduced "x-ray limit" and means for measuring that limit. The gauge comprises an ion gauge of the Bayard-Alpert type having a short collector and having means for varying the grid-collector voltage. The "x-ray limit" (i.e. the collector current resulting from x-rays striking the collector) may then be determined by the formula: ##EQU1## where: I.sub.x ="x-ray limit", I.sub.l and I.sub.h =the collector current at the lower and higher grid voltage respectively; and, .alpha.=the ratio of the collector current due to positive ions at the higher voltage to that at the lower voltage.
1991-02-15
picked, Ce11Pop" .xmonth, CeliPcpUpA .hour’ . Phase kND $80) = 0 ELSE IF (stationinfol36’ [stations. picked, CellIP-- CpA .n=nh, CeSUP pP.hour . Phiase...CellGrid, irt (322,24. 281,314, RightCeliGridAction, ShoCe11~ta, DcNot-hingPr-oc, bJii ne (lfepnIi lfs t.Xj05,efIghplit. Y4, 60,16,white, blak , black...8217.Hilite(oc,yy); with CellPI~p do begin if (SubCells (Hcnth,Hr] .X < (Get~4axX - RightsideStatsA .width - SubCellIs (Month, Hour) Width - SubCellP~ cpA
NASA Astrophysics Data System (ADS)
Lupi, Alessandro; Bovino, Stefano; Capelo, Pedro R.; Volonteri, Marta; Silk, Joseph
2018-03-01
In this study, we present a suite of high-resolution numerical simulations of an isolated galaxy to test a sub-grid framework to consistently follow the formation and dissociation of H2 with non-equilibrium chemistry. The latter is solved via the package KROME, coupled to the mesh-less hydrodynamic code GIZMO. We include the effect of star formation (SF), modelled with a physically motivated prescription independent of H2, supernova feedback and mass-losses from low-mass stars, extragalactic and local stellar radiation, and dust and H2 shielding, to investigate the emergence of the observed correlation between H2 and SF rate surface densities. We present two different sub-grid models and compare them with on-the-fly radiative transfer (RT) calculations, to assess the main differences and limits of the different approaches. We also discuss a sub-grid clumping factor model to enhance the H2 formation, consistent with our SF prescription, which is crucial, at the achieved resolution, to reproduce the correlation with H2. We find that both sub-grid models perform very well relative to the RT simulation, giving comparable results, with moderate differences, but at much lower computational cost. We also find that, while the Kennicutt-Schmidt relation for the total gas is not strongly affected by the different ingredients included in the simulations, the H2-based counterpart is much more sensitive, because of the crucial role played by the dissociating radiative flux and the gas shielding.
Effects of multiple-scale driving on turbulence statistics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yoo, Hyunju; Cho, Jungyeon, E-mail: hyunju527@gmail.com, E-mail: jcho@cnu.ac.kr
2014-01-01
Turbulence is ubiquitous in astrophysical fluids such as the interstellar medium and the intracluster medium. In turbulence studies, it is customary to assume that fluid is driven on a single scale. However, in astrophysical fluids, there can be many different driving mechanisms that act on different scales. If there are multiple energy-injection scales, the process of energy cascade and turbulence dynamo will be different compared with the case of the single energy-injection scale. In this work, we perform three-dimensional incompressible/compressible magnetohydrodynamic turbulence simulations. We drive turbulence in Fourier space in two wavenumber ranges, 2≤k≤√12 (large scale) and 15 ≲ kmore » ≲ 26 (small scale). We inject different amount of energy in each range by changing the amplitude of forcing in the range. We present the time evolution of the kinetic and magnetic energy densities and discuss the turbulence dynamo in the presence of energy injections at two scales. We show how kinetic, magnetic, and density spectra are affected by the two-scale energy injections and we discuss the observational implications. In the case ε {sub L} < ε {sub S}, where ε {sub L} and ε {sub S} are energy-injection rates at the large and small scales, respectively, our results show that even a tiny amount of large-scale energy injection can significantly change the properties of turbulence. On the other hand, when ε {sub L} ≳ ε {sub S}, the small-scale driving does not influence the turbulence statistics much unless ε {sub L} ∼ ε {sub S}.« less
NASA Technical Reports Server (NTRS)
Mccormick, S.; Quinlan, D.
1989-01-01
The fast adaptive composite grid method (FAC) is an algorithm that uses various levels of uniform grids (global and local) to provide adaptive resolution and fast solution of PDEs. Like all such methods, it offers parallelism by using possibly many disconnected patches per level, but is hindered by the need to handle these levels sequentially. The finest levels must therefore wait for processing to be essentially completed on all the coarser ones. A recently developed asynchronous version of FAC, called AFAC, completely eliminates this bottleneck to parallelism. This paper describes timing results for AFAC, coupled with a simple load balancing scheme, applied to the solution of elliptic PDEs on an Intel iPSC hypercube. These tests include performance of certain processes necessary in adaptive methods, including moving grids and changing refinement. A companion paper reports on numerical and analytical results for estimating convergence factors of AFAC applied to very large scale examples.
NASA Astrophysics Data System (ADS)
Gao, Yang; Leung, L. Ruby; Zhao, Chun; Hagos, Samson
2017-03-01
Simulating summer precipitation is a significant challenge for climate models that rely on cumulus parameterizations to represent moist convection processes. Motivated by recent advances in computing that support very high-resolution modeling, this study aims to systematically evaluate the effects of model resolution and convective parameterizations across the gray zone resolutions. Simulations using the Weather Research and Forecasting model were conducted at grid spacings of 36 km, 12 km, and 4 km for two summers over the conterminous U.S. The convection-permitting simulations at 4 km grid spacing are most skillful in reproducing the observed precipitation spatial distributions and diurnal variability. Notable differences are found between simulations with the traditional Kain-Fritsch (KF) and the scale-aware Grell-Freitas (GF) convection schemes, with the latter more skillful in capturing the nocturnal timing in the Great Plains and North American monsoon regions. The GF scheme also simulates a smoother transition from convective to large-scale precipitation as resolution increases, resulting in reduced sensitivity to model resolution compared to the KF scheme. Nonhydrostatic dynamics has a positive impact on precipitation over complex terrain even at 12 km and 36 km grid spacings. With nudging of the winds toward observations, we show that the conspicuous warm biases in the Southern Great Plains are related to precipitation biases induced by large-scale circulation biases, which are insensitive to model resolution. Overall, notable improvements in simulating summer rainfall and its diurnal variability through convection-permitting modeling and scale-aware parameterizations suggest promising venues for improving climate simulations of water cycle processes.
Three-dimensional micro-scale strain mapping in living biological soft tissues.
Moo, Eng Kuan; Sibole, Scott C; Han, Sang Kuy; Herzog, Walter
2018-04-01
Non-invasive characterization of the mechanical micro-environment surrounding cells in biological tissues at multiple length scales is important for the understanding of the role of mechanics in regulating the biosynthesis and phenotype of cells. However, there is a lack of imaging methods that allow for characterization of the cell micro-environment in three-dimensional (3D) space. The aims of this study were (i) to develop a multi-photon laser microscopy protocol capable of imprinting 3D grid lines onto living tissue at a high spatial resolution, and (ii) to develop image processing software capable of analyzing the resulting microscopic images and performing high resolution 3D strain analyses. Using articular cartilage as the biological tissue of interest, we present a novel two-photon excitation imaging technique for measuring the internal 3D kinematics in intact cartilage at sub-micrometer resolution, spanning length scales from the tissue to the cell level. Using custom image processing software, we provide accurate and robust 3D micro-strain analysis that allows for detailed qualitative and quantitative assessment of the 3D tissue kinematics. This novel technique preserves tissue structural integrity post-scanning, therefore allowing for multiple strain measurements at different time points in the same specimen. The proposed technique is versatile and opens doors for experimental and theoretical investigations on the relationship between tissue deformation and cell biosynthesis. Studies of this nature may enhance our understanding of the mechanisms underlying cell mechano-transduction, and thus, adaptation and degeneration of soft connective tissues. We presented a novel two-photon excitation imaging technique for measuring the internal 3D kinematics in intact cartilage at sub-micrometer resolution, spanning from tissue length scale to cellular length scale. Using a custom image processing software (lsmgridtrack), we provide accurate and robust micro-strain analysis that allowed for detailed qualitative and quantitative assessment of the 3D tissue kinematics. The approach presented here can also be applied to other biological tissues such as meniscus and annulus fibrosus, as well as tissue-engineered tissues for the characterization of their mechanical properties. This imaging technique opens doors for experimental and theoretical investigation on the relationship between tissue deformation and cell biosynthesis. Studies of this nature may enhance our understanding of the mechanisms underlying cell mechano-transduction, and thus, adaptation and degeneration of soft connective tissues. Copyright © 2018 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Alessandri, Andrea; Catalano, Franco; De Felice, Matteo; Van Den Hurk, Bart; Doblas Reyes, Francisco; Boussetta, Souhail; Balsamo, Gianpaolo; Miller, Paul
2016-04-01
The EC-Earth earth system model has been recently developed to include the dynamics of vegetation. In its original formulation, vegetation variability is simply operated by the Leaf Area Index (LAI), which affects climate basically by changing the vegetation physiological resistance to evapotranspiration. This coupling has been found to have only a weak effect on the surface climate modeled by EC-Earth. In reality, the effective sub-grid vegetation fractional coverage will vary seasonally and at interannual time-scales in response to leaf-canopy growth, phenology and senescence. Therefore it affects biophysical parameters such as the albedo, surface roughness and soil field capacity. To adequately represent this effect in EC-Earth, we included an exponential dependence of the vegetation cover on the LAI. By comparing two sets of simulations performed with and without the new variable fractional-coverage parameterization, spanning retrospective predictions at the decadal (5-years), seasonal and sub-seasonal time-scales, we show for the first time a significant multi-scale enhancement of vegetation impacts in climate simulation and prediction over land. Particularly large effects at multiple time scales are shown over boreal winter middle-to-high latitudes over Canada, West US, Eastern Europe, Russia and eastern Siberia due to the implemented time-varying shadowing effect by tree-vegetation on snow surfaces. Over Northern Hemisphere boreal forest regions the improved representation of vegetation cover tends to correct the winter warm biases, improves the climate change sensitivity, the decadal potential predictability as well as the skill of forecasts at seasonal and sub-seasonal time-scales. Significant improvements of the prediction of 2m temperature and rainfall are also shown over transitional land surface hot spots. Both the potential predictability at decadal time-scale and seasonal-forecasts skill are enhanced over Sahel, North American Great Plains, Nordeste Brazil and South East Asia, mainly related to improved performance in the surface evapotranspiration.
WRF nested large-eddy simulations of deep convection during SEAC4RS
NASA Astrophysics Data System (ADS)
Heath, Nicholas Kyle
Deep convection is an important component of atmospheric circulations that affects many aspects of weather and climate. Therefore, improved understanding and realistic simulations of deep convection are critical to both operational and climate forecasts. Large-eddy simulations (LESs) often are used with observations to enhance understanding of convective processes. This study develops and evaluates a nested-LES method using the Weather Research and Forecasting (WRF) model. Our goal is to evaluate the extent to which the WRF nested-LES approach is useful for studying deep convection during a real-world case. The method was applied on 2 September 2013, a day of continental convection having a robust set of ground and airborne data available for evaluation. A three domain mesoscale WRF simulation is run first. Then, the finest mesoscale output (1.35 km grid length) is used to separately drive nested-LES domains with grid lengths of 450 and 150 m. Results reveal that the nested-LES approach reasonably simulates a broad spectrum of observations, from reflectivity distributions to vertical velocity profiles, during the study period. However, reducing the grid spacing does not necessarily improve results for our case, with the 450 m simulation outperforming the 150 m version. We find that simulated updrafts in the 150 m simulation are too narrow to overcome the negative effects of entrainment, thereby generating convection that is weaker than observed. Increasing the sub-grid mixing length in the 150 m simulation leads to deeper, more realistic convection, but comes at the expense of delaying the onset of the convection. Overall, results show that both the 450 m and 150 m simulations are influenced considerably by the choice of sub-grid mixing length used in the LES turbulence closure. Finally, the simulations and observations are used to study the processes forcing strong midlevel cloud-edge downdrafts that were observed on 2 September. Results suggest that these downdrafts are forced by evaporative cooling due to mixing near cloud edge and by vertical perturbation pressure gradient forces acting to restore mass continuity around neighboring updrafts. We conclude that the WRF nested-LES approach provides an effective method for studying deep convection for our real-world case. The method can be used to provide insight into physical processes that are important to understanding observations. The WRF nested-LES approach could be adapted for other case studies in which high-resolution observations are available for validation.
Fuchs, K.; Tittgemeyer, M.; Ryberg, T.; Wenzel, F.; Mooney, W.
2002-01-01
We infer the fine structure of a sub-Moho boundary layer (SMBL) at the top of the lithospheric mantle from high-resolution seismic observations of Peaceful Nuclear Explosions (PNE) on superlong-range profiles in Russia. Densely recorded seismograms permit recognition of previously unknown features of teleseismic propagation of the well known Pn and Sn phases, such as a band of incoherent, scattered, high-frequency seismic energy, developing consistently from station to station, apparent velocities of sub-Moho material, and high-frequency energy to distances of more than 3000 km with a coda band, incoherent at 10 km spacing and yet consistently observed to the end of the profiles. Estimates of the other key elements of the SMBL were obtained by finite difference calculations of wave propagation in elastic 2D models from a systematic grid search through parameter space. The SMBL consists of randomly distributed, mild velocity fluctuations of 2% or schlieren of high aspect ratios (???40) with long horizontal extent (???20 km) and therefore as thin as 0.5 km only; SMBL thickness is 60-100 km. It is suggested that the SMBL is of global significance as the physical base of the platewide observed high-frequency phases Pn and Sn. It is shown that wave propagation in the SMBL waveguide is insensitive to the background velocity distribution on which its schlieren are superimposed. This explains why the Pn and Sn phases traverse geological provinces of various age, heat flow, crustal thickness, and tectonic regimes. Their propagation appears to be independent of age. temperature, pressure, and stress. Dynamic stretching of mantle material during subduction or flow, possibly combined with chemical differentiation have to be considered as scale-forming processes in the upper mantle. However, it is difficult to distinguish with the present sets of Pn/Sn array data whether (and also where) the boundary layer is a frozen-in feature of paleo-processes or whether it is a response to an on-going processes; nevertheless, the derived quantitative estimates of the SMBL properties provide important constraints for any hypothesis on scale-forming processes. Models to be tested by future numerical and field experiments are, for example, repeated subduction-convection stretching of oceanic lithosphere (marble-cake model) and schlieren formation at mid-ocean ridges. It is also proposed that the modeling of the observed blocking of Sn and Pn propagation at active plate margins offers a new tool to study the depth range of tectonics below the crust-mantle boundary. Finally, the deduced schlieren structure of the SMBL closes an important scale gap of three to four orders of magnitude between structural dimensions studied in petrological analysis of mantle samples (xenoliths or outcrop of oceanic lithosphere) and those imaged in classical seismological studies of the lithosphere.
Can we estimate biogeochemical uptake rates in sediments from reach-scale data or vice versa?
NASA Astrophysics Data System (ADS)
Gonzalez-Pinzon, R.; Garayburu-Caruso, V. A.
2017-12-01
Hydrologists and stream ecologists want to understand how reactive transport processes from sub-meter to reach scales aggregate to determine nutrient and carbon export across watersheds. Mesocosm (sub-meter) scale experiments offer the advantage of being tractable and affordable but may be spatially and temporally irrelevant for describing watershed-scale processes. While reach scale experiments sample larger heterogeneities, they provide aggregated information that does not allow for easy detection of hot-spots and hot-moments, and might still be irrelevant for describing watershed processes if they are not conducted under varying flow conditions. We conducted mesocosm (column) and reach-scale experiments along a first-to-eight stream order continuum using nutrient and resazurin tracers to investigate how information collected at the sub-meter scale (mesocosom experiments) compares to that collected at the reach scale, and vice versa. Our work highlights the difficulty of finding useful patterns not only across stream orders (i.e., for the same type of experiment) but also across experiments. Our results offer quantitative perspective on why hydrologists and stream ecologists must depart from the status quo of conducting solute-specific (e.g., only N), site-specific (primarily headwaters) and single-season (mainly summer) experiments to understand controls on nutrient retention.
NASA Astrophysics Data System (ADS)
Normani, S. D.; Sykes, J. F.; Jensen, M. R.
2009-04-01
A high resolution sub-regional scale (84 km2) density-dependent, fracture zone network groundwater flow model with hydromechanical coupling and pseudo-permafrost, was developed from a larger 5734 km2 regional scale groundwater flow model of a Canadian Shield setting in fractured crystalline rock. The objective of the work is to illustrate aspects of regional and sub-regional groundwater flow that are relevant to the long-term performance of a hypothetical nuclear fuel repository. The discrete fracture dual continuum numerical model FRAC3DVS-OPG was used for all simulations. A discrete fracture zone network model delineated from surface features was superimposed onto an 789887 element flow domain mesh. Orthogonal fracture faces (between adjacent finite element grid blocks) were used to best represent the irregular discrete fracture zone network. The crystalline rock between these structural discontinuities was assigned properties characteristic of those reported for the Canadian Shield at the Underground Research Laboratory at Pinawa, Manitoba. Interconnectivity of permeable fracture features is an important pathway for the possibly relatively rapid migration of average water particles and subsequent reduction in residence times. The multiple 121000 year North American continental scale paleoclimate simulations are provided by W.R. Peltier using the University of Toronto Glacial Systems Model (UofT GSM). Values of ice sheet normal stress, and proglacial lake depth from the UofT GSM are applied to the sub-regional model as surface boundary conditions, using a freshwater head equivalent to the normal stress imposed by the ice sheet at its base. Permafrost depth is applied as a permeability reduction to both three-dimensional grid blocks and fractures that lie within the time varying permafrost zone. Two different paleoclimate simulations are applied to the sub-regional model to investigate the effect on the depth of glacial meltwater migration into the subsurface. In addition, different conceptualizations of fracture permeability with depth, and various hydromechanical loading efficiencies are used to investigate glacial meltwater penetration. The importance of density dependent flow, due to pore waters deep in the Canadian Shield with densities of up to 1200 kg/m3 and total dissolved solids concentrations in excess of 300 g/L, is also illustrated. Performance measures used in the assessment include depth of glacial meltwater penetration using a tracer, and mean life expectancy. Consistent with the findings from isotope and geochemical assessments, the analyses support the conclusion that for the discrete fracture zone and matrix properties simulated in this study, glacial meltwaters would not likely impact a deep geologic repository in a crystalline rock setting.
Human skin surface evaluation by image processing
NASA Astrophysics Data System (ADS)
Zhu, Liangen; Zhan, Xuemin; Xie, Fengying
2003-12-01
Human skin gradually lose its tension and becomes very dry as time flies by. Use of cosmetics is effective to prevent skin aging. Recently, there are many choices of products of cosmetics. To show their effects, It is desirable to develop a way to evaluate quantificationally skin surface condition. In this paper, An automatic skin evaluating method is proposed. The skin surface has the pattern called grid-texture. This pattern is composed of the valleys that spread vertically, horizontally, and obliquely and the hills separated by them. Changes of the grid are closely linked to the skin surface condition. They can serve as a good indicator for the skin condition. By measuring the skin grid using digital image processing technologies, we can evaluate skin surface about its aging, health, and alimentary status. In this method, the skin grid is first detected to form a closed net. Then, some skin parameters such as Roughness, tension, scale and gloss can be calculated from the statistical measurements of the net. Through analyzing these parameters, the condition of the skin can be monitored.
Modelling Near-Surface Metallic Clutter Without the Excruciating Pain
NASA Astrophysics Data System (ADS)
Downs, C. M.; Weiss, C. J.; Bach, J.; Williams, J. T.
2016-12-01
An ongoing problem in modeling electromagnetic (EM) interactions with the near-surface and related anthropogenic metal clutter is the large difference in length scale between the clutter dimensions and their resulting EM response. For example, observational evidence shows that cables, pipes and rail lines can have a strong influence far from where they are located, even in situations where these artefacts are volumetrically insignificant over the scale of the model. This poses a significant modeling problem for understanding geohazards in urban environments, for example, because of the very fine numerical discretization required for accurate representation of an artefact embedded in a larger computational domain. We adopt a sub-grid approximation and impose a boundary condition along grid edges to capture the vanishing fields of a perfect conductor. We work in a Cartesian system where the EM fields are solved via finite volumes in the frequency domain in terms of the Lorenz gauged magnetic vector (A) and electric scalar (Phi) potentials. The electric fied is given simply by A-grad(Phi), and set identically to zero along edges of the mesh that coincide with the center of long, slender metallic conductors. A simple extension to bulky artefacts like blocks or slabs involves endowing all such edges in their interior with the same "internal" boundary condition. In essence, we apply the "perfect electric conductor" boundary condition to select edges interior to the modeling domain. We note a few minor numerical consequences of this approach, namely: the zero-E field internal boundary condition destroys the symmetry of the finite volume coefficient matrix; and, the accuracy of the representation of the conducting artefact is restricted by the relatively coarse discretization mesh. The former is overcome with the use of preconditioned bi-conjugate gradient methods instead of the quasi-minimal-residual method. Both are matrix-free iterative solvers - thus avoiding unnecessary storage- and both exhibit generally good convergence for well-posed problems. The latter is more difficult to overcome without either modifying the mesh (potentially degrading the condition number of the coefficient matrix) or with novel mesh sub-gridding. Initial results show qualitative agreement with the expected physics.
New Approaches to Quantifying Transport Model Error in Atmospheric CO2 Simulations
NASA Technical Reports Server (NTRS)
Ott, L.; Pawson, S.; Zhu, Z.; Nielsen, J. E.; Collatz, G. J.; Gregg, W. W.
2012-01-01
In recent years, much progress has been made in observing CO2 distributions from space. However, the use of these observations to infer source/sink distributions in inversion studies continues to be complicated by difficulty in quantifying atmospheric transport model errors. We will present results from several different experiments designed to quantify different aspects of transport error using the Goddard Earth Observing System, Version 5 (GEOS-5) Atmospheric General Circulation Model (AGCM). In the first set of experiments, an ensemble of simulations is constructed using perturbations to parameters in the model s moist physics and turbulence parameterizations that control sub-grid scale transport of trace gases. Analysis of the ensemble spread and scales of temporal and spatial variability among the simulations allows insight into how parameterized, small-scale transport processes influence simulated CO2 distributions. In the second set of experiments, atmospheric tracers representing model error are constructed using observation minus analysis statistics from NASA's Modern-Era Retrospective Analysis for Research and Applications (MERRA). The goal of these simulations is to understand how errors in large scale dynamics are distributed, and how they propagate in space and time, affecting trace gas distributions. These simulations will also be compared to results from NASA's Carbon Monitoring System Flux Pilot Project that quantified the impact of uncertainty in satellite constrained CO2 flux estimates on atmospheric mixing ratios to assess the major factors governing uncertainty in global and regional trace gas distributions.
Final Report for Department of Energy Project DE-SC0012198
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lucchese, Robert; Poliakoff, Erwin; Trallero-Herrero, Carlos
The study of the motion of atoms in molecules is important to understanding many areas of physical and life sciences. Such motion occurs on many different times scales, with electronic motion occurring on a sub-femtosecond time scale, and simple vibrational motion in tens to hundreds of femtoseconds. One way to follow such processes in real time is by the use of short-pulsed lasers, and in particular by studying time-resolved photoionization and the related process of high-harmonic generation (HHG). Thus there has been much effort to develop the tools necessary to probe molecular systems using short pulse lasers and understanding themore » sensitivity of the different possible probes to the time dependent geometric structure as well as the electronic structure of molecules. Our research has particularly focused on the connection between high-field processes and the more traditional weak field photoionization processes. Strong field and weak field processes can be connected through models that involve the same matrix elements. We have demonstrated in our study of HHG from SF6 that the spectrum is sensitive to the interplay between the angular dependence of the ionization step and recombination step. In our study of rescattering spectroscopy, we have shown that with a combination of experiment and theory, we can use this high-field spectroscopy to determine molecular structure in molecules such as C2H4. We have also developed new computational tools based on overset grids to enable studies on larger molecular systems which use much more robust numerical approaches so that the resulting code can be a tool that non-specialists can use to study related systems.« less
Dong, Nan; Yang, Xiaohuan; Cai, Hongyan; Xu, Fengjiao
2017-01-01
The research on the grid size suitability is important to provide improvement in accuracies of gridded population distribution. It contributes to reveal the actual spatial distribution of population. However, currently little research has been done in this area. Many well-modeled gridded population dataset are basically built at a single grid scale. If the grid cell size is not appropriate, it will result in spatial information loss or data redundancy. Therefore, in order to capture the desired spatial variation of population within the area of interest, it is necessary to conduct research on grid size suitability. This study summarized three expressed levels to analyze grid size suitability, which include location expressed level, numeric information expressed level, and spatial relationship expressed level. This study elaborated the reasons for choosing the five indexes to explore expression suitability. These five indexes are consistency measure, shape index rate, standard deviation of population density, patches diversity index, and the average local variance. The suitable grid size was determined by constructing grid size-indicator value curves and suitable grid size scheme. Results revealed that the three expressed levels on 10m grid scale are satisfying. And the population distribution raster data with 10m grid size provide excellent accuracy without loss. The 10m grid size is recommended as the appropriate scale for generating a high-quality gridded population distribution in our study area. Based on this preliminary study, it indicates the five indexes are coordinated with each other and reasonable and effective to assess grid size suitability. We also suggest choosing these five indexes in three perspectives of expressed level to carry out the research on grid size suitability of gridded population distribution.
Dong, Nan; Yang, Xiaohuan; Cai, Hongyan; Xu, Fengjiao
2017-01-01
The research on the grid size suitability is important to provide improvement in accuracies of gridded population distribution. It contributes to reveal the actual spatial distribution of population. However, currently little research has been done in this area. Many well-modeled gridded population dataset are basically built at a single grid scale. If the grid cell size is not appropriate, it will result in spatial information loss or data redundancy. Therefore, in order to capture the desired spatial variation of population within the area of interest, it is necessary to conduct research on grid size suitability. This study summarized three expressed levels to analyze grid size suitability, which include location expressed level, numeric information expressed level, and spatial relationship expressed level. This study elaborated the reasons for choosing the five indexes to explore expression suitability. These five indexes are consistency measure, shape index rate, standard deviation of population density, patches diversity index, and the average local variance. The suitable grid size was determined by constructing grid size-indicator value curves and suitable grid size scheme. Results revealed that the three expressed levels on 10m grid scale are satisfying. And the population distribution raster data with 10m grid size provide excellent accuracy without loss. The 10m grid size is recommended as the appropriate scale for generating a high-quality gridded population distribution in our study area. Based on this preliminary study, it indicates the five indexes are coordinated with each other and reasonable and effective to assess grid size suitability. We also suggest choosing these five indexes in three perspectives of expressed level to carry out the research on grid size suitability of gridded population distribution. PMID:28122050
NASA Astrophysics Data System (ADS)
Skamarock, W. C.
2015-12-01
One of the major problems in atmospheric model applications is the representation of deep convection within the models; explicit simulation of deep convection on fine meshes performs much better than sub-grid parameterized deep convection on coarse meshes. Unfortunately, the high cost of explicit convective simulation has meant it has only been used to down-scale global simulations in weather prediction and regional climate applications, typically using traditional one-way interactive nesting technology. We have been performing real-time weather forecast tests using a global non-hydrostatic atmospheric model (the Model for Prediction Across Scales, MPAS) that employs a variable-resolution unstructured Voronoi horizontal mesh (nominally hexagons) to span hydrostatic to nonhydrostatic scales. The smoothly varying Voronoi mesh eliminates many downscaling problems encountered using traditional one- or two-way grid nesting. Our test weather forecasts cover two periods - the 2015 Spring Forecast Experiment conducted at the NOAA Storm Prediction Center during the month of May in which we used a 50-3 km mesh, and the PECAN field program examining nocturnal convection over the US during the months of June and July in which we used a 15-3 km mesh. An important aspect of this modeling system is that the model physics be scale-aware, particularly the deep convection parameterization. These MPAS simulations employ the Grell-Freitas scale-aware convection scheme. Our test forecasts show that the scheme produces a gradual transition in the deep convection, from the deep unstable convection being handled entirely by the convection scheme on the coarse mesh regions (dx > 15 km), to the deep convection being almost entirely explicit on the 3 km NA region of the meshes. We will present results illustrating the performance of critical aspects of the MPAS model in these tests.
Immersion and dry scanner extensions for sub-10nm production nodes
NASA Astrophysics Data System (ADS)
Weichselbaum, Stefan; Bornebroek, Frank; de Kort, Toine; Droste, Richard; de Graaf, Roelof F.; van Ballegoij, Rob; Botter, Herman; McLaren, Matthew G.; de Boeij, Wim P.
2015-03-01
Progressing towards the 10nm and 7nm imaging node, pattern-placement and layer-to-layer overlay requirements keep on scaling down and drives system improvements in immersion (ArFi) and dry (ArF/KrF) scanners. A series of module enhancements in the NXT platform have been introduced; among others, the scanner is equipped with exposure stages with better dynamics and thermal control. Grid accuracy improvements with respect to calibration, setup, stability, and layout dependency tighten MMO performance and enable mix and match scanner operation. The same platform improvements also benefit focus control. Improvements in detectability and reproducibility of low contrast alignment marks enhance the alignment solution window for 10nm logic processes and beyond. The system's architecture allows dynamic use of high-order scanner optimization based on advanced actuators of projection lens and scanning stages. This enables a holistic optimization approach for the scanner, the mask, and the patterning process. Productivity scanner design modifications esp. stage speeds and optimization in metrology schemes provide lower layer costs for customers using immersion lithography as well as conventional dry technology. Imaging, overlay, focus, and productivity data is presented, that demonstrates 10nm and 7nm node litho-capability for both (immersion & dry) platforms.
NASA Astrophysics Data System (ADS)
Rollinson, C.; Simkins, J.; Fer, I.; Desai, A. R.; Dietze, M.
2017-12-01
Simulations of ecosystem dynamics and comparisons with empirical data require accurate, continuous, and often sub-daily meteorology records that are spatially aligned to the scale of the empirical data. A wealth of meteorology data for the past, present, and future is available through site-specific observations, modern reanalysis products, and gridded GCM simulations. However, these products are mismatched in spatial and temporal resolution, often with both different means and seasonal patterns. We have designed and implemented a two-step meteorological downscaling and ensemble generation method that combines multiple meteorology data products through debiasing and temporal downscaling protocols. Our methodology is designed to preserve the covariance among seven meteorological variables for use as drivers in ecosystem model simulations: temperature, precipitation, short- and longwave radiation, surface pressure, humidity, and wind. Furthermore, our method propagates uncertainty through the downscaling process and results in ensembles of meteorology that can be compared to paleoclimate reconstructions and used to analyze the effects of both high- and low-frequency climate anomalies on ecosystem dynamics. Using a multiple linear regression approach, we have combined hourly, 0.125-degree gridded data from the NLDAS (1980-present) with CRUNCEP (1901-2010) and CMIP5 historical (1850-2005), past millennium (850-1849), and future (1950-2100) GCM simulations. This has resulted in an ensemble of continuous, hourly-resolved meteorology from from the paleo era into the future with variability in weather events as well as low-frequency climatic changes. We investigate the influence of extreme sub-daily weather phenomena versus long-term climatic changes in an ensemble of ecosystem models that range in atmospheric and biological complexity. Through data assimilation with paleoclimate reconstructions of past climate, we can improve data-model comparisons using observations of vegetation change from the past 1200 years. Accounting for driver uncertainty in model evaluation can help determine the relative influence of structural versus parameterization errors in ecosystem modelings.
GreenView and GreenLand Applications Development on SEE-GRID Infrastructure
NASA Astrophysics Data System (ADS)
Mihon, Danut; Bacu, Victor; Gorgan, Dorian; Mészáros, Róbert; Gelybó, Györgyi; Stefanut, Teodor
2010-05-01
The GreenView and GreenLand applications [1] have been developed through the SEE-GRID-SCI (SEE-GRID eInfrastructure for regional eScience) FP7 project co-funded by the European Commission [2]. The development of environment applications is a challenge for Grid technologies and software development methodologies. This presentation exemplifies the development of the GreenView and GreenLand applications over the SEE-GRID infrastructure by the Grid Application Development Methodology [3]. Today's environmental applications are used in vary domains of Earth Science such as meteorology, ground and atmospheric pollution, ground metal detection or weather prediction. These applications run on satellite images (e.g. Landsat, MERIS, MODIS, etc.) and the accuracy of output results depends mostly of the quality of these images. The main drawback of such environmental applications regards the need of computation power and storage power (some images are almost 1GB in size), in order to process such a large data volume. Actually, almost applications requiring high computation resources have approached the migration onto the Grid infrastructure. This infrastructure offers the computing power by running the atomic application components on different Grid nodes in sequential or parallel mode. The middleware used between the Grid infrastructure and client applications is ESIP (Environment Oriented Satellite Image Processing Platform), which is based on gProcess platform [4]. In its current format, gProcess is used for launching new processes on the Grid nodes, but also for monitoring the execution status of these processes. This presentation highlights two case studies of Grid based environmental applications, GreenView and GreenLand [5]. GreenView is used in correlation with MODIS (Moderate Resolution Imaging Spectroradiometer) satellite images and meteorological datasets, in order to produce pseudo colored temperature and vegetation maps for different geographical CEE (Central Eastern Europe) regions. On the other hand, GreenLand is used for generating maps for different vegetation indexes (e.g. NDVI, EVI, SAVI, GEMI) based on Landsat satellite images. Both applications are using interpolation and random value generation algorithms, but also specific formulas for computing vegetation index values. The GreenView and GreenLand applications have been experimented over the SEE-GRID infrastructure and the performance evaluation is reported in [6]. The improvement of the execution time (obtained through a better parallelization of jobs), the extension of geographical areas to other parts of the Earth, and new user interaction techniques on spatial data and large set of satellite images are the goals of the future work. References [1] GreenView application on Wiki, http://wiki.egee-see.org/index.php/GreenView [2] SEE-GRID-SCI Project, http://www.see-grid-sci.eu/ [3] Gorgan D., Stefanut T., Bâcu V., Mihon D., Grid based Environment Application Development Methodology, SCICOM, 7th International Conference on "Large-Scale Scientific Computations", 4-8 June, 2009, Sozopol, Bulgaria, (To be published by Springer), (2009). [4] Gorgan D., Bacu V., Stefanut T., Rodila D., Mihon D., Grid based Satellite Image Processing Platform for Earth Observation Applications Development. IDAACS'2009 - IEEE Fifth International Workshop on "Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications", 21-23 September, Cosenza, Italy, IEEE Published in Computer Press, 247-252 (2009). [5] Mihon D., Bacu V., Stefanut T., Gorgan D., "Grid Based Environment Application Development - GreenView Application". ICCP2009 - IEEE 5th International Conference on Intelligent Computer Communication and Processing, 27 Aug, 2009 Cluj-Napoca. Published by IEEE Computer Press, pp. 275-282 (2009). [6] Danut Mihon, Victor Bacu, Dorian Gorgan, Róbert Mészáros, Györgyi Gelybó, Teodor Stefanut, Practical Considerations on the GreenView Application Development and Execution over SEE-GRID. SEE-GRID-SCI User Forum, 9-10 Dec 2009, Bogazici University, Istanbul, Turkey, ISBN: 978-975-403-510-0, pp. 167-175 (2009).
Novel three-dimensional dandelion-like TiO{sub 2} structure with high photocatalytic activity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bai Xuelian; Xie Bin; Pan Nan
2008-03-15
Pure rutile phase crystalline TiO{sub 2} powder with novel 3D dandelion-like structure was synthesized by using a facile hydrothermal method with TiCl{sub 3} as the main starting material. In such a 3D structure, the nanometer-scale construction elements aggregate together and form a micrometer-scale artificial unit. The typical 3D dandelion structure has an average diameter of 1.5-2 {mu}m and is packed radially by nanorods with [001] preference growth direction. Each individual nanorod is hundreds of nanometers in length, and tens of nanometers in diameter. The obtained 3D dandelion-like TiO{sub 2} powder has a high photocatalytic activity, which is equivalent to thatmore » of the commercial available P25 titania powder. Mechanisms of the formation of the dandelion-like structure were also discussed. A different oxidation process of Ti(III) to Ti(IV) during hydrothermal was suggested. - Graphical abstract: Rutile-phase TiO{sub 2} powders with novel 3D dandelion-like structures were synthesized. This kind of 3D artificial hierarchical titania structure has the advantage of reserving the novel nanometer-scale properties while providing us the easiness of storing and handling as we routinely enjoyed for the micrometer-scale materials. A different oxidation process of Ti(III) to Ti(IV) during hydrothermal process was suggested.« less
FY2017 Electrification Annual Progress Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
During fiscal year 2017 (FY 2017), the U.S. Department of Energy (DOE) Vehicle Technologies Office (VTO) funded early stage research & development (R&D) projects that address Batteries and Electrification of the U.S. transportation sector. The VTO Electrification Sub-Program is composed of Electric Drive Technologies, and Grid Integration activities. The Electric Drive Technologies group conducts R&D projects that advance Electric Motors and Power Electronics technologies. The Grid and Charging Infrastructure group conducts R&D projects that advance Grid Modernization and Electric Vehicle Charging technologies. This document presents a brief overview of the Electrification Sub-Program and progress reports for its R&D projects. Eachmore » of the progress reports provide a project overview and highlights of the technical results that were accomplished in FY 2017.« less
Spatial Variability of Snowpack Properties On Small Slopes
NASA Astrophysics Data System (ADS)
Pielmeier, C.; Kronholm, K.; Schneebeli, M.; Schweizer, J.
The spatial variability of alpine snowpacks is created by a variety of parameters like deposition, wind erosion, sublimation, melting, temperature, radiation and metamor- phism of the snow. Spatial variability is thought to strongly control the avalanche initi- ation and failure propagation processes. Local snowpack measurements are currently the basis for avalanche warning services and there exist contradicting hypotheses about the spatial continuity of avalanche active snow layers and interfaces. Very little about the spatial variability of the snowpack is known so far, therefore we have devel- oped a systematic and objective method to measure the spatial variability of snowpack properties, layering and its relation to stability. For a complete coverage, the analysis of the spatial variability has to entail all scales from mm to km. In this study the small to medium scale spatial variability is investigated, i.e. the range from centimeters to tenths of meters. During the winter 2000/2001 we took systematic measurements in lines and grids on a flat snow test field with grid distances from 5 cm to 0.5 m. Fur- thermore, we measured systematic grids with grid distances between 0.5 m and 2 m in undisturbed flat fields and on small slopes above the tree line at the Choerbschhorn, in the region of Davos, Switzerland. On 13 days we measured the spatial pattern of the snowpack stratigraphy with more than 110 snow micro penetrometer measure- ments at slopes and flat fields. Within this measuring grid we placed 1 rutschblock and 12 stuffblock tests to measure the stability of the snowpack. With the large num- ber of measurements we are able to use geostatistical methods to analyse the spatial variability of the snowpack. Typical correlation lengths are calculated from semivari- ograms. Discerning the systematic trends from random spatial variability is analysed using statistical models. Scale dependencies are shown and recurring scaling patterns are outlined. The importance of the small and medium scale spatial variability for the larger (kilometer) scale spatial variability as well as for the avalanche formation are discussed. Finally, an outlook on spatial models for the snowpack variability is given.
Large-scale production of anhydrous nitric acid and nitric acid solutions of dinitrogen pentoxide
Harrar, Jackson E.; Quong, Roland; Rigdon, Lester P.; McGuire, Raymond R.
2001-01-01
A method and apparatus are disclosed for a large scale, electrochemical production of anhydrous nitric acid and N.sub.2 O.sub.5. The method includes oxidizing a solution of N.sub.2 O.sub.4 /aqueous-HNO.sub.3 at the anode, while reducing aqueous HNO.sub.3 at the cathode, in a flow electrolyzer constructed of special materials. N.sub.2 O.sub.4 is produced at the cathode and may be separated and recycled as a feedstock for use in the anolyte. The process is controlled by regulating the electrolysis current until the desired products are obtained. The chemical compositions of the anolyte and catholyte are monitored by measurement of the solution density and the concentrations of N.sub.2 O.sub.4.
NASA Astrophysics Data System (ADS)
Ramsdale, Jason; Balme, Matthew; Conway, Susan
2015-04-01
An International Space Science Institute (ISSI) team project has been convened to study the northern plains of Mars. The northern plains are younger and at lower elevation than the majority of the martian surface and are thought to be the remnants of an ancient ocean. Understanding the surface geology and geomorphology of the Northern Plains is complex, because the surface has been subtly modified many times, making traditional unit-boundaries hard to define. Our ISSI team project aims to answer the following questions: 1) "What is the distribution of ice-related landforms in the northern plains, and can it be related to distinct latitude bands or different geological or geomorphological units?" 2) "What is the relationship between the latitude dependent mantle (LDM; a draping unit believed to comprise of ice and dust thought to be deposited under periods of high axial obliquity) and (i) landforms indicative of ground ice, and (ii) other geological units in the northern plains?" 3) "What are the distributions and associations of recent landforms indicative of thaw of ice or snow?" With increasing coverage of high-resolution images of the surface of we are able to identify increasing numbers and varieties of small-scale landforms on Mars. Many such landforms are too small to represent on regional maps, yet determining their presence or absence across large areas can form the observational basis for developing hypotheses on the nature and history of an area. The combination of improved spatial resolution with near-continuous coverage increases the time required to analyse the data. This becomes problematic when attempting regional or global-scale studies of metre-scale landforms. Here, we describe an approach to mapping small features across large areas. Rather than traditional mapping with points, lines and polygons, we used a grid "tick box" approach to locate specific landforms. The mapping strips were divided into 15×150 grid of squares, each approximately 20×20 km, for each study area. Orbital images at 6-15m/pix were then viewed systematically for each grid square and the presence or absence of each of the basic suite of landforms recorded. The landforms were recorded as being either "present", "dominant", "possible", or "absent" in each grid square. The result is a series of coarse-resolution "rasters" showing the distribution of the different types of landforms across the strip. We have found this approach to be efficient, scalable and appropriate for teams of people mapping remotely. It is easily scalable because, carrying the "absent" values forward to finer grids from the larger grids would mean only areas with positive values for that landform would need to be examined to increase the resolution for the whole strip. As each sub-grid only requires the presence or absence of a landform ascertaining, it therefore removes an individual's decision as to where to draw boundaries, making the method efficient and repeatable.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Guba, O.; Taylor, M. A.; Ullrich, P. A.
2014-11-27
We evaluate the performance of the Community Atmosphere Model's (CAM) spectral element method on variable-resolution grids using the shallow-water equations in spherical geometry. We configure the method as it is used in CAM, with dissipation of grid scale variance, implemented using hyperviscosity. Hyperviscosity is highly scale selective and grid independent, but does require a resolution-dependent coefficient. For the spectral element method with variable-resolution grids and highly distorted elements, we obtain the best results if we introduce a tensor-based hyperviscosity with tensor coefficients tied to the eigenvalues of the local element metric tensor. The tensor hyperviscosity is constructed so that, formore » regions of uniform resolution, it matches the traditional constant-coefficient hyperviscosity. With the tensor hyperviscosity, the large-scale solution is almost completely unaffected by the presence of grid refinement. This later point is important for climate applications in which long term climatological averages can be imprinted by stationary inhomogeneities in the truncation error. We also evaluate the robustness of the approach with respect to grid quality by considering unstructured conforming quadrilateral grids generated with a well-known grid-generating toolkit and grids generated by SQuadGen, a new open source alternative which produces lower valence nodes.« less
Guba, O.; Taylor, M. A.; Ullrich, P. A.; ...
2014-06-25
We evaluate the performance of the Community Atmosphere Model's (CAM) spectral element method on variable resolution grids using the shallow water equations in spherical geometry. We configure the method as it is used in CAM, with dissipation of grid scale variance implemented using hyperviscosity. Hyperviscosity is highly scale selective and grid independent, but does require a resolution dependent coefficient. For the spectral element method with variable resolution grids and highly distorted elements, we obtain the best results if we introduce a tensor-based hyperviscosity with tensor coefficients tied to the eigenvalues of the local element metric tensor. The tensor hyperviscosity ismore » constructed so that for regions of uniform resolution it matches the traditional constant coefficient hyperviscsosity. With the tensor hyperviscosity the large scale solution is almost completely unaffected by the presence of grid refinement. This later point is important for climate applications where long term climatological averages can be imprinted by stationary inhomogeneities in the truncation error. We also evaluate the robustness of the approach with respect to grid quality by considering unstructured conforming quadrilateral grids generated with a well-known grid-generating toolkit and grids generated by SQuadGen, a new open source alternative which produces lower valence nodes.« less
NASA Technical Reports Server (NTRS)
Dominguez, Anthony; Kleissl, Jan P.; Luvall, Jeffrey C.
2011-01-01
Large-eddy Simulation (LES) was used to study convective boundary layer (CBL) flow through suburban regions with both large and small scale heterogeneities in surface temperature. Constant remotely sensed surface temperatures were applied at the surface boundary at resolutions of 10 m, 90 m, 200 m, and 1 km. Increasing the surface resolution from 1 km to 200 m had the most significant impact on the mean and turbulent flow characteristics as the larger scale heterogeneities became resolved. While previous studies concluded that scales of heterogeneity much smaller than the CBL inversion height have little impact on the CBL characteristics, we found that further increasing the surface resolution (resolving smaller scale heterogeneities) results in an increase in mean surface heat flux, thermal blending height, and potential temperature profile. The results of this study will help to better inform sub-grid parameterization for meso-scale meteorological models. The simulation tool developed through this study (combining LES and high resolution remotely sensed surface conditions) is a significant step towards future studies on the micro-scale meteorology in urban areas.
NASA Astrophysics Data System (ADS)
Nuber, André; Manukyan, Edgar; Maurer, Hansruedi
2014-05-01
Conventional methods of interpreting seismic data rely on filtering and processing limited portions of the recorded wavefield. Typically, either reflections, refractions or surface waves are considered in isolation. Particularly in near-surface engineering and environmental investigations (depths less than, say 100 m), these wave types often overlap in time and are difficult to separate. Full waveform inversion is a technique that seeks to exploit and interpret the full information content of the seismic records without the need for separating events first; it yields models of the subsurface at sub-wavelength resolution. We use a finite element modelling code to solve the 2D elastic isotropic wave equation in the frequency domain. This code is part of a Gauss-Newton inversion scheme which we employ to invert for the P- and S-wave velocities as well as for density in the subsurface. For shallow surface data the use of an elastic forward solver is essential because surface waves often dominate the seismograms. This leads to high sensitivities (partial derivatives contained in the Jacobian matrix of the Gauss-Newton inversion scheme) and thus large model updates close to the surface. Reflections from deeper structures may also include useful information, but the large sensitivities of the surface waves often preclude this information from being fully exploited. We have developed two methods that balance the sensitivity distributions and thus may help resolve the deeper structures. The first method includes equilibrating the columns of the Jacobian matrix prior to every inversion step by multiplying them with individual scaling factors. This is expected to also balance the model updates throughout the entire subsurface model. It can be shown that this procedure is mathematically equivalent to balancing the regularization weights of the individual model parameters. A proper choice of the scaling factors required to balance the Jacobian matrix is critical. We decided to normalise the columns of the Jacobian based on their absolute column sum, but defining an upper threshold for the scaling factors. This avoids particularly small and therefore insignificant sensitivities being over-boosted, which would produce unstable results. The second method proposed includes adjusting the inversion cell size with depth. Multiple cells of the forward modelling grid are merged to form larger inversion cells (typical ratios between forward and inversion cells are in the order of 1:100). The irregular inversion grid is adapted to the expected resolution power of full waveform inversion. Besides stabilizing the inversion, this approach also reduces the number of model parameters to be recovered. Consequently, the computational costs and the memory consumption are reduced significantly. This is particularly critical when Gauss-Newton type inversion schemes are employed. Extensive tests with synthetic data demonstrated that both methods stabilise the inversion and improve the inversion results. The two methods have some redundancy, which can be seen when both are applied simultaneously, that is, when scaling of the Jacobian matrix is applied to an irregular inversion grid. The calculated scaling factors are quite balanced and span a much smaller range than in the case of a regular inversion grid.
Bench-scale reactors were used to test a novel thermo-oxidation process on municipal wastewater treatment plant (WWTP) waste activated sludge (WAS) using hydrogen peroxide (H2O2) to achieve a Class A sludge product appropriate for land application. Reactor ...
Feenstra, Roeland; Christen, David; Paranthaman, Mariappan
1999-01-01
A method is disclosed for fabricating YBa.sub.2 Cu.sub.3 O.sub.7 superconductor layers with the capability of carrying large superconducting currents on a metallic tape (substrate) supplied with a biaxially textured oxide buffer layer. The method represents a simplification of previously established techniques and provides processing requirements compatible with scale-up to long wire (tape) lengths and high processing speeds. This simplification has been realized by employing the BaF.sub.2 method to grow a YBa.sub.2 Cu.sub.3 O.sub.7 film on a metallic substrate having a biaxially textured oxide buffer layer.
NASA Technical Reports Server (NTRS)
Heineck, James; Schairer, Edward; Ramasamy, Manikandan; Roozeboom, Nettie
2016-01-01
This paper describes simultaneous optical measurements of a sub-scale helicopter rotor in the U.S. Army Hover Chamber at NASA Ames Research Center. The measurements included thermal imaging of the rotor blades to detect boundary layer transition; retro-reflective background-oriented schlieren (RBOS) to visualize vortices; and stereo photogrammetry to measure displacements of the rotor blades, to compute spatial coordinates of the vortices from the RBOS data, and to map the thermal imaging data to a three-dimensional surface grid. The test also included an exploratory effort to measure flow near the rotor tip by tomographic particle image velocimetry (tomo PIV)an effort that yielded valuable experience but little data. The thermal imaging was accomplished using an image-derotation method that allowed long integration times without image blur. By mapping the thermal image data to a surface grid it was possible to accurately locate transition in spatial coordinates along the length of the rotor blade.
A numerical solution method for acoustic radiation from axisymmetric bodies
NASA Technical Reports Server (NTRS)
Caruthers, John E.; Raviprakash, G. K.
1995-01-01
A new and very efficient numerical method for solving equations of the Helmholtz type is specialized for problems having axisymmetric geometry. It is then demonstrated by application to the classical problem of acoustic radiation from a vibrating piston set in a stationary infinite plane. The method utilizes 'Green's Function Discretization', to obtain an accurate resolution of the waves using only 2-3 points per wave. Locally valid free space Green's functions, used in the discretization step, are obtained by quadrature. Results are computed for a range of grid spacing/piston radius ratios at a frequency parameter, omega R/c(sub 0), of 2 pi. In this case, the minimum required grid resolution appears to be fixed by the need to resolve a step boundary condition at the piston edge rather than by the length scale imposed by the wave length of the acoustic radiation. It is also demonstrated that a local near-field radiation boundary procedure allows the domain to be truncated very near the radiating source with little effect on the solution.
NASA Astrophysics Data System (ADS)
Hill, James C.; Liu, Zhenping; Fox, Rodney O.; Passalacqua, Alberto; Olsen, Michael G.
2015-11-01
The multi-inlet vortex reactor (MIVR) has been developed to provide a platform for rapid mixing in the application of flash nanoprecipitation (FNP) for manufacturing functional nanoparticles. Unfortunately, commonly used RANS methods are unable to accurately model this complex swirling flow. Large eddy simulations have also been problematic, as expensive fine grids to accurately model the flow are required. These dilemmas led to the strategy of applying a Delayed Detached Eddy Simulation (DDES) method to the vortex reactor. In the current work, the turbulent swirling flow inside a scaled-up MIVR has been investigated by using a dynamic DDES model. In the DDES model, the eddy viscosity has a form similar to the Smagorinsky sub-grid viscosity in LES and allows the implementation of a dynamic procedure to determine its coefficient. The complex recirculating back flow near the reactor center has been successfully captured by using this dynamic DDES model. Moreover, the simulation results are found to agree with experimental data for mean velocity and Reynolds stresses.
NASA Astrophysics Data System (ADS)
Powell, R. L.; Goulden, M.; Peterson, S.; Roberts, D. A.; Still, C. J.
2015-12-01
Temperature is a primary environmental control on biological systems and processes at a range of spatial and temporal scales, from controlling biochemical processes such as photosynthesis to influencing continental-scale species distribution. The Landsat satellite series provides a long record (since the mid-1980s) of relatively high spatial resolution thermal infrared (TIR) imagery, from which we derive land surface temperature (LST) grids. Here, we investigate fine spatial resolution factors that influence Landsat-derived LST over a spectrally and spatially heterogeneous landscape. We focus on paired sites (inside/outside a 1994 fire scar) within a pinyon-juniper scrubland in Southern California. The sites have nearly identical micro-meteorology and vegetation species composition, but distinctly different vegetation abundance and structure. The tower at the unburned site includes a number of in-situ imaging tools to quantify vegetation properties, including a thermal camera on a pan-tilt mount, allowing hourly characterization of landscape component temperatures (e.g., sunlit canopy, bare soil, leaf litter). We use these in-situ measurements to assess the impact of fine-scale landscape heterogeneity on estimates of LST, including sensitivity to (i) the relative abundance of component materials, (ii) directional effects due to solar and viewing geometry, (iii) duration of sunlit exposure for each compositional type, and (iv) air temperature. To scale these properties to Landsat spatial resolution (~100-m), we characterize the sub-pixel composition of landscape components (in addition to shade) by applying spectral mixture analysis (SMA) to the Landsat Operational Land Imager (OLI) spectral bands and test the sensitivity of the relationships established with the in-situ data at this coarser scale. The effects of vegetation abundance and cover height versus other controls on satellite-derived estimates of LST will be assessed by comparing estimates at the burned vs. unburned sites across multiple seasons (~30 dates).
Semihard processes with BLM renormalization scale setting
DOE Office of Scientific and Technical Information (OSTI.GOV)
Caporale, Francesco; Ivanov, Dmitry Yu.; Murdaca, Beatrice
We apply the BLM scale setting procedure directly to amplitudes (cross sections) of several semihard processes. It is shown that, due to the presence of β{sub 0}-terms in the NLA results for the impact factors, the obtained optimal renormalization scale is not universal, but depends both on the energy and on the process in question. We illustrate this general conclusion considering the following semihard processes: (i) inclusive production of two forward high-p{sub T} jets separated by large interval in rapidity (Mueller-Navelet jets); (ii) high-energy behavior of the total cross section for highly virtual photons; (iii) forward amplitude of the productionmore » of two light vector mesons in the collision of two virtual photons.« less
Exploring Crisis Management in U.S. Small Businesses
NASA Astrophysics Data System (ADS)
Williams, Jon
As a critical infrastructure, the US electricity grid supplies electricity to 340 million people within eight separate regions. The power infrastructure is vulnerable to many types of disasters capable of severing supplies of electricity. The impact on the employees and communities when small- and medium-size enterprises are shut down due to disasters can be severe. The purpose of the quantitative comparative study was to explore small- and medium-size enterprises crisis management strategies in the case of power infrastructure vulnerabilities. Perceptions of small business leaders were probed about crisis management planning relevant to three secondary factors: prior experience of crises, threat perceptions, and planning self-efficacy. Participants completed an adapted questionnaire instrument based on a five-point Likert scale for six sub-factors including resilience through planning, financial impact, operational crisis management, the perfect storm, the aftermath of survival, and atrophy. The instrument also measured three additional factors to include, prior experience of crises, threat perceptions, and planning self-efficacy, across seven types of crises. The results of this study indicated that of the 276 respondents, 104 had no crisis plans, but 172 did have crisis plans. Of those who had implemented crisis plans, 19% had specific provisions to address power outages or attacks on the electrical grid. Of the respondents who had not planned for power outages nor experienced significant losses of power, a statistically significant number acknowledged an external threat to their business. The majority of respondents indicated that long-term planning was related to resilience; however, the migration of crisis understanding into the planning process or implementation was not implemented. This heightened awareness of potential crises without the corresponding development and implementation of mitigation crisis plans requires additional research to understand drivers effecting the decision making process with crisis managers.
Analysis of TRMM-LIS Lightning and Related Microphysics Using a Cell-Scale Database
NASA Technical Reports Server (NTRS)
Leroy, Anita; Petersen, Walter A.
2010-01-01
Previous studies of tropical lightning activity using Tropical Rainfall Measurement Mission (TRMM) Lightning Imaging Sensor (LIS) data performed analyses of lightning behavior over mesoscale "feature" scales or over uniform grids. In order to study lightning and the governing ice microphysics intrinsic to thunderstorms at a more process-specific scale (i.e., the scale over which electrification processes and lightning occur in a "unit" thunderstorm), a new convective cell-scale database was developed by analyzing and refining the University of Utah's Precipitation Features database and retaining precipitation data parameters computed from the TRMM precipitation radar (PR), microwave imager (TMI) and LIS instruments. The resulting data base was to conduct a limited four-year study of tropical continental convection occurring over the Amazon Basin, Congo, Maritime Continent and the western Pacific Ocean. The analysis reveals expected strong correlations between lightning flash counts per cell and ice proxies, such as ice water path, minimum and average 85GHz brightness temperatures, and 18dBz echo top heights above the freezing level in all regimes, as well as regime-specific relationships between lighting flash counts and PR-derived surface rainfall rates. Additionally, radar CFADs were used to partition the 3D structure of cells in each regime at different flash counts. The resulting cell-scale analyses are compared to previous mesoscale feature and gridded studies wherever possible.
The GILDA t-Infrastructure: grid training activities in Africa and future opportunities
NASA Astrophysics Data System (ADS)
Ardizzone, V.; Barbera, R.; Ciuffo, L.; Giorgio, E.
2009-04-01
Scientists, educators, and students from many parts of the worlds are not able to take advantage of ICT because the digital divide is growing and prevents less developed countries to exploit its benefits. Instead of becoming more empowered and involved in worldwide developments, they are becoming increasingly marginalised as the world of education and science becomes increasingly Internet-dependent. The Grid Infn Laboratory for Dissemination Activities (GILDA) spreads since almost five years the awareness of Grid technology to a large audience, training new communities and fostering new organisations to provide resources. The knowledge dissemination process guided by the training activities is a key factor to ensure that all users can fully understand the characteristics of the Grid services offered by large existing e-Infrastructure. GILDA is becoming a "de facto" standard in training infrastructures (t-Infrastructures) and it is adopted by many grid projects worldwide. In this contribution we will report on the latest status of GILDA services and on the training activities recently carried out in sub-Saharan Africa (Malawi and South Africa). Particular care will be devoted to show how GILDA can be "cloned" to satisfy both education and research demands of African Organisations. The opportunities to benefit from GILDA in the framework of the EPIKH project as well as the plans of the European Commission on grid training and education for the 2010-2011 calls of its 7th Framework Programme will be presented and discussed.
Parameterizing Grid-Averaged Longwave Fluxes for Inhomogeneous Marine Boundary Layer Clouds
NASA Technical Reports Server (NTRS)
Barker, Howard W.; Wielicki, Bruce A.
1997-01-01
This paper examines the relative impacts on grid-averaged longwave flux transmittance (emittance) for Marine Boundary Layer (MBL) cloud fields arising from horizontal variability of optical depth tau and cloud sides, First, using fields of Landsat-inferred tau and a Monte Carlo photon transport algorithm, it is demonstrated that mean all-sky transmittances for 3D variable MBL clouds can be computed accurately by the conventional method of linearly weighting clear and cloudy transmittances by their respective sky fractions. Then, the approximations of decoupling cloud and radiative properties and assuming independent columns are shown to be adequate for computation of mean flux transmittance. Since real clouds have nonzero geometric thicknesses, cloud fractions A'(sub c) presented to isotropic beams usually exceed the more familiar vertically projected cloud fractions A(sub c). It is shown, however, that when A(sub c)less than or equal to 0.9, biases for all-sky transmittance stemming from use of A(sub c) as opposed to A'(sub c) are roughly 2-5 times smaller than, and opposite in sign to, biases due to neglect of horizontal variability of tau. By neglecting variable tau, all-sky transmittances are underestimated often by more than 0.1 for A(sub c) near 0.75 and this translates into relative errors that can exceed 40% (corresponding errors for all-sky emittance are about 20% for most values of A(sub c). Thus, priority should be given to development of General Circulation Model (GCM) parameterizations that account for the effects of horizontal variations in unresolved tau, effects of cloud sides are of secondary importance. On this note, an efficient stochastic model for computing grid-averaged cloudy-sky flux transmittances is furnished that assumes that distributions of tau, for regions comparable in size to GCM grid cells, can be described adequately by gamma distribution functions. While the plane-parallel, homogeneous model underestimates cloud transmittance by about an order of magnitude when 3D variable cloud transmittances are less than or equal to 0.2 and by approx. 20% to 100% otherwise, the stochastic model reduces these biases often by more than 80%.
Physics-based distributed snow models in the operational arena: Current and future challenges
NASA Astrophysics Data System (ADS)
Winstral, A. H.; Jonas, T.; Schirmer, M.; Helbig, N.
2017-12-01
The demand for modeling tools robust to climate change and weather extremes along with coincident increases in computational capabilities have led to an increase in the use of physics-based snow models in operational applications. Current operational applications include the WSL-SLF's across Switzerland, ASO's in California, and USDA-ARS's in Idaho. While the physics-based approaches offer many advantages there remain limitations and modeling challenges. The most evident limitation remains computation times that often limit forecasters to a single, deterministic model run. Other limitations however remain less conspicuous amidst the assumptions that these models require little to no calibration based on their foundation on physical principles. Yet all energy balance snow models seemingly contain parameterizations or simplifications of processes where validation data are scarce or present understanding is limited. At the research-basin scale where many of these models were developed these modeling elements may prove adequate. However when applied over large areas, spatially invariable parameterizations of snow albedo, roughness lengths and atmospheric exchange coefficients - all vital to determining the snowcover energy balance - become problematic. Moreover as we apply models over larger grid cells, the representation of sub-grid variability such as the snow-covered fraction adds to the challenges. Here, we will demonstrate some of the major sensitivities of distributed energy balance snow models to particular model constructs, the need for advanced and spatially flexible methods and parameterizations, and prompt the community for open dialogue and future collaborations to further modeling capabilities.
Landlab: A numerical modeling framework for evolving Earth surfaces from mountains to the coast
NASA Astrophysics Data System (ADS)
Gasparini, N. M.; Adams, J. M.; Tucker, G. E.; Hobley, D. E. J.; Hutton, E.; Istanbulluoglu, E.; Nudurupati, S. S.
2016-02-01
Landlab is an open-source, user-friendly, component-based modeling framework for exploring the evolution of Earth's surface. Landlab itself is not a model. Instead, it is a computational framework that facilitates the development of numerical models of coupled earth surface processes. The Landlab Python library includes a gridding engine and process components, along with support functions for tasks such as reading in DEM data and input variables, setting boundary conditions, and plotting and outputting data. Each user of Landlab builds his or her own unique model. The first step in building a Landlab model is generally initializing a grid, either regular (raster) or irregular (e.g. delaunay or radial), and process components. This initialization process involves reading in relevant parameter values and data. The process components act on the grid to alter grid properties over time. For example, a component exists that can track the growth, death, and succession of vegetation over time. There are also several components that evolve surface elevation, through processes such as fluvial sediment transport and linear diffusion, among others. Users can also build their own process components, taking advantage of existing functions in Landlab such as those that identify grid connectivity and calculate gradients and flux divergence. The general nature of the framework makes it applicable to diverse environments - from bedrock rivers to a pile of sand - and processes acting over a range of spatial and temporal scales. In this poster we illustrate how a user builds a model using Landlab and propose a number of ways in which Landlab can be applied in coastal environments - from dune migration to channelization of barrier islands. We seek input from the coastal community as to how the process component library can be expanded to explore the diverse phenomena that act to shape coastal environments.
Numerical modeling and analysis of the effect of complex Greek topography on tornadogenesis
NASA Astrophysics Data System (ADS)
Matsangouras, I. T.; Pytharoulis, I.; Nastos, P. T.
2014-07-01
Tornadoes have been reported in Greece over recent decades in specific sub-geographical areas and have been associated with strong synoptic forcing. While it has been established that meteorological conditions over Greece are affected at various scales by the significant variability of topography, the Ionian Sea to the west and the Aegean Sea to the east, there is still uncertainty regarding topography's importance on tornadic generation and development. The aim of this study is to investigate the role of topography in significant tornadogenesis events that were triggered under strong synoptic scale forcing over Greece. Three tornado events that occurred over the last years in Thebes (Boeotia, 17 November 2007), Vrastema (Chalkidiki, 12 February 2010) and Vlychos (Lefkada, 20 September 2011) were selected for numerical experiments. These events were associated with synoptic scale forcing, while their intensities were T4-T5 (on the TORRO scale), causing significant damage. The simulations were performed using the non-hydrostatic weather research and forecasting model (WRF), initialized by European Centre for Medium-Range Weather Forecasts (ECMWF) gridded analyses, with telescoping nested grids that allow for the representation of atmospheric circulations ranging from the synoptic scale down to the mesoscale. In the experiments, the topography of the inner grid was modified by: (a) 0% (actual topography) and (b) -100% (without topography), making an effort to determine whether the occurrence of tornadoes - mainly identified by various severe weather instability indices - could be indicated by modifying topography. The principal instability variables employed consisted of the bulk Richardson number (BRN) shear, the energy helicity index (EHI), the storm-relative environmental helicity (SRH), and the maximum convective available potential energy (MCAPE, for parcels with maximum θe). Additionally, a model verification was conducted for every sensitivity experiment accompanied by analysis of the absolute vorticity budget. Numerical simulations revealed that the complex topography constituted an important factor during the 17 November 2007 and 12 February 2010 events, based on EHI, SRH, BRN, and MCAPE analyses. Conversely, topography around the 20 September 2011 event was characterized as the least significant factor based on EHI, SRH, BRN, and MCAPE analyses.
Spatial and Temporal Uncertainty of Crop Yield Aggregations
NASA Technical Reports Server (NTRS)
Porwollik, Vera; Mueller, Christoph; Elliott, Joshua; Chryssanthacopoulos, James; Iizumi, Toshichika; Ray, Deepak K.; Ruane, Alex C.; Arneth, Almut; Balkovic, Juraj; Ciais, Philippe;
2016-01-01
The aggregation of simulated gridded crop yields to national or regional scale requires information on temporal and spatial patterns of crop-specific harvested areas. This analysis estimates the uncertainty of simulated gridded yield time series related to the aggregation with four different harvested area data sets. We compare aggregated yield time series from the Global Gridded Crop Model Inter-comparison project for four crop types from 14 models at global, national, and regional scale to determine aggregation-driven differences in mean yields and temporal patterns as measures of uncertainty. The quantity and spatial patterns of harvested areas differ for individual crops among the four datasets applied for the aggregation. Also simulated spatial yield patterns differ among the 14 models. These differences in harvested areas and simulated yield patterns lead to differences in aggregated productivity estimates, both in mean yield and in the temporal dynamics. Among the four investigated crops, wheat yield (17% relative difference) is most affected by the uncertainty introduced by the aggregation at the global scale. The correlation of temporal patterns of global aggregated yield time series can be as low as for soybean (r = 0.28).For the majority of countries, mean relative differences of nationally aggregated yields account for10% or less. The spatial and temporal difference can be substantial higher for individual countries. Of the top-10 crop producers, aggregated national multi-annual mean relative difference of yields can be up to 67% (maize, South Africa), 43% (wheat, Pakistan), 51% (rice, Japan), and 427% (soybean, Bolivia).Correlations of differently aggregated yield time series can be as low as r = 0.56 (maize, India), r = 0.05*Corresponding (wheat, Russia), r = 0.13 (rice, Vietnam), and r = -0.01 (soybean, Uruguay). The aggregation to sub-national scale in comparison to country scale shows that spatial uncertainties can cancel out in countries with large harvested areas per crop type. We conclude that the aggregation uncertainty can be substantial for crop productivity and production estimations in the context of food security, impact assessment, and model evaluation exercises.
Energy Management and Optimization Methods for Grid Energy Storage Systems
Byrne, Raymond H.; Nguyen, Tu A.; Copp, David A.; ...
2017-08-24
Today, the stability of the electric power grid is maintained through real time balancing of generation and demand. Grid scale energy storage systems are increasingly being deployed to provide grid operators the flexibility needed to maintain this balance. Energy storage also imparts resiliency and robustness to the grid infrastructure. Over the last few years, there has been a significant increase in the deployment of large scale energy storage systems. This growth has been driven by improvements in the cost and performance of energy storage technologies and the need to accommodate distributed generation, as well as incentives and government mandates. Energymore » management systems (EMSs) and optimization methods are required to effectively and safely utilize energy storage as a flexible grid asset that can provide multiple grid services. The EMS needs to be able to accommodate a variety of use cases and regulatory environments. In this paper, we provide a brief history of grid-scale energy storage, an overview of EMS architectures, and a summary of the leading applications for storage. These serve as a foundation for a discussion of EMS optimization methods and design.« less
Energy Management and Optimization Methods for Grid Energy Storage Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Byrne, Raymond H.; Nguyen, Tu A.; Copp, David A.
Today, the stability of the electric power grid is maintained through real time balancing of generation and demand. Grid scale energy storage systems are increasingly being deployed to provide grid operators the flexibility needed to maintain this balance. Energy storage also imparts resiliency and robustness to the grid infrastructure. Over the last few years, there has been a significant increase in the deployment of large scale energy storage systems. This growth has been driven by improvements in the cost and performance of energy storage technologies and the need to accommodate distributed generation, as well as incentives and government mandates. Energymore » management systems (EMSs) and optimization methods are required to effectively and safely utilize energy storage as a flexible grid asset that can provide multiple grid services. The EMS needs to be able to accommodate a variety of use cases and regulatory environments. In this paper, we provide a brief history of grid-scale energy storage, an overview of EMS architectures, and a summary of the leading applications for storage. These serve as a foundation for a discussion of EMS optimization methods and design.« less
GLAD: a system for developing and deploying large-scale bioinformatics grid.
Teo, Yong-Meng; Wang, Xianbing; Ng, Yew-Kwong
2005-03-01
Grid computing is used to solve large-scale bioinformatics problems with gigabytes database by distributing the computation across multiple platforms. Until now in developing bioinformatics grid applications, it is extremely tedious to design and implement the component algorithms and parallelization techniques for different classes of problems, and to access remotely located sequence database files of varying formats across the grid. In this study, we propose a grid programming toolkit, GLAD (Grid Life sciences Applications Developer), which facilitates the development and deployment of bioinformatics applications on a grid. GLAD has been developed using ALiCE (Adaptive scaLable Internet-based Computing Engine), a Java-based grid middleware, which exploits the task-based parallelism. Two bioinformatics benchmark applications, such as distributed sequence comparison and distributed progressive multiple sequence alignment, have been developed using GLAD.
Family and peer support matter for precoital and coital behaviors among adolescents in Lima
Bayer, Angela M.; Cabrera, Lilia Z.; Gilman, Robert H.; Hindin, Michelle J.; Tsui, Amy O.
2015-01-01
We analyzed the association between sub-scales developed with adolescents and the outcomes of precoital behaviors and vaginal sex in Lima, Peru. Adolescent participants in key informant sessions operationalized concepts identified during qualitative concept mapping into several sub-scales. Face and content validity testing and pilot application with respondent debriefing were used to refine the sub-scales. Three hundred 15–17 year olds were surveyed about the sub-scales, socio-demographics and sexual behaviors. Exploratory factor analysis confirmed six sub-scales, self-image, goals and decision-making, family education, parental rules/control, school support and peer support, which we regressed on the outcomes. Twice as many males as females reported more than three precoital behaviors and vaginal sex. Higher peer support reduced the likelihood of vaginal sex and precoital behaviors and higher family education reduced precoital behaviors. Results affirm the importance of including adolescents in the entire research process and of sex education with family- and peer-based strategies. PMID:25305443
NASA Astrophysics Data System (ADS)
Grigoras, Costin; Carminati, Federico; Vladimirovna Datskova, Olga; Schreiner, Steffen; Lee, Sehoon; Zhu, Jianlin; Gheata, Mihaela; Gheata, Andrei; Saiz, Pablo; Betev, Latchezar; Furano, Fabrizio; Mendez Lorenzo, Patricia; Grigoras, Alina Gabriela; Bagnasco, Stefano; Peters, Andreas Joachim; Saiz Santos, Maria Dolores
2011-12-01
With the LHC and ALICE entering a full operation and production modes, the amount of Simulation and RAW data processing and end user analysis computational tasks are increasing. The efficient management of all these tasks, all of which have large differences in lifecycle, amounts of processed data and methods to analyze the end result, required the development and deployment of new tools in addition to the already existing Grid infrastructure. To facilitate the management of the large scale simulation and raw data reconstruction tasks, ALICE has developed a production framework called a Lightweight Production Manager (LPM). The LPM is automatically submitting jobs to the Grid based on triggers and conditions, for example after a physics run completion. It follows the evolution of the job and publishes the results on the web for worldwide access by the ALICE physicists. This framework is tightly integrated with the ALICE Grid framework AliEn. In addition to the publication of the job status, LPM is also allowing a fully authenticated interface to the AliEn Grid catalogue, to browse and download files, and in the near future will provide simple types of data analysis through ROOT plugins. The framework is also being extended to allow management of end user jobs.
Implement a Sub-grid Turbulent Orographic Form Drag in WRF and its application to Tibetan Plateau
NASA Astrophysics Data System (ADS)
Zhou, X.; Yang, K.; Wang, Y.; Huang, B.
2017-12-01
Sub-grid-scale orographic variation exerts turbulent form drag on atmospheric flows. The Weather Research and Forecasting model (WRF) includes a turbulent orographic form drag (TOFD) scheme that adds the stress to the surface layer. In this study, another TOFD scheme has been incorporated in WRF3.7, which exerts an exponentially decaying drag on each model layer. To investigate the effect of the new scheme, WRF with the old and new one was used to simulate the climate over the complex terrain of the Tibetan Plateau. The two schemes were evaluated in terms of the direct impact (on wind) and the indirect impact (on air temperature, surface pressure and precipitation). Both in winter and summer, the new TOFD scheme reduces the mean bias in the surface wind, and clearly reduces the root mean square error (RMSEs) in comparisons with the station measurements (Figure 1). Meanwhile, the 2-m air temperature and surface pressure is also improved (Figure 2) due to the more warm air northward transport across south boundary of TP in winter. The 2-m air temperature is hardly improved in summer but the precipitation improvement is more obvious, with reduced mean bias and RMSEs. This is due to the weakening of water vapor flux (at low-level flow with the new scheme) crossing the Himalayan Mountains from South Asia.
NASA Astrophysics Data System (ADS)
Gao, K.; Harris, L.; Chen, J. H.; Lin, S. J.
2017-12-01
Skillful subseasonal prediction of hurricane activity (from two weeks to less than a season) is important for early preparedness and reducing the hurricane damage in coastal regions. In this study, we will present evaluations of the performance of GFDL HiRAM (High-Resolution Atmospheric Model) for the simulation and prediction of the North Atlantic hurricane activity on the sub-seasonal time scale. A series of sub-seasonal (30-day duration) retrospective predictions were performed over the years 2000-2014 using two configurations of HiRAM: a) global uniform 25km-resolution grid and b) two-way nested grid with a 8km-resolution nest over North Atlantic. The analysis of hurricane structure from the two sets of simulations indicates the two-way-nesting method is an efficient way to improve the representation of hurricanes in global models: the two-way nested configuration produces realistic hurricane inner-core size and structure, which leads to improved lifetime maximum intensity distribution. Both configurations show very promising performance in the subseasonal hurricane genesis prediction, but the two-way nested configuration shows better performance in the prediction of major hurricane (Categories 3-5) activity because of the improved intensity simulation. We will also present the analysis of how the phase and magnitude of MJO, as well as the initial SST anomaly affect the model's prediction skill.
Global Electricity Trade Network: Structures and Implications
Ji, Ling; Jia, Xiaoping; Chiu, Anthony S. F.; Xu, Ming
2016-01-01
Nations increasingly trade electricity, and understanding the structure of the global power grid can help identify nations that are critical for its reliability. This study examines the global grid as a network with nations as nodes and international electricity trade as links. We analyze the structure of the global electricity trade network and find that the network consists of four sub-networks, and provide a detailed analysis of the largest network, Eurasia. Russia, China, Ukraine, and Azerbaijan have high betweenness measures in the Eurasian sub-network, indicating the degrees of centrality of the positions they hold. The analysis reveals that the Eurasian sub-network consists of seven communities based on the network structure. We find that the communities do not fully align with geographical proximity, and that the present international electricity trade in the Eurasian sub-network causes an approximately 11 million additional tons of CO2 emissions. PMID:27504825
Global Electricity Trade Network: Structures and Implications.
Ji, Ling; Jia, Xiaoping; Chiu, Anthony S F; Xu, Ming
2016-01-01
Nations increasingly trade electricity, and understanding the structure of the global power grid can help identify nations that are critical for its reliability. This study examines the global grid as a network with nations as nodes and international electricity trade as links. We analyze the structure of the global electricity trade network and find that the network consists of four sub-networks, and provide a detailed analysis of the largest network, Eurasia. Russia, China, Ukraine, and Azerbaijan have high betweenness measures in the Eurasian sub-network, indicating the degrees of centrality of the positions they hold. The analysis reveals that the Eurasian sub-network consists of seven communities based on the network structure. We find that the communities do not fully align with geographical proximity, and that the present international electricity trade in the Eurasian sub-network causes an approximately 11 million additional tons of CO2 emissions.
Landázuri, Andrea C.; Sáez, A. Eduardo; Anthony, T. Renée
2016-01-01
This work presents fluid flow and particle trajectory simulation studies to determine the aspiration efficiency of a horizontally oriented occupational air sampler using computational fluid dynamics (CFD). Grid adaption and manual scaling of the grids were applied to two sampler prototypes based on a 37-mm cassette. The standard k–ε model was used to simulate the turbulent air flow and a second order streamline-upwind discretization scheme was used to stabilize convective terms of the Navier–Stokes equations. Successively scaled grids for each configuration were created manually and by means of grid adaption using the velocity gradient in the main flow direction. Solutions were verified to assess iterative convergence, grid independence and monotonic convergence. Particle aspiration efficiencies determined for both prototype samplers were undistinguishable, indicating that the porous filter does not play a noticeable role in particle aspiration. Results conclude that grid adaption is a powerful tool that allows to refine specific regions that require lots of detail and therefore better resolve flow detail. It was verified that adaptive grids provided a higher number of locations with monotonic convergence than the manual grids and required the least computational effort. PMID:26949268
Muko, Soyoka; Shimatani, Ichiro K; Nozawa, Yoko
2014-07-01
Spatial distributions of individuals are conventionally analysed by representing objects as dimensionless points, in which spatial statistics are based on centre-to-centre distances. However, if organisms expand without overlapping and show size variations, such as is the case for encrusting corals, interobject spacing is crucial for spatial associations where interactions occur. We introduced new pairwise statistics using minimum distances between objects and demonstrated their utility when examining encrusting coral community data. We also calculated the conventional point process statistics and the grid-based statistics to clarify the advantages and limitations of each spatial statistical method. For simplicity, coral colonies were approximated by disks in these demonstrations. Focusing on short-distance effects, the use of minimum distances revealed that almost all coral genera were aggregated at a scale of 1-25 cm. However, when fragmented colonies (ramets) were treated as a genet, a genet-level analysis indicated weak or no aggregation, suggesting that most corals were randomly distributed and that fragmentation was the primary cause of colony aggregations. In contrast, point process statistics showed larger aggregation scales, presumably because centre-to-centre distances included both intercolony spacing and colony sizes (radius). The grid-based statistics were able to quantify the patch (aggregation) scale of colonies, but the scale was strongly affected by the colony size. Our approach quantitatively showed repulsive effects between an aggressive genus and a competitively weak genus, while the grid-based statistics (covariance function) also showed repulsion although the spatial scale indicated from the statistics was not directly interpretable in terms of ecological meaning. The use of minimum distances together with previously proposed spatial statistics helped us to extend our understanding of the spatial patterns of nonoverlapping objects that vary in size and the associated specific scales. © 2013 The Authors. Journal of Animal Ecology © 2013 British Ecological Society.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Michael J. Bockelie
2002-01-04
This DOE SBIR Phase II final report summarizes research that has been performed to develop a parallel adaptive tool for modeling steady, two phase turbulent reacting flow. The target applications for the new tool are full scale, fossil-fuel fired boilers and furnaces such as those used in the electric utility industry, chemical process industry and mineral/metal process industry. The type of analyses to be performed on these systems are engineering calculations to evaluate the impact on overall furnace performance due to operational, process or equipment changes. To develop a Computational Fluid Dynamics (CFD) model of an industrial scale furnace requiresmore » a carefully designed grid that will capture all of the large and small scale features of the flowfield. Industrial systems are quite large, usually measured in tens of feet, but contain numerous burners, air injection ports, flames and localized behavior with dimensions that are measured in inches or fractions of inches. To create an accurate computational model of such systems requires capturing length scales within the flow field that span several orders of magnitude. In addition, to create an industrially useful model, the grid can not contain too many grid points - the model must be able to execute on an inexpensive desktop PC in a matter of days. An adaptive mesh provides a convenient means to create a grid that can capture both fine flow field detail within a very large domain with a ''reasonable'' number of grid points. However, the use of an adaptive mesh requires the development of a new flow solver. To create the new simulation tool, we have combined existing reacting CFD modeling software with new software based on emerging block structured Adaptive Mesh Refinement (AMR) technologies developed at Lawrence Berkeley National Laboratory (LBNL). Specifically, we combined: -physical models, modeling expertise, and software from existing combustion simulation codes used by Reaction Engineering International; -mesh adaption, data management, and parallelization software and technology being developed by users of the BoxLib library at LBNL; and -solution methods for problems formulated on block structured grids that were being developed in collaboration with technical staff members at the University of Utah Center for High Performance Computing (CHPC) and at LBNL. The combustion modeling software used by Reaction Engineering International represents an investment of over fifty man-years of development, conducted over a period of twenty years. Thus, it was impractical to achieve our objective by starting from scratch. The research program resulted in an adaptive grid, reacting CFD flow solver that can be used only on limited problems. In current form the code is appropriate for use on academic problems with simplified geometries. The new solver is not sufficiently robust or sufficiently general to be used in a ''production mode'' for industrial applications. The principle difficulty lies with the multi-level solver technology. The use of multi-level solvers on adaptive grids with embedded boundaries is not yet a mature field and there are many issues that remain to be resolved. From the lessons learned in this SBIR program, we have started work on a new flow solver with an AMR capability. The new code is based on a conventional cell-by-cell mesh refinement strategy used in unstructured grid solvers that employ hexahedral cells. The new solver employs several of the concepts and solution strategies developed within this research program. The formulation of the composite grid problem for the new solver has been designed to avoid the embedded boundary complications encountered in this SBIR project. This follow-on effort will result in a reacting flow CFD solver with localized mesh capability that can be used to perform engineering calculations on industrial problems in a production mode.« less
NASA Astrophysics Data System (ADS)
Kossieris, Panagiotis; Makropoulos, Christos; Onof, Christian; Koutsoyiannis, Demetris
2018-01-01
Many hydrological applications, such as flood studies, require the use of long rainfall data at fine time scales varying from daily down to 1 min time step. However, in the real world there is limited availability of data at sub-hourly scales. To cope with this issue, stochastic disaggregation techniques are typically employed to produce possible, statistically consistent, rainfall events that aggregate up to the field data collected at coarser scales. A methodology for the stochastic disaggregation of rainfall at fine time scales was recently introduced, combining the Bartlett-Lewis process to generate rainfall events along with adjusting procedures to modify the lower-level variables (i.e., hourly) so as to be consistent with the higher-level one (i.e., daily). In the present paper, we extend the aforementioned scheme, initially designed and tested for the disaggregation of daily rainfall into hourly depths, for any sub-hourly time scale. In addition, we take advantage of the recent developments in Poisson-cluster processes incorporating in the methodology a Bartlett-Lewis model variant that introduces dependence between cell intensity and duration in order to capture the variability of rainfall at sub-hourly time scales. The disaggregation scheme is implemented in an R package, named HyetosMinute, to support disaggregation from daily down to 1-min time scale. The applicability of the methodology was assessed on a 5-min rainfall records collected in Bochum, Germany, comparing the performance of the above mentioned model variant against the original Bartlett-Lewis process (non-random with 5 parameters). The analysis shows that the disaggregation process reproduces adequately the most important statistical characteristics of rainfall at wide range of time scales, while the introduction of the model with dependent intensity-duration results in a better performance in terms of skewness, rainfall extremes and dry proportions.
Grid Research | Grid Modernization | NREL
Grid Research Grid Research NREL addresses the challenges of today's electric grid through high researcher in a lab Integrated Devices and Systems Developing and evaluating grid technologies and integrated Controls Developing methods for real-time operations and controls of power systems at any scale Photo of
Impacts of uncertainties in European gridded precipitation observations on regional climate analysis
Gobiet, Andreas
2016-01-01
ABSTRACT Gridded precipitation data sets are frequently used to evaluate climate models or to remove model output biases. Although precipitation data are error prone due to the high spatio‐temporal variability of precipitation and due to considerable measurement errors, relatively few attempts have been made to account for observational uncertainty in model evaluation or in bias correction studies. In this study, we compare three types of European daily data sets featuring two Pan‐European data sets and a set that combines eight very high‐resolution station‐based regional data sets. Furthermore, we investigate seven widely used, larger scale global data sets. Our results demonstrate that the differences between these data sets have the same magnitude as precipitation errors found in regional climate models. Therefore, including observational uncertainties is essential for climate studies, climate model evaluation, and statistical post‐processing. Following our results, we suggest the following guidelines for regional precipitation assessments. (1) Include multiple observational data sets from different sources (e.g. station, satellite, reanalysis based) to estimate observational uncertainties. (2) Use data sets with high station densities to minimize the effect of precipitation undersampling (may induce about 60% error in data sparse regions). The information content of a gridded data set is mainly related to its underlying station density and not to its grid spacing. (3) Consider undercatch errors of up to 80% in high latitudes and mountainous regions. (4) Analyses of small‐scale features and extremes are especially uncertain in gridded data sets. For higher confidence, use climate‐mean and larger scale statistics. In conclusion, neglecting observational uncertainties potentially misguides climate model development and can severely affect the results of climate change impact assessments. PMID:28111497
Prein, Andreas F; Gobiet, Andreas
2017-01-01
Gridded precipitation data sets are frequently used to evaluate climate models or to remove model output biases. Although precipitation data are error prone due to the high spatio-temporal variability of precipitation and due to considerable measurement errors, relatively few attempts have been made to account for observational uncertainty in model evaluation or in bias correction studies. In this study, we compare three types of European daily data sets featuring two Pan-European data sets and a set that combines eight very high-resolution station-based regional data sets. Furthermore, we investigate seven widely used, larger scale global data sets. Our results demonstrate that the differences between these data sets have the same magnitude as precipitation errors found in regional climate models. Therefore, including observational uncertainties is essential for climate studies, climate model evaluation, and statistical post-processing. Following our results, we suggest the following guidelines for regional precipitation assessments. (1) Include multiple observational data sets from different sources (e.g. station, satellite, reanalysis based) to estimate observational uncertainties. (2) Use data sets with high station densities to minimize the effect of precipitation undersampling (may induce about 60% error in data sparse regions). The information content of a gridded data set is mainly related to its underlying station density and not to its grid spacing. (3) Consider undercatch errors of up to 80% in high latitudes and mountainous regions. (4) Analyses of small-scale features and extremes are especially uncertain in gridded data sets. For higher confidence, use climate-mean and larger scale statistics. In conclusion, neglecting observational uncertainties potentially misguides climate model development and can severely affect the results of climate change impact assessments.
A Systematic Multi-Time Scale Solution for Regional Power Grid Operation
NASA Astrophysics Data System (ADS)
Zhu, W. J.; Liu, Z. G.; Cheng, T.; Hu, B. Q.; Liu, X. Z.; Zhou, Y. F.
2017-10-01
Many aspects need to be taken into consideration in a regional grid while making schedule plans. In this paper, a systematic multi-time scale solution for regional power grid operation considering large scale renewable energy integration and Ultra High Voltage (UHV) power transmission is proposed. In the time scale aspect, we discuss the problem from month, week, day-ahead, within-day to day-behind, and the system also contains multiple generator types including thermal units, hydro-plants, wind turbines and pumped storage stations. The 9 subsystems of the scheduling system are described, and their functions and relationships are elaborated. The proposed system has been constructed in a provincial power grid in Central China, and the operation results further verified the effectiveness of the system.
CO{sub 2} Reuse in Petrochemical Facilities
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jason Trembly; Brian Turk; Maruthi Pavani
2010-12-31
To address public concerns regarding the consequences of climate change from anthropogenic carbon dioxide (CO{sub 2}) emissions, the U.S. Department of Energy's National Energy Technology Laboratory (DOE/NETL) is actively funding a CO{sub 2} management program to develop technologies capable of mitigating CO{sub 2} emissions from power plant and industrial facilities. Over the past decade, this program has focused on reducing the costs of carbon capture and storage technologies. Recently, DOE/NETL launched an alternative CO{sub 2} mitigation program focused on beneficial CO{sub 2} reuse to support the development of technologies that mitigate emissions by converting CO{sub 2} into valuable chemicals andmore » fuels. RTI, with DOE/NETL support, has been developing an innovative beneficial CO{sub 2} reuse process for converting CO{sub 2} into substitute natural gas (SNG) by using by-product hydrogen (H{sub 2)-containing fuel gas from petrochemical facilities. This process leveraged commercial reactor technology currently used in fluid catalytic crackers in petroleum refining and a novel nickel (Ni)-based catalyst developed by RTI. The goal was to generate an SNG product that meets the pipeline specifications for natural gas, making the SNG product completely compatible with the existing natural gas infrastructure. RTI's technology development efforts focused on demonstrating the technical feasibility of this novel CO{sub 2} reuse process and obtaining the necessary engineering information to design a pilot demonstration unit for converting about 4 tons per day (tons/day) of CO{sub 2} into SNG at a suitable host site. This final report describes the results of the Phase I catalyst and process development efforts. The methanation activity of several commercial fixed-bed catalysts was evaluated under fluidized-bed conditions in a bench-scale reactor to identify catalyst performance targets. RTI developed two fluidizable Ni-based catalyst formulations (Cat-1 and Cat-3) that demonstrated equal or better performance than that of commercial methanation catalysts. The Cat-1 and Cat-3 formulations were successfully scaled up using commercial manufacturing equipment at the Sud-Chemie Inc. pilot-plant facility in Louisville, KY. Pilot transport reactor testing with RTI's Cat-1 formulation at Kellog Brown & Root's Technology Center demonstrated the ability of the process to achieve high single-pass CO{sub 2} conversion. Using information acquired from bench- and pilot-scale testing, a basic engineering design package was prepared for a 4-ton/day CO{sub 2} pilot demonstration unit, including process and instrumentation diagrams, equipment list, control philosophy, and preliminary cost estimate.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Meng, F.; Banks, J. W.; Henshaw, W. D.
We describe a new partitioned approach for solving conjugate heat transfer (CHT) problems where the governing temperature equations in different material domains are time-stepped in a implicit manner, but where the interface coupling is explicit. The new approach, called the CHAMP scheme (Conjugate Heat transfer Advanced Multi-domain Partitioned), is based on a discretization of the interface coupling conditions using a generalized Robin (mixed) condition. The weights in the Robin condition are determined from the optimization of a condition derived from a local stability analysis of the coupling scheme. The interface treatment combines ideas from optimized-Schwarz methods for domain-decomposition problems togethermore » with the interface jump conditions and additional compatibility jump conditions derived from the governing equations. For many problems (i.e. for a wide range of material properties, grid-spacings and time-steps) the CHAMP algorithm is stable and second-order accurate using no sub-time-step iterations (i.e. a single implicit solve of the temperature equation in each domain). In extreme cases (e.g. very fine grids with very large time-steps) it may be necessary to perform one or more sub-iterations. Each sub-iteration generally increases the range of stability substantially and thus one sub-iteration is likely sufficient for the vast majority of practical problems. The CHAMP algorithm is developed first for a model problem and analyzed using normal-mode the- ory. The theory provides a mechanism for choosing optimal parameters in the mixed interface condition. A comparison is made to the classical Dirichlet-Neumann (DN) method and, where applicable, to the optimized- Schwarz (OS) domain-decomposition method. For problems with different thermal conductivities and dif- fusivities, the CHAMP algorithm outperforms the DN scheme. For domain-decomposition problems with uniform conductivities and diffusivities, the CHAMP algorithm performs better than the typical OS scheme with one grid-cell overlap. Lastly, the CHAMP scheme is also developed for general curvilinear grids and CHT ex- amples are presented using composite overset grids that confirm the theory and demonstrate the effectiveness of the approach.« less
Optical Delineation of Benthic Habitat Using an Autonomous Underwater Vehicle
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moline, Mark A.; Woodruff, Dana L.; Evans, Nathan R.
To improve understanding and characterization of coastal regions, there has been an increasing emphasis on autonomous systems that can sample the ocean on relevant scales. Autonomous underwater vehicles (AUVs) with active propulsion are especially well suited for studies of the coastal ocean because they are able to provide systematic and near-synoptic spatial observations. With this capability, science users are beginning to integrate sensor suits for a broad range of specific and often novel applications. Here, the relatively mature Remote Environmental Monitoring Units (REMUS) AUV system is configured with multi-spectral radiometers to delineate benthic habitat in Sequim Bay, WA. The vehiclemore » was deployed in a grid pattern along 5 km of coastline in depths from 30 to less than 2 meters. Similar to satellite and/or aerial remote sensing, the bandwidth ratios from the downward looking radiance sensor and upward looking irradiance sensor were used to identify beds of eelgrass on sub-meter scales. Strong correlations were found between the optical reflectance signals and the geo-referenced in situ data collected with underwater video within the grid. Results demonstrate the ability of AUVs to map littoral habitats at high resolution and highlight the overall utility of the REMUS vehicle for nearshore oceanography.« less
NASA Astrophysics Data System (ADS)
Ullman, D. J.; Schmittner, A.; Danabasoglu, G.; Norton, N. J.; Müller, M.
2016-02-01
Oscillations in the moon's orbit around the earth modulate regional tidal dissipation with a periodicity of 18.6 years. In regions where the diurnal tidal constituents dominate diapycnal mixing, this Lunar Nodal Cycle (LNC) may be significant enough to influence ocean circulation, sea surface temperature, and climate variability. Such periodicity in the LNC as an external forcing may provide a mechanistic source for Pacific decadal variability (i.e. Pacific Decadal Oscillation, PDO) where diurnal tidal constituents are strong. We have introduced three enhancements to the latest version of the Community Earth System Model (CESM) to better simulate tidal-forced mixing. First, we have produced a sub-grid scale bathymetry scheme that better resolves the vertical distribution of the barotropic energy flux in regions where the native CESM grid does not resolve high spatial-scale bathymetric features. Second, we test a number of alternative barotropic tidal constituent energy flux fields that are derived from various satellite altimeter observations and tidal models. Third, we introduce modulations of the individual diurnal and semi-diurnal tidal constituents, ranging from monthly to decadal periods, as derived from the full lunisolar tidal potential. Using both ocean-only and fully-coupled configurations, we test the influence of these enhancements, particularly the LNC modulations, on ocean mixing and bidecadal climate variability in CESM.
EPIC-Simulated and MODIS-Derived Leaf Area Index (LAI) ...
Leaf Area Index (LAI) is an important parameter in assessing vegetation structure for characterizing forest canopies over large areas at broad spatial scales using satellite remote sensing data. However, satellite-derived LAI products can be limited by obstructed atmospheric conditions yielding sub-optimal values, or complete non-returns. The United States Environmental Protection Agency’s Exposure Methods and Measurements and Computational Exposure Divisions are investigating the viability of supplemental modelled LAI inputs into satellite-derived data streams to support various regional and local scale air quality models for retrospective and future climate assessments. In this present study, one-year (2002) of plot level stand characteristics at four study sites located in Virginia and North Carolina are used to calibrate species-specific plant parameters in a semi-empirical biogeochemical model. The Environmental Policy Integrated Climate (EPIC) model was designed primarily for managed agricultural field crop ecosystems, but also includes managed woody species that span both xeric and mesic sites (e.g., mesquite, pine, oak, etc.). LAI was simulated using EPIC at a 4 km2 and 12 km2 grid coincident with the regional Community Multiscale Air Quality Model (CMAQ) grid. LAI comparisons were made between model-simulated and MODIS-derived LAI. Field/satellite-upscaled LAI was also compared to the corresponding MODIS LAI value. Preliminary results show field/satel
NASA Astrophysics Data System (ADS)
Ault, T.; Schwartz, M. D.; Zurita-Milla, R.; Weltzin, J. F.; Betancourt, J. L.
2015-12-01
Climate change is expected to modify the timing of seasonal transitions this century, impacting wildlife migrations, ecosystem function, and agricultural activity. Tracking seasonal transitions in a consistent manner across space and through time requires indices that can be used for monitoring and managing biophysical and ecological systems during the coming decades. Here a new gridded dataset of spring indices is described and used to understand interannual, decadal, and secular trends across the coterminous US. This dataset is derived from daily interpolated meteorological data, and results are compared with historical station data to ensure the trends and variations are robust. Regional trends in the first leaf index range from -0.8 to -1.6 days per decade, while first bloom index trends are between -0.4 and -1.2 for most regions. However, these trends are modulated by interannual to multidecadal variations, which are substantial throughout the regions considered here. These findings emphasize the important role large-scale climate modes of variability play in modulating spring onset on interannual to multidecadal timescales. Finally, there is some potential for successful sub-seasonal forecasts of spring onset, as indices from most regions are significantly correlated with antecedent large-scale modes of variability.
NASA Astrophysics Data System (ADS)
Miguez-Macho, Gonzalo; Stenchikov, Georgiy L.; Robock, Alan
2004-07-01
It is well known that regional climate simulations are sensitive to the size and position of the domain chosen for calculations. Here we study the physical mechanisms of this sensitivity. We conducted simulations with the Regional Atmospheric Modeling System (RAMS) for June 2000 over North America at 50 km horizontal resolution using a 7500 km × 5400 km grid and NCEP/NCAR reanalysis as boundary conditions. The position of the domain was displaced in several directions, always maintaining the U.S. in the interior, out of the buffer zone along the lateral boundaries. Circulation biases developed a large scale structure, organized by the Rocky Mountains, resulting from a systematic shifting of the synoptic wave trains that crossed the domain. The distortion of the large-scale circulation was produced by interaction of the modeled flow with the lateral boundaries of the nested domain and varied when the position of the grid was altered. This changed the large-scale environment among the different simulations and translated into diverse conditions for the development of the mesoscale processes that produce most of precipitation for the Great Plains in the summer season. As a consequence, precipitation results varied, sometimes greatly, among the experiments with the different grid positions. To eliminate the dependence of results on the position of the domain, we used spectral nudging of waves longer than 2500 km above the boundary layer. Moisture was not nudged at any level. This constrained the synoptic scales to follow reanalysis while allowing the model to develop the small-scale dynamics responsible for the rainfall. Nudging of the large scales successfully eliminated the variation of precipitation results when the grid was moved. We suggest that this technique is necessary for all downscaling studies with regional models with domain sizes of a few thousand kilometers and larger embedded in global models.
Continuous data assimilation for downscaling large-footprint soil moisture retrievals
NASA Astrophysics Data System (ADS)
Altaf, Muhammad U.; Jana, Raghavendra B.; Hoteit, Ibrahim; McCabe, Matthew F.
2016-10-01
Soil moisture is a key component of the hydrologic cycle, influencing processes leading to runoff generation, infiltration and groundwater recharge, evaporation and transpiration. Generally, the measurement scale for soil moisture is found to be different from the modeling scales for these processes. Reducing this mismatch between observation and model scales in necessary for improved hydrological modeling. An innovative approach to downscaling coarse resolution soil moisture data by combining continuous data assimilation and physically based modeling is presented. In this approach, we exploit the features of Continuous Data Assimilation (CDA) which was initially designed for general dissipative dynamical systems and later tested numerically on the incompressible Navier-Stokes equation, and the Benard equation. A nudging term, estimated as the misfit between interpolants of the assimilated coarse grid measurements and the fine grid model solution, is added to the model equations to constrain the model's large scale variability by available measurements. Soil moisture fields generated at a fine resolution by a physically-based vadose zone model (HYDRUS) are subjected to data assimilation conditioned upon coarse resolution observations. This enables nudging of the model outputs towards values that honor the coarse resolution dynamics while still being generated at the fine scale. Results show that the approach is feasible to generate fine scale soil moisture fields across large extents, based on coarse scale observations. Application of this approach is likely in generating fine and intermediate resolution soil moisture fields conditioned on the radiometerbased, coarse resolution products from remote sensing satellites.
TopoSCALE v.1.0: downscaling gridded climate data in complex terrain
NASA Astrophysics Data System (ADS)
Fiddes, J.; Gruber, S.
2014-02-01
Simulation of land surface processes is problematic in heterogeneous terrain due to the the high resolution required of model grids to capture strong lateral variability caused by, for example, topography, and the lack of accurate meteorological forcing data at the site or scale it is required. Gridded data products produced by atmospheric models can fill this gap, however, often not at an appropriate spatial resolution to drive land-surface simulations. In this study we describe a method that uses the well-resolved description of the atmospheric column provided by climate models, together with high-resolution digital elevation models (DEMs), to downscale coarse-grid climate variables to a fine-scale subgrid. The main aim of this approach is to provide high-resolution driving data for a land-surface model (LSM). The method makes use of an interpolation of pressure-level data according to topographic height of the subgrid. An elevation and topography correction is used to downscale short-wave radiation. Long-wave radiation is downscaled by deriving a cloud-component of all-sky emissivity at grid level and using downscaled temperature and relative humidity fields to describe variability with elevation. Precipitation is downscaled with a simple non-linear lapse and optionally disaggregated using a climatology approach. We test the method in comparison with unscaled grid-level data and a set of reference methods, against a large evaluation dataset (up to 210 stations per variable) in the Swiss Alps. We demonstrate that the method can be used to derive meteorological inputs in complex terrain, with most significant improvements (with respect to reference methods) seen in variables derived from pressure levels: air temperature, relative humidity, wind speed and incoming long-wave radiation. This method may be of use in improving inputs to numerical simulations in heterogeneous and/or remote terrain, especially when statistical methods are not possible, due to lack of observations (i.e. remote areas or future periods).
Inter-model variability in hydrological extremes projections for Amazonian sub-basins
NASA Astrophysics Data System (ADS)
Andres Rodriguez, Daniel; Garofolo, Lucas; Lázaro de Siqueira Júnior, José; Samprogna Mohor, Guilherme; Tomasella, Javier
2014-05-01
Irreducible uncertainties due to knowledge's limitations, chaotic nature of climate system and human decision-making process drive uncertainties in Climate Change projections. Such uncertainties affect the impact studies, mainly when associated to extreme events, and difficult the decision-making process aimed at mitigation and adaptation. However, these uncertainties allow the possibility to develop exploratory analyses on system's vulnerability to different sceneries. The use of different climate model's projections allows to aboard uncertainties issues allowing the use of multiple runs to explore a wide range of potential impacts and its implications for potential vulnerabilities. Statistical approaches for analyses of extreme values are usually based on stationarity assumptions. However, nonstationarity is relevant at the time scales considered for extreme value analyses and could have great implications in dynamic complex systems, mainly under climate change transformations. Because this, it is required to consider the nonstationarity in the statistical distribution parameters. We carried out a study of the dispersion in hydrological extremes projections using climate change projections from several climate models to feed the Distributed Hydrological Model of the National Institute for Spatial Research, MHD-INPE, applied in Amazonian sub-basins. This model is a large-scale hydrological model that uses a TopModel approach to solve runoff generation processes at the grid-cell scale. MHD-INPE model was calibrated for 1970-1990 using observed meteorological data and comparing observed and simulated discharges by using several performance coeficients. Hydrological Model integrations were performed for present historical time (1970-1990) and for future period (2010-2100). Because climate models simulate the variability of the climate system in statistical terms rather than reproduce the historical behavior of climate variables, the performances of the model's runs during the historical period, when feed with climate model data, were tested using descriptors of the Flow Duration Curves. The analyses of projected extreme values were carried out considering the nonstationarity of the GEV distribution parameters and compared with extremes events in present time. Results show inter-model variability in a broad dispersion on projected extreme's values. Such dispersion implies different degrees of socio-economic impacts associated to extreme hydrological events. Despite the no existence of one optimum result, this variability allows the analyses of adaptation strategies and its potential vulnerabilities.
NASA Astrophysics Data System (ADS)
Arunachalam, S.; Baek, B. H.; Vennam, P. L.; Woody, M. C.; Omary, M.; Binkowski, F.; Fleming, G.
2012-12-01
Commercial aircraft emit substantial amounts of pollutants during their complete activity cycle that ranges from landing-and-takeoff (LTO) at airports to cruising in upper elevations of the atmosphere, and affect both air quality and climate. Since these emissions are not uniformly emitted over the earth, and have substantial temporal and spatial variability, it is vital to accurately evaluate and quantify the relative impacts of aviation emissions on ambient air quality. Regional-scale air quality modeling applications do not routinely include these aircraft emissions from all cycles. Federal Aviation Administration (FAA) has developed the Aviation Environmental Design Tool (AEDT), a software system that dynamically models aircraft performance in space and time to calculate fuel burn and emissions from gate-to-gate for all commercial aviation activity from all airports globally. To process in-flight aircraft emissions and to provide a realistic representation of these for treatment in grid-based air quality models, we have developed an interface processor called AEDTproc that accurately distributes full-flight chorded emissions in time and space to create gridded, hourly model-ready emissions input data. Unlike the traditional emissions modeling approach of treating aviation emissions as ground-level sources or processing emissions only from the LTO cycles in regional-scale air quality studies, AEDTproc distributes chorded inventories of aircraft emissions during LTO cycles and cruise activities into a time-variant 3-D gridded structure. We will present results of processed 2006 global emissions from AEDT over a continental U.S. modeling domain to support a national-scale air quality assessment of the incremental impacts of aircraft emissions on surface air quality. This includes about 13.6 million flights within the U.S. out of 31.2 million flights globally. We will focus on assessing spatio-temporal variability of these commercial aircraft emissions, and comparing upper tropospheric budgets of NOx from aircraft and lightning sources in the modeling domain.
Examples of data assimilation in mesoscale models
NASA Technical Reports Server (NTRS)
Carr, Fred; Zack, John; Schmidt, Jerry; Snook, John; Benjamin, Stan; Stauffer, David
1993-01-01
The keynote address was the problem of physical initialization of mesoscale models. The classic purpose of physical or diabatic initialization is to reduce or eliminate the spin-up error caused by the lack, at the initial time, of the fully developed vertical circulations required to support regions of large rainfall rates. However, even if a model has no spin-up problem, imposition of observed moisture and heating rate information during assimilation can improve quantitative precipitation forecasts, especially early in the forecast. The two key issues in physical initialization are the choice of assimilating technique and sources of hydrologic/hydrometeor data. Another example of data assimilation in mesoscale models was presented in a series of meso-beta scale model experiments with and 11 km version of the MASS model designed to investigate the sensitivity of convective initiation forced by thermally direct circulations resulting from differential surface heating to four dimensional assimilation of surface and radar data. The results of these simulations underscore the need to accurately initialize and simulate grid and sub-grid scale clouds in meso- beta scale models. The status of the application of the CSU-RAMS mesoscale model by the NOAA Forecast Systems Lab for producing real-time forecasts with 10-60 km mesh resolutions over (4000 km)(exp 2) domains for use by the aviation community was reported. Either MAPS or LAPS model data are used to initialize the RAMS model on a 12-h cycle. The use of MAPS (Mesoscale Analysis and Prediction System) model was discussed. Also discussed was the mesobeta-scale data assimilation using a triply-nested nonhydrostatic version of the MM5 model.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Khangaonkar, Tarang; Yang, Zhaoqing
2011-01-01
Estuarine and coastal hydrodynamic processes are sometimes neglected in the design and planning of nearshore restoration actions. Despite best intentions, efforts to restore nearshore habitats can result in poor outcomes if circulation and transport which also affect freshwater-saltwater interactions are not properly addressed. Limitations due to current land use can lead to selection of sub-optimal restoration alternatives that may result in undesirable consequences, such as flooding, deterioration of water quality, and erosion, requiring immediate remedies and costly repairs. Uncertainty with achieving restoration goals, such as recovery of tidal exchange, supply of sediment and nutrients, and establishment of fish migration pathways,more » may be minimized by using numerical models designed for application to the nearshore environment. A high resolution circulation and transport model of the Puget Sound, in the state of Washington, was developed to assist with nearshore habitat restoration design and analysis, and to answer the question “can we achieve beneficial restoration outcomes at small local scale, as well as at a large estuary-wide scale?” The Puget Sound model is based on an unstructured grid framework to define the complex Puget Sound shoreline using a finite volume coastal ocean model (FVCOM). The capability of the model for simulating the important nearshore processes, such as circulation in complex multiple tidal channels, wetting and drying of tide flats, and water quality and sediment transport as part of restoration feasibility, are illustrated through examples of restoration projects in Puget Sound.« less
NASA Astrophysics Data System (ADS)
Takemi, T.; Yasui, M.
2005-12-01
Recent studies on dust emission and transport have been concerning the small-scale atmospheric processes in order to incorporate them as a subgrid-scale effect in large-scale numerical prediction models. In the present study, we investigated the dynamical processes and mechanisms of dust emission, mixing, and transport induced by boundary-layer and cumulus convection under a fair-weather condition over a Chinese desert. We performed a set of sensitivity experiments as well as a control simulation in order to examine the effects of vertical wind shear, upper-level wind speed, and moist convection by using a simplified and idealized modeling framework. The results of the control experiment showed that surface dust emission was at first caused before the noon time by intense convective motion which not only developed in the boundary layer but also penetrated into the free troposphere. In the afternoon hours, boundary-layer dry convection actively mixed and transported dust within the boundary layer. Some of the convective cells penetrated above the boundary layer, which led to the generation of cumulus clouds and hence gradually increased the dust content in the free troposphere. Coupled effects of the dry and moist convection played an important role in inducing surface dust emission and transporting dust vertically. This was clearly demonstrated through the comparison of the results between the control and the sensitivity experiments. The results of the control simulation were compared with lidar measurements. The simulation well captured the observed diurnal features of the upward transport of dust. We also examined the dependence of the simulated results on grid resolution: the grid size was changed from 250 m up to 4 km. It was found that there was a significant difference between the 2-km and 4-km grids. If a cumulus parameterization was added to the 4-km grid run, the column content was comparable to the other cases. This result suggests that subgrid parameterizations are required if the grid size is larger than the order of 1 km in a fair-weather condition.
NASA Astrophysics Data System (ADS)
Bonacorsi, D.; Gutsche, O.
The Worldwide LHC Computing Grid (WLCG) project decided in March 2009 to perform scale tests of parts of its overall Grid infrastructure before the start of the LHC data taking. The "Scale Test for the Experiment Program" (STEP'09) was performed mainly in June 2009 -with more selected tests in September- October 2009 -and emphasized the simultaneous test of the computing systems of all 4 LHC experiments. CMS tested its Tier-0 tape writing and processing capabilities. The Tier-1 tape systems were stress tested using the complete range of Tier-1 work-flows: transfer from Tier-0 and custody of data on tape, processing and subsequent archival, redistribution of datasets amongst all Tier-1 sites as well as burst transfers of datasets to Tier-2 sites. The Tier-2 analysis capacity was tested using bulk analysis job submissions to backfill normal user activity. In this talk, we will report on the different performed tests and present their post-mortem analysis.
Grid-Enabled Quantitative Analysis of Breast Cancer
2010-10-01
large-scale, multi-modality computerized image analysis . The central hypothesis of this research is that large-scale image analysis for breast cancer...research, we designed a pilot study utilizing large scale parallel Grid computing harnessing nationwide infrastructure for medical image analysis . Also
An Integrated Hydrogen Production-CO2 Capture Process from Fossil Fuel
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhicheng Wang
The new technology concept integrates two significant complementary hydrogen production and CO{sub 2}-sequestration approaches that have been developed at Oak Ridge National Laboratory (ORNL) and Clark Atlanta University. The process can convert biomass into hydrogen and char. Hydrogen can be efficiently used for stationary power and mobile applications, or it can be synthesized into Ammonia which can be used for CO{sub 2}-sequestration, while char can be used for making time-release fertilizers (NH{sub 4}HCO{sub 3}) by absorption of CO{sub 2} and other acid gases from exhaust flows. Fertilizers are then used for the growth of biomass back to fields. This projectmore » includes bench scale experiments and pilot scale tests. The Combustion and Emission Lab at Clark Atlanta University has conducted the bench scale experiments. The facility used for pilot scale tests was built in Athens, GA. The overall yield from this process is 7 wt% hydrogen and 32 wt% charcoal/activated carbon of feedstock (peanut shell). The value of co-product activated carbon is about $1.1/GJ and this coproduct reduced the selling price of hydrogen. And the selling price of hydrogen is estimated to be $6.95/GJ. The green house experimental results show that the samples added carbon-fertilizers have effectively growth increase of three different types of plants and improvement ability of keeping fertilizer in soil to avoid the fertilizer leaching with water.« less
Subgrid-scale parameterization and low-frequency variability: a response theory approach
NASA Astrophysics Data System (ADS)
Demaeyer, Jonathan; Vannitsem, Stéphane
2016-04-01
Weather and climate models are limited in the possible range of resolved spatial and temporal scales. However, due to the huge space- and time-scale ranges involved in the Earth System dynamics, the effects of many sub-grid processes should be parameterized. These parameterizations have an impact on the forecasts or projections. It could also affect the low-frequency variability present in the system (such as the one associated to ENSO or NAO). An important question is therefore to know what is the impact of stochastic parameterizations on the Low-Frequency Variability generated by the system and its model representation. In this context, we consider a stochastic subgrid-scale parameterization based on the Ruelle's response theory and proposed in Wouters and Lucarini (2012). We test this approach in the context of a low-order coupled ocean-atmosphere model, detailed in Vannitsem et al. (2015), for which a part of the atmospheric modes is considered as unresolved. A natural separation of the phase-space into a slow invariant set and its fast complement allows for an analytical derivation of the different terms involved in the parameterization, namely the average, the fluctuation and the long memory terms. Its application to the low-order system reveals that a considerable correction of the low-frequency variability along the invariant subset can be obtained. This new approach of scale separation opens new avenues of subgrid-scale parameterizations in multiscale systems used for climate forecasts. References: Vannitsem S, Demaeyer J, De Cruz L, Ghil M. 2015. Low-frequency variability and heat transport in a low-order nonlinear coupled ocean-atmosphere model. Physica D: Nonlinear Phenomena 309: 71-85. Wouters J, Lucarini V. 2012. Disentangling multi-level systems: averaging, correlations and memory. Journal of Statistical Mechanics: Theory and Experiment 2012(03): P03 003.
NASA Astrophysics Data System (ADS)
Kim, Jongho; Ivanov, Valeriy Y.; Katopodes, Nikolaos D.
2013-09-01
A novel two-dimensional, physically based model of soil erosion and sediment transport coupled to models of hydrological and overland flow processes has been developed. The Hairsine-Rose formulation of erosion and deposition processes is used to account for size-selective sediment transport and differentiate bed material into original and deposited soil layers. The formulation is integrated within the framework of the hydrologic and hydrodynamic model tRIBS-OFM, Triangulated irregular network-based, Real-time Integrated Basin Simulator-Overland Flow Model. The integrated model explicitly couples the hydrodynamic formulation with the advection-dominated transport equations for sediment of multiple particle sizes. To solve the system of equations including both the Saint-Venant and the Hairsine-Rose equations, the finite volume method is employed based on Roe's approximate Riemann solver on an unstructured grid. The formulation yields space-time dynamics of flow, erosion, and sediment transport at fine scale. The integrated model has been successfully verified with analytical solutions and empirical data for two benchmark cases. Sensitivity tests to grid resolution and the number of used particle sizes have been carried out. The model has been validated at the catchment scale for the Lucky Hills watershed located in southeastern Arizona, USA, using 10 events for which catchment-scale streamflow and sediment yield data were available. Since the model is based on physical laws and explicitly uses multiple types of watershed information, satisfactory results were obtained. The spatial output has been analyzed and the driving role of topography in erosion processes has been discussed. It is expected that the integrated formulation of the model has the promise to reduce uncertainties associated with typical parameterizations of flow and erosion processes. A potential for more credible modeling of earth-surface processes is thus anticipated.
LARGE SCALE METHOD FOR THE PRODUCTION AND PURIFICATION OF CURIUM
Higgins, G.H.; Crane, W.W.T.
1959-05-19
A large-scale process for production and purification of Cm/sup 242/ is described. Aluminum slugs containing Am are irradiated and declad in a NaOH-- NaHO/sub 3/ solution at 85 to 100 deg C. The resulting slurry filtered and washed with NaOH, NH/sub 4/OH, and H/sub 2/O. Recovery of Cm from filtrate and washings is effected by an Fe(OH)/sub 3/ precipitation. The precipitates are then combined and dissolved ln HCl and refractory oxides centrifuged out. These oxides are then fused with Na/sub 2/CO/sub 3/ and dissolved in HCl. The solution is evaporated and LiCl solution added. The Cm, rare earths, and anionic impurities are adsorbed on a strong-base anfon exchange resin. Impurities are eluted with LiCl--HCl solution, rare earths and Cm are eluted by HCl. Other ion exchange steps further purify the Cm. The Cm is then precipitated as fluoride and used in this form or further purified and processed. (T.R.H.)
Low Quality Natural Gas Sulfur Removal and Recovery CNG Claus Sulfur Recovery Process
DOE Office of Scientific and Technical Information (OSTI.GOV)
Klint, V.W.; Dale, P.R.; Stephenson, C.
1997-10-01
Increased use of natural gas (methane) in the domestic energy market will force the development of large non-producing gas reserves now considered to be low quality. Large reserves of low quality natural gas (LQNG) contaminated with hydrogen sulfide (H{sub 2}S), carbon dioxide (CO{sub 2}) and nitrogen (N) are available but not suitable for treatment using current conventional gas treating methods due to economic and environmental constraints. A group of three technologies have been integrated to allow for processing of these LQNG reserves; the Controlled Freeze Zone (CFZ) process for hydrocarbon / acid gas separation; the Triple Point Crystallizer (TPC) processmore » for H{sub 2}S / C0{sub 2} separation and the CNG Claus process for recovery of elemental sulfur from H{sub 2}S. The combined CFZ/TPC/CNG Claus group of processes is one program aimed at developing an alternative gas treating technology which is both economically and environmentally suitable for developing these low quality natural gas reserves. The CFZ/TPC/CNG Claus process is capable of treating low quality natural gas containing >10% C0{sub 2} and measurable levels of H{sub 2}S and N{sub 2} to pipeline specifications. The integrated CFZ / CNG Claus Process or the stand-alone CNG Claus Process has a number of attractive features for treating LQNG. The processes are capable of treating raw gas with a variety of trace contaminant components. The processes can also accommodate large changes in raw gas composition and flow rates. The combined processes are capable of achieving virtually undetectable levels of H{sub 2}S and significantly less than 2% CO in the product methane. The separation processes operate at pressure and deliver a high pressure (ca. 100 psia) acid gas (H{sub 2}S) stream for processing in the CNG Claus unit. This allows for substantial reductions in plant vessel size as compared to conventional Claus / Tail gas treating technologies. A close integration of the components of the CNG Claus process also allow for use of the methane/H{sub 2}S separation unit as a Claus tail gas treating unit by recycling the CNG Claus tail gas stream. This allows for virtually 100 percent sulfur recovery efficiency (virtually zero SO{sub 2} emissions) by recycling the sulfur laden tail gas to extinction. The use of the tail gas recycle scheme also deemphasizes the conventional requirement in Claus units to have high unit conversion efficiency and thereby make the operation much less affected by process upsets and feed gas composition changes. The development of these technologies has been ongoing for many years and both the CFZ and the TPC processes have been demonstrated at large pilot plant scales. On the other hand, prior to this project, the CNG Claus process had not been proven at any scale. Therefore, the primary objective of this portion of the program was to design, build and operate a pilot scale CNG Claus unit and demonstrate the required fundamental reaction chemistry and also demonstrate the viability of a reasonably sized working unit.« less
NASA Astrophysics Data System (ADS)
Klingbeil, Knut; Lemarié, Florian; Debreu, Laurent; Burchard, Hans
2018-05-01
The state of the art of the numerics of hydrostatic structured-grid coastal ocean models is reviewed here. First, some fundamental differences in the hydrodynamics of the coastal ocean, such as the large surface elevation variation compared to the mean water depth, are contrasted against large scale ocean dynamics. Then the hydrodynamic equations as they are used in coastal ocean models as well as in large scale ocean models are presented, including parameterisations for turbulent transports. As steps towards discretisation, coordinate transformations and spatial discretisations based on a finite-volume approach are discussed with focus on the specific requirements for coastal ocean models. As in large scale ocean models, splitting of internal and external modes is essential also for coastal ocean models, but specific care is needed when drying & flooding of intertidal flats is included. As one obvious characteristic of coastal ocean models, open boundaries occur and need to be treated in a way that correct model forcing from outside is transmitted to the model domain without reflecting waves from the inside. Here, also new developments in two-way nesting are presented. Single processes such as internal inertia-gravity waves, advection and turbulence closure models are discussed with focus on the coastal scales. Some overview on existing hydrostatic structured-grid coastal ocean models is given, including their extensions towards non-hydrostatic models. Finally, an outlook on future perspectives is made.
Matthew P. Peters; Louis R. Iverson; Anantha M. Prasad; Steve N. Matthews
2013-01-01
Fine-scale soil (SSURGO) data were processed at the county level for 37 states within the eastern United States, initially for use as predictor variables in a species distribution model called DISTRIB II. Values from county polygon files converted into a continuous 30-m raster grid were aggregated to 4-km cells and integrated with other environmental and site condition...
On the use of Schwarz-Christoffel conformal mappings to the grid generation for global ocean models
NASA Astrophysics Data System (ADS)
Xu, S.; Wang, B.; Liu, J.
2015-10-01
In this article we propose two grid generation methods for global ocean general circulation models. Contrary to conventional dipolar or tripolar grids, the proposed methods are based on Schwarz-Christoffel conformal mappings that map areas with user-prescribed, irregular boundaries to those with regular boundaries (i.e., disks, slits, etc.). The first method aims at improving existing dipolar grids. Compared with existing grids, the sample grid achieves a better trade-off between the enlargement of the latitudinal-longitudinal portion and the overall smooth grid cell size transition. The second method addresses more modern and advanced grid design requirements arising from high-resolution and multi-scale ocean modeling. The generated grids could potentially achieve the alignment of grid lines to the large-scale coastlines, enhanced spatial resolution in coastal regions, and easier computational load balance. Since the grids are orthogonal curvilinear, they can be easily utilized by the majority of ocean general circulation models that are based on finite difference and require grid orthogonality. The proposed grid generation algorithms can also be applied to the grid generation for regional ocean modeling where complex land-sea distribution is present.
From GCM grid cell to agricultural plot: scale issues affecting modelling of climate impact
Baron, Christian; Sultan, Benjamin; Balme, Maud; Sarr, Benoit; Traore, Seydou; Lebel, Thierry; Janicot, Serge; Dingkuhn, Michael
2005-01-01
General circulation models (GCM) are increasingly capable of making relevant predictions of seasonal and long-term climate variability, thus improving prospects of predicting impact on crop yields. This is particularly important for semi-arid West Africa where climate variability and drought threaten food security. Translating GCM outputs into attainable crop yields is difficult because GCM grid boxes are of larger scale than the processes governing yield, involving partitioning of rain among runoff, evaporation, transpiration, drainage and storage at plot scale. This study analyses the bias introduced to crop simulation when climatic data is aggregated spatially or in time, resulting in loss of relevant variation. A detailed case study was conducted using historical weather data for Senegal, applied to the crop model SARRA-H (version for millet). The study was then extended to a 10°N–17° N climatic gradient and a 31 year climate sequence to evaluate yield sensitivity to the variability of solar radiation and rainfall. Finally, a down-scaling model called LGO (Lebel–Guillot–Onibon), generating local rain patterns from grid cell means, was used to restore the variability lost by aggregation. Results indicate that forcing the crop model with spatially aggregated rainfall causes yield overestimations of 10–50% in dry latitudes, but nearly none in humid zones, due to a biased fraction of rainfall available for crop transpiration. Aggregation of solar radiation data caused significant bias in wetter zones where radiation was limiting yield. Where climatic gradients are steep, these two situations can occur within the same GCM grid cell. Disaggregation of grid cell means into a pattern of virtual synoptic stations having high-resolution rainfall distribution removed much of the bias caused by aggregation and gave realistic simulations of yield. It is concluded that coupling of GCM outputs with plot level crop models can cause large systematic errors due to scale incompatibility. These errors can be avoided by transforming GCM outputs, especially rainfall, to simulate the variability found at plot level. PMID:16433096
The International Symposium on Grids and Clouds
NASA Astrophysics Data System (ADS)
The International Symposium on Grids and Clouds (ISGC) 2012 will be held at Academia Sinica in Taipei from 26 February to 2 March 2012, with co-located events and workshops. The conference is hosted by the Academia Sinica Grid Computing Centre (ASGC). 2012 is the decennium anniversary of the ISGC which over the last decade has tracked the convergence, collaboration and innovation of individual researchers across the Asia Pacific region to a coherent community. With the continuous support and dedication from the delegates, ISGC has provided the primary international distributed computing platform where distinguished researchers and collaboration partners from around the world share their knowledge and experiences. The last decade has seen the wide-scale emergence of e-Infrastructure as a critical asset for the modern e-Scientist. The emergence of large-scale research infrastructures and instruments that has produced a torrent of electronic data is forcing a generational change in the scientific process and the mechanisms used to analyse the resulting data deluge. No longer can the processing of these vast amounts of data and production of relevant scientific results be undertaken by a single scientist. Virtual Research Communities that span organisations around the world, through an integrated digital infrastructure that connects the trust and administrative domains of multiple resource providers, have become critical in supporting these analyses. Topics covered in ISGC 2012 include: High Energy Physics, Biomedicine & Life Sciences, Earth Science, Environmental Changes and Natural Disaster Mitigation, Humanities & Social Sciences, Operations & Management, Middleware & Interoperability, Security and Networking, Infrastructure Clouds & Virtualisation, Business Models & Sustainability, Data Management, Distributed Volunteer & Desktop Grid Computing, High Throughput Computing, and High Performance, Manycore & GPU Computing.