Hybrid Surface Mesh Adaptation for Climate Modeling
Ahmed Khamayseh; Valmor de Almeida; Glen Hansen
2008-10-01
Solution-driven mesh adaptation is becoming quite popular for spatial error control in the numerical simulation of complex computational physics applications, such as climate modeling. Typically, spatial adaptation is achieved by element subdivision (h adaptation) with a primary goal of resolving the local length scales of interest. A second, less-popular method of spatial adaptivity is called “mesh motion” (r adaptation); the smooth repositioning of mesh node points aimed at resizing existing elements to capture the local length scales. This paper proposes an adaptation method based on a combination of both element subdivision and node point repositioning (rh adaptation). By combining these two methods using the notion of a mobility function, the proposed approach seeks to increase the flexibility and extensibility of mesh motion algorithms while providing a somewhat smoother transition between refined regions than is produced by element subdivision alone. Further, in an attempt to support the requirements of a very general class of climate simulation applications, the proposed method is designed to accommodate unstructured, polygonal mesh topologies in addition to the most popular mesh types.
Adjoint Methods for Guiding Adaptive Mesh Refinement in Tsunami Modeling
NASA Astrophysics Data System (ADS)
Davis, B. N.; LeVeque, R. J.
2016-12-01
One difficulty in developing numerical methods for tsunami modeling is the fact that solutions contain time-varying regions where much higher resolution is required than elsewhere in the domain, particularly when tracking a tsunami propagating across the ocean. The open source GeoClaw software deals with this issue by using block-structured adaptive mesh refinement to selectively refine around propagating waves. For problems where only a target area of the total solution is of interest (e.g., one coastal community), a method that allows identifying and refining the grid only in regions that influence this target area would significantly reduce the computational cost of finding a solution. In this work, we show that solving the time-dependent adjoint equation and using a suitable inner product with the forward solution allows more precise refinement of the relevant waves. We present the adjoint methodology first in one space dimension for illustration and in a broad context since it could also be used in other adaptive software, and potentially for other tsunami applications beyond adaptive refinement. We then show how this adjoint method has been integrated into the adaptive mesh refinement strategy of the open source GeoClaw software and present tsunami modeling results showing that the accuracy of the solution is maintained and the computational time required is significantly reduced through the integration of the adjoint method into adaptive mesh refinement.
Model-Based Nonrigid Motion Analysis Using Natural Feature Adaptive Mesh
Zhang, Y.; Goldgof, D.B.; Sarkar, S.; Tsap, L.V.
2000-04-25
The success of nonrigid motion analysis using physical finite element model is dependent on the mesh that characterizes the object's geometric structure. We suggest a deformable mesh adapted to the natural features of images. The adaptive mesh requires much fewer number of nodes than the fixed mesh which was used in our previous work. We demonstrate the higher efficiency of the adaptive mesh in the context of estimating burn scar elasticity relative to normal skin elasticity using the observed 2D image sequence. Our results show that the scar assessment method based on the physical model using natural feature adaptive mesh can be applied to images which do not have artificial markers.
Thermal-chemical Mantle Convection Models With Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Leng, W.; Zhong, S.
2008-12-01
In numerical modeling of mantle convection, resolution is often crucial for resolving small-scale features. New techniques, adaptive mesh refinement (AMR), allow local mesh refinement wherever high resolution is needed, while leaving other regions with relatively low resolution. Both computational efficiency for large- scale simulation and accuracy for small-scale features can thus be achieved with AMR. Based on the octree data structure [Tu et al. 2005], we implement the AMR techniques into the 2-D mantle convection models. For pure thermal convection models, benchmark tests show that our code can achieve high accuracy with relatively small number of elements both for isoviscous cases (i.e. 7492 AMR elements v.s. 65536 uniform elements) and for temperature-dependent viscosity cases (i.e. 14620 AMR elements v.s. 65536 uniform elements). We further implement tracer-method into the models for simulating thermal-chemical convection. By appropriately adding and removing tracers according to the refinement of the meshes, our code successfully reproduces the benchmark results in van Keken et al. [1997] with much fewer elements and tracers compared with uniform-mesh models (i.e. 7552 AMR elements v.s. 16384 uniform elements, and ~83000 tracers v.s. ~410000 tracers). The boundaries of the chemical piles in our AMR code can be easily refined to the scales of a few kilometers for the Earth's mantle and the tracers are concentrated near the chemical boundaries to precisely trace the evolvement of the boundaries. It is thus very suitable for our AMR code to study the thermal-chemical convection problems which need high resolution to resolve the evolvement of chemical boundaries, such as the entrainment problems [Sleep, 1988].
Nonhydrostatic adaptive mesh dynamics for multiscale climate models (Invited)
NASA Astrophysics Data System (ADS)
Collins, W.; Johansen, H.; McCorquodale, P.; Colella, P.; Ullrich, P. A.
2013-12-01
Many of the atmospheric phenomena with the greatest potential impact in future warmer climates are inherently multiscale. Such meteorological systems include hurricanes and tropical cyclones, atmospheric rivers, and other types of hydrometeorological extremes. These phenomena are challenging to simulate in conventional climate models due to the relatively coarse uniform model resolutions relative to the native nonhydrostatic scales of the phenomonological dynamics. To enable studies of these systems with sufficient local resolution for the multiscale dynamics yet with sufficient speed for climate-change studies, we have adapted existing adaptive mesh dynamics for the DOE-NSF Community Atmosphere Model (CAM). In this talk, we present an adaptive, conservative finite volume approach for moist non-hydrostatic atmospheric dynamics. The approach is based on the compressible Euler equations on 3D thin spherical shells, where the radial direction is treated implicitly (using a fourth-order Runga-Kutta IMEX scheme) to eliminate time step constraints from vertical acoustic waves. Refinement is performed only in the horizontal directions. The spatial discretization is the equiangular cubed-sphere mapping, with a fourth-order accurate discretization to compute flux averages on faces. By using both space-and time-adaptive mesh refinement, the solver allocates computational effort only where greater accuracy is needed. The resulting method is demonstrated to be fourth-order accurate for model problems, and robust at solution discontinuities and stable for large aspect ratios. We present comparisons using a simplified physics package for dycore comparisons of moist physics. Hadley cell lifting an advected tracer into upper atmosphere, with horizontal adaptivity
NASA Technical Reports Server (NTRS)
Stapleton, Scott; Gries, Thomas; Waas, Anthony M.; Pineda, Evan J.
2014-01-01
Enhanced finite elements are elements with an embedded analytical solution that can capture detailed local fields, enabling more efficient, mesh independent finite element analysis. The shape functions are determined based on the analytical model rather than prescribed. This method was applied to adhesively bonded joints to model joint behavior with one element through the thickness. This study demonstrates two methods of maintaining the fidelity of such elements during adhesive non-linearity and cracking without increasing the mesh needed for an accurate solution. The first method uses adaptive shape functions, where the shape functions are recalculated at each load step based on the softening of the adhesive. The second method is internal mesh adaption, where cracking of the adhesive within an element is captured by further discretizing the element internally to represent the partially cracked geometry. By keeping mesh adaptations within an element, a finer mesh can be used during the analysis without affecting the global finite element model mesh. Examples are shown which highlight when each method is most effective in reducing the number of elements needed to capture adhesive nonlinearity and cracking. These methods are validated against analogous finite element models utilizing cohesive zone elements.
Efficient Plasma Ion Source Modeling With Adaptive Mesh Refinement (Abstract)
Kim, J.S.; Vay, J.L.; Friedman, A.; Grote, D.P.
2005-03-15
Ion beam drivers for high energy density physics and inertial fusion energy research require high brightness beams, so there is little margin of error allowed for aberration at the emitter. Thus, accurate plasma ion source computer modeling is required to model the plasma sheath region and time-dependent effects correctly.A computer plasma source simulation module that can be used with a powerful heavy ion fusion code, WARP, or as a standalone code, is being developed. In order to treat the plasma sheath region accurately and efficiently, the module will have the capability of handling multiple spatial scale problems by using Adaptive Mesh Refinement (AMR). We will report on our progress on the project.
Adaptive Mesh Refinement in Reactive Transport Modeling of Subsurface Environments
NASA Astrophysics Data System (ADS)
Molins, S.; Day, M.; Trebotich, D.; Graves, D. T.
2015-12-01
Adaptive mesh refinement (AMR) is a numerical technique for locally adjusting the resolution of computational grids. AMR makes it possible to superimpose levels of finer grids on the global computational grid in an adaptive manner allowing for more accurate calculations locally. AMR codes rely on the fundamental concept that the solution can be computed in different regions of the domain with different spatial resolutions. AMR codes have been applied to a wide range of problem including (but not limited to): fully compressible hydrodynamics, astrophysical flows, cosmological applications, combustion, blood flow, heat transfer in nuclear reactors, and land ice and atmospheric models for climate. In subsurface applications, in particular, reactive transport modeling, AMR may be particularly useful in accurately capturing concentration gradients (hence, reaction rates) that develop in localized areas of the simulation domain. Accurate evaluation of reaction rates is critical in many subsurface applications. In this contribution, we will discuss recent applications that bring to bear AMR capabilities on reactive transport problems from the pore scale to the flood plain scale.
Parallel Adaptive Mesh Refinement
Diachin, L; Hornung, R; Plassmann, P; WIssink, A
2005-03-04
As large-scale, parallel computers have become more widely available and numerical models and algorithms have advanced, the range of physical phenomena that can be simulated has expanded dramatically. Many important science and engineering problems exhibit solutions with localized behavior where highly-detailed salient features or large gradients appear in certain regions which are separated by much larger regions where the solution is smooth. Examples include chemically-reacting flows with radiative heat transfer, high Reynolds number flows interacting with solid objects, and combustion problems where the flame front is essentially a two-dimensional sheet occupying a small part of a three-dimensional domain. Modeling such problems numerically requires approximating the governing partial differential equations on a discrete domain, or grid. Grid spacing is an important factor in determining the accuracy and cost of a computation. A fine grid may be needed to resolve key local features while a much coarser grid may suffice elsewhere. Employing a fine grid everywhere may be inefficient at best and, at worst, may make an adequately resolved simulation impractical. Moreover, the location and resolution of fine grid required for an accurate solution is a dynamic property of a problem's transient features and may not be known a priori. Adaptive mesh refinement (AMR) is a technique that can be used with both structured and unstructured meshes to adjust local grid spacing dynamically to capture solution features with an appropriate degree of resolution. Thus, computational resources can be focused where and when they are needed most to efficiently achieve an accurate solution without incurring the cost of a globally-fine grid. Figure 1.1 shows two example computations using AMR; on the left is a structured mesh calculation of a impulsively-sheared contact surface and on the right is the fuselage and volume discretization of an RAH-66 Comanche helicopter [35]. Note the
Parallel Adaptive Mesh Refinement Library
NASA Technical Reports Server (NTRS)
Mac-Neice, Peter; Olson, Kevin
2005-01-01
Parallel Adaptive Mesh Refinement Library (PARAMESH) is a package of Fortran 90 subroutines designed to provide a computer programmer with an easy route to extension of (1) a previously written serial code that uses a logically Cartesian structured mesh into (2) a parallel code with adaptive mesh refinement (AMR). Alternatively, in its simplest use, and with minimal effort, PARAMESH can operate as a domain-decomposition tool for users who want to parallelize their serial codes but who do not wish to utilize adaptivity. The package builds a hierarchy of sub-grids to cover the computational domain of a given application program, with spatial resolution varying to satisfy the demands of the application. The sub-grid blocks form the nodes of a tree data structure (a quad-tree in two or an oct-tree in three dimensions). Each grid block has a logically Cartesian mesh. The package supports one-, two- and three-dimensional models.
NASA Astrophysics Data System (ADS)
Soghrati, Soheil; Xiao, Fei; Nagarajan, Anand
2016-12-01
A Conforming to Interface Structured Adaptive Mesh Refinement (CISAMR) technique is introduced for the automated transformation of a structured grid into a conforming mesh with appropriate element aspect ratios. The CISAMR algorithm is composed of three main phases: (i) Structured Adaptive Mesh Refinement (SAMR) of the background grid; (ii) r-adaptivity of the nodes of elements cut by the crack; (iii) sub-triangulation of the elements deformed during the r-adaptivity process and those with hanging nodes generated during the SAMR process. The required considerations for the treatment of crack tips and branching cracks are also discussed in this manuscript. Regardless of the complexity of the problem geometry and without using iterative smoothing or optimization techniques, CISAMR ensures that aspect ratios of conforming elements are lower than three. Multiple numerical examples are presented to demonstrate the application of CISAMR for modeling linear elastic fracture problems with intricate morphologies.
NASA Astrophysics Data System (ADS)
Soghrati, Soheil; Xiao, Fei; Nagarajan, Anand
2017-04-01
A Conforming to Interface Structured Adaptive Mesh Refinement (CISAMR) technique is introduced for the automated transformation of a structured grid into a conforming mesh with appropriate element aspect ratios. The CISAMR algorithm is composed of three main phases: (i) Structured Adaptive Mesh Refinement (SAMR) of the background grid; (ii) r-adaptivity of the nodes of elements cut by the crack; (iii) sub-triangulation of the elements deformed during the r-adaptivity process and those with hanging nodes generated during the SAMR process. The required considerations for the treatment of crack tips and branching cracks are also discussed in this manuscript. Regardless of the complexity of the problem geometry and without using iterative smoothing or optimization techniques, CISAMR ensures that aspect ratios of conforming elements are lower than three. Multiple numerical examples are presented to demonstrate the application of CISAMR for modeling linear elastic fracture problems with intricate morphologies.
Advanced numerical methods in mesh generation and mesh adaptation
Lipnikov, Konstantine; Danilov, A; Vassilevski, Y; Agonzal, A
2010-01-01
Numerical solution of partial differential equations requires appropriate meshes, efficient solvers and robust and reliable error estimates. Generation of high-quality meshes for complex engineering models is a non-trivial task. This task is made more difficult when the mesh has to be adapted to a problem solution. This article is focused on a synergistic approach to the mesh generation and mesh adaptation, where best properties of various mesh generation methods are combined to build efficiently simplicial meshes. First, the advancing front technique (AFT) is combined with the incremental Delaunay triangulation (DT) to build an initial mesh. Second, the metric-based mesh adaptation (MBA) method is employed to improve quality of the generated mesh and/or to adapt it to a problem solution. We demonstrate with numerical experiments that combination of all three methods is required for robust meshing of complex engineering models. The key to successful mesh generation is the high-quality of the triangles in the initial front. We use a black-box technique to improve surface meshes exported from an unattainable CAD system. The initial surface mesh is refined into a shape-regular triangulation which approximates the boundary with the same accuracy as the CAD mesh. The DT method adds robustness to the AFT. The resulting mesh is topologically correct but may contain a few slivers. The MBA uses seven local operations to modify the mesh topology. It improves significantly the mesh quality. The MBA method is also used to adapt the mesh to a problem solution to minimize computational resources required for solving the problem. The MBA has a solid theoretical background. In the first two experiments, we consider the convection-diffusion and elasticity problems. We demonstrate the optimal reduction rate of the discretization error on a sequence of adaptive strongly anisotropic meshes. The key element of the MBA method is construction of a tensor metric from hierarchical edge
Adaptive, Tactical Mesh Networking: Control Base MANET Model
2010-09-01
pp. 316–320 Available: IEEE Xplore , http://ieeexplore.ieee.org [Accessed: June 9, 2010]. [5] N. Sidiropoulos, “Multiuser Transmit Beamforming...Mobile Mesh Segments of TNT Testbed .......... 11 Figure 5. Infrastructure and Ad Hoc Mode of IEEE 802.11................................ 13 Figure...6. The Power Spectral Density of OFDM................................................ 14 Figure 7. A Typical IEEE 802.16 Network
Unstructured mesh generation and adaptivity
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1995-01-01
An overview of current unstructured mesh generation and adaptivity techniques is given. Basic building blocks taken from the field of computational geometry are first described. Various practical mesh generation techniques based on these algorithms are then constructed and illustrated with examples. Issues of adaptive meshing and stretched mesh generation for anisotropic problems are treated in subsequent sections. The presentation is organized in an education manner, for readers familiar with computational fluid dynamics, wishing to learn more about current unstructured mesh techniques.
Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; ...
2015-06-30
The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as muchmore » geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer.« less
Ibrahim, Ahmad M.; Wilson, Paul P.H.; Sawan, Mohamed E.; Mosher, Scott W.; Peplow, Douglas E.; Wagner, John C.; Evans, Thomas M.; Grove, Robert E.
2015-06-30
The CADIS and FW-CADIS hybrid Monte Carlo/deterministic techniques dramatically increase the efficiency of neutronics modeling, but their use in the accurate design analysis of very large and geometrically complex nuclear systems has been limited by the large number of processors and memory requirements for their preliminary deterministic calculations and final Monte Carlo calculation. Three mesh adaptivity algorithms were developed to reduce the memory requirements of CADIS and FW-CADIS without sacrificing their efficiency improvement. First, a macromaterial approach enhances the fidelity of the deterministic models without changing the mesh. Second, a deterministic mesh refinement algorithm generates meshes that capture as much geometric detail as possible without exceeding a specified maximum number of mesh elements. Finally, a weight window coarsening algorithm decouples the weight window mesh and energy bins from the mesh and energy group structure of the deterministic calculations in order to remove the memory constraint of the weight window map from the deterministic mesh resolution. The three algorithms were used to enhance an FW-CADIS calculation of the prompt dose rate throughout the ITER experimental facility. Using these algorithms resulted in a 23.3% increase in the number of mesh tally elements in which the dose rates were calculated in a 10-day Monte Carlo calculation and, additionally, increased the efficiency of the Monte Carlo simulation by a factor of at least 3.4. The three algorithms enabled this difficult calculation to be accurately solved using an FW-CADIS simulation on a regular computer cluster, eliminating the need for a world-class super computer.
NASA Astrophysics Data System (ADS)
Jacobs, C. T.; Collins, G. S.; Piggott, M. D.; Kramer, S. C.; Wilson, C. R. G.
2013-02-01
Small-scale experiments of volcanic ash particle settling in water have demonstrated that ash particles can either settle slowly and individually, or rapidly and collectively as a gravitationally unstable ash-laden plume. This has important implications for the emplacement of tephra deposits on the seabed. Numerical modelling has the potential to extend the results of laboratory experiments to larger scales and explore the conditions under which plumes may form and persist, but many existing models are computationally restricted by the fixed mesh approaches that they employ. In contrast, this paper presents a new multiphase flow model that uses an adaptive unstructured mesh approach. As a simulation progresses, the mesh is optimized to focus numerical resolution in areas important to the dynamics and decrease it where it is not needed, thereby potentially reducing computational requirements. Model verification is performed using the method of manufactured solutions, which shows the correct solution convergence rates. Model validation and application considers 2-D simulations of plume formation in a water tank which replicate published laboratory experiments. The numerically predicted settling velocities for both individual particles and plumes, as well as instability behaviour, agree well with experimental data and observations. Plume settling is clearly hindered by the presence of a salinity gradient, and its influence must therefore be taken into account when considering particles in bodies of saline water. Furthermore, individual particles settle in the laminar flow regime while plume settling is shown (by plume Reynolds numbers greater than unity) to be in the turbulent flow regime, which has a significant impact on entrainment and settling rates. Mesh adaptivity maintains solution accuracy while providing a substantial reduction in computational requirements when compared to the same simulation performed using a fixed mesh, highlighting the benefits of an
A Predictive Model of Fragmentation using Adaptive Mesh Refinement and a Hierarchical Material Model
Koniges, A E; Masters, N D; Fisher, A C; Anderson, R W; Eder, D C; Benson, D; Kaiser, T B; Gunney, B T; Wang, P; Maddox, B R; Hansen, J F; Kalantar, D H; Dixit, P; Jarmakani, H; Meyers, M A
2009-03-03
Fragmentation is a fundamental material process that naturally spans spatial scales from microscopic to macroscopic. We developed a mathematical framework using an innovative combination of hierarchical material modeling (HMM) and adaptive mesh refinement (AMR) to connect the continuum to microstructural regimes. This framework has been implemented in a new multi-physics, multi-scale, 3D simulation code, NIF ALE-AMR. New multi-material volume fraction and interface reconstruction algorithms were developed for this new code, which is leading the world effort in hydrodynamic simulations that combine AMR with ALE (Arbitrary Lagrangian-Eulerian) techniques. The interface reconstruction algorithm is also used to produce fragments following material failure. In general, the material strength and failure models have history vector components that must be advected along with other properties of the mesh during remap stage of the ALE hydrodynamics. The fragmentation models are validated against an electromagnetically driven expanding ring experiment and dedicated laser-based fragmentation experiments conducted at the Jupiter Laser Facility. As part of the exit plan, the NIF ALE-AMR code was applied to a number of fragmentation problems of interest to the National Ignition Facility (NIF). One example shows the added benefit of multi-material ALE-AMR that relaxes the requirement that material boundaries must be along mesh boundaries.
An Adaptive Mesh Algorithm: Mesh Structure and Generation
Scannapieco, Anthony J.
2016-06-21
The purpose of Adaptive Mesh Refinement is to minimize spatial errors over the computational space not to minimize the number of computational elements. The additional result of the technique is that it may reduce the number of computational elements needed to retain a given level of spatial accuracy. Adaptive mesh refinement is a computational technique used to dynamically select, over a region of space, a set of computational elements designed to minimize spatial error in the computational model of a physical process. The fundamental idea is to increase the mesh resolution in regions where the physical variables are represented by a broad spectrum of modes in k-space, hence increasing the effective global spectral coverage of those physical variables. In addition, the selection of the spatially distributed elements is done dynamically by cyclically adjusting the mesh to follow the spectral evolution of the system. Over the years three types of AMR schemes have evolved; block, patch and locally refined AMR. In block and patch AMR logical blocks of various grid sizes are overlaid to span the physical space of interest, whereas in locally refined AMR no logical blocks are employed but locally nested mesh levels are used to span the physical space. The distinction between block and patch AMR is that in block AMR the original blocks refine and coarsen entirely in time, whereas in patch AMR the patches change location and zone size with time. The type of AMR described herein is a locally refi ned AMR. In the algorithm described, at any point in physical space only one zone exists at whatever level of mesh that is appropriate for that physical location. The dynamic creation of a locally refi ned computational mesh is made practical by a judicious selection of mesh rules. With these rules the mesh is evolved via a mesh potential designed to concentrate the nest mesh in regions where the physics is modally dense, and coarsen zones in regions where the physics is modally
Adaptive Mesh Expansion Model (AMEM) for Liver Segmentation from CT Image
Wang, Xuehu; Yang, Jian; Ai, Danni; Zheng, Yongchang; Tang, Songyuan; Wang, Yongtian
2015-01-01
This study proposes a novel adaptive mesh expansion model (AMEM) for liver segmentation from computed tomography images. The virtual deformable simplex model (DSM) is introduced to represent the mesh, in which the motion of each vertex can be easily manipulated. The balloon, edge, and gradient forces are combined with the binary image to construct the external force of the deformable model, which can rapidly drive the DSM to approach the target liver boundaries. Moreover, tangential and normal forces are combined with the gradient image to control the internal force, such that the DSM degree of smoothness can be precisely controlled. The triangular facet of the DSM is adaptively decomposed into smaller triangular components, which can significantly improve the segmentation accuracy of the irregularly sharp corners of the liver. The proposed method is evaluated on the basis of different criteria applied to 10 clinical data sets. Experiments demonstrate that the proposed AMEM algorithm is effective and robust and thus outperforms six other up-to-date algorithms. Moreover, AMEM can achieve a mean overlap error of 6.8% and a mean volume difference of 2.7%, whereas the average symmetric surface distance and the root mean square symmetric surface distance can reach 1.3 mm and 2.7 mm, respectively. PMID:25769030
A Parallel Ocean Model With Adaptive Mesh Refinement Capability For Global Ocean Prediction
Herrnstein, Aaron R.
2005-12-01
An ocean model with adaptive mesh refinement (AMR) capability is presented for simulating ocean circulation on decade time scales. The model closely resembles the LLNL ocean general circulation model with some components incorporated from other well known ocean models when appropriate. Spatial components are discretized using finite differences on a staggered grid where tracer and pressure variables are defined at cell centers and velocities at cell vertices (B-grid). Horizontal motion is modeled explicitly with leapfrog and Euler forward-backward time integration, and vertical motion is modeled semi-implicitly. New AMR strategies are presented for horizontal refinement on a B-grid, leapfrog time integration, and time integration of coupled systems with unequal time steps. These AMR capabilities are added to the LLNL software package SAMRAI (Structured Adaptive Mesh Refinement Application Infrastructure) and validated with standard benchmark tests. The ocean model is built on top of the amended SAMRAI library. The resulting model has the capability to dynamically increase resolution in localized areas of the domain. Limited basin tests are conducted using various refinement criteria and produce convergence trends in the model solution as refinement is increased. Carbon sequestration simulations are performed on decade time scales in domains the size of the North Atlantic and the global ocean. A suggestion is given for refinement criteria in such simulations. AMR predicts maximum pH changes and increases in CO_{2} concentration near the injection sites that are virtually unattainable with a uniform high resolution due to extremely long run times. Fine scale details near the injection sites are achieved by AMR with shorter run times than the finest uniform resolution tested despite the need for enhanced parallel performance. The North Atlantic simulations show a reduction in passive tracer errors when AMR is applied instead of a uniform coarse resolution. No
Anderson, R W; Pember, R B; Elliot, N S
2000-09-26
A new method for the solution of the unsteady Euler equations has been developed. The method combines staggered grid Lagrangian techniques with structured local adaptive mesh refinement (AMR). This method is a precursor to a more general adaptive arbitrary Lagrangian Eulerian (ALE-AMR) algorithm under development, which will facilitate the solution of problems currently at and beyond the boundary of soluble problems by traditional ALE methods by focusing computational resources where they are required. Many of the core issues involved in the development of the ALE-AMR method hinge upon the integration of AMR with a Lagrange step, which is the focus of the work described here. The novel components of the method are mainly driven by the need to reconcile traditional AMR techniques, which are typically employed on stationary meshes with cell-centered quantities, with the staggered grids and grid motion employed by Lagrangian methods. These new algorithmic components are first developed in one dimension and are then generalized to two dimensions. Solutions of several model problems involving shock hydrodynamics are presented and discussed.
Adaptive Mesh Refinement in CTH
Crawford, David
1999-05-04
This paper reports progress on implementing a new capability of adaptive mesh refinement into the Eulerian multimaterial shock- physics code CTH. The adaptivity is block-based with refinement and unrefinement occurring in an isotropic 2:1 manner. The code is designed to run on serial, multiprocessor and massive parallel platforms. An approximate factor of three in memory and performance improvements over comparable resolution non-adaptive calculations has-been demonstrated for a number of problems.
Adaptive Mesh Refinement for Microelectronic Device Design
NASA Technical Reports Server (NTRS)
Cwik, Tom; Lou, John; Norton, Charles
1999-01-01
Finite element and finite volume methods are used in a variety of design simulations when it is necessary to compute fields throughout regions that contain varying materials or geometry. Convergence of the simulation can be assessed by uniformly increasing the mesh density until an observable quantity stabilizes. Depending on the electrical size of the problem, uniform refinement of the mesh may be computationally infeasible due to memory limitations. Similarly, depending on the geometric complexity of the object being modeled, uniform refinement can be inefficient since regions that do not need refinement add to the computational expense. In either case, convergence to the correct (measured) solution is not guaranteed. Adaptive mesh refinement methods attempt to selectively refine the region of the mesh that is estimated to contain proportionally higher solution errors. The refinement may be obtained by decreasing the element size (h-refinement), by increasing the order of the element (p-refinement) or by a combination of the two (h-p refinement). A successful adaptive strategy refines the mesh to produce an accurate solution measured against the correct fields without undue computational expense. This is accomplished by the use of a) reliable a posteriori error estimates, b) hierarchal elements, and c) automatic adaptive mesh generation. Adaptive methods are also useful when problems with multi-scale field variations are encountered. These occur in active electronic devices that have thin doped layers and also when mixed physics is used in the calculation. The mesh needs to be fine at and near the thin layer to capture rapid field or charge variations, but can coarsen away from these layers where field variations smoothen and charge densities are uniform. This poster will present an adaptive mesh refinement package that runs on parallel computers and is applied to specific microelectronic device simulations. Passive sensors that operate in the infrared portion of
Multiphase flow modelling of explosive volcanic eruptions using adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Jacobs, Christian T.; Collins, Gareth S.; Piggott, Matthew D.; Kramer, Stephan C.
2014-05-01
Explosive volcanic eruptions generate highly energetic plumes of hot gas and ash particles that produce diagnostic deposits and pose an extreme environmental hazard. The formation, dispersion and collapse of these volcanic plumes are complex multiscale processes that are extremely challenging to simulate numerically. Accurate description of particle and droplet aggregation, movement and settling requires a model capable of capturing the dynamics on a range of scales (from cm to km) and a model that can correctly describe the important multiphase interactions that take place. However, even the most advanced models of eruption dynamics to date are restricted by the fixed mesh-based approaches that they employ. The research presented herein describes the development of a compressible multiphase flow model within Fluidity, a combined finite element / control volume computational fluid dynamics (CFD) code, for the study of explosive volcanic eruptions. Fluidity adopts a state-of-the-art adaptive unstructured mesh-based approach to discretise the domain and focus numerical resolution only in areas important to the dynamics, while decreasing resolution where it is not needed as a simulation progresses. This allows the accurate but economical representation of the flow dynamics throughout time, and potentially allows large multi-scale problems to become tractable in complex 3D domains. The multiphase flow model is verified with the method of manufactured solutions, and validated by simulating published gas-solid shock tube experiments and comparing the numerical results against pressure gauge data. The application of the model considers an idealised 7 km by 7 km domain in which the violent eruption of hot gas and volcanic ash high into the atmosphere is simulated. Although the simulations do not correspond to a particular eruption case study, the key flow features observed in a typical explosive eruption event are successfully captured. These include a shock wave resulting
toolkit computational mesh conceptual model.
Baur, David G.; Edwards, Harold Carter; Cochran, William K.; Williams, Alan B.; Sjaardema, Gregory D.
2010-03-01
The Sierra Toolkit computational mesh is a software library intended to support massively parallel multi-physics computations on dynamically changing unstructured meshes. This domain of intended use is inherently complex due to distributed memory parallelism, parallel scalability, heterogeneity of physics, heterogeneous discretization of an unstructured mesh, and runtime adaptation of the mesh. Management of this inherent complexity begins with a conceptual analysis and modeling of this domain of intended use; i.e., development of a domain model. The Sierra Toolkit computational mesh software library is designed and implemented based upon this domain model. Software developers using, maintaining, or extending the Sierra Toolkit computational mesh library must be familiar with the concepts/domain model presented in this report.
Anderson, R W; Pember, R B; Elliott, N S
2001-10-22
A new method that combines staggered grid Arbitrary Lagrangian-Eulerian (ALE) techniques with structured local adaptive mesh refinement (AMR) has been developed for solution of the Euler equations. This method facilitates the solution of problems currently at and beyond the boundary of soluble problems by traditional ALE methods by focusing computational resources where they are required through dynamic adaption. Many of the core issues involved in the development of the combined ALEAMR method hinge upon the integration of AMR with a staggered grid Lagrangian integration method. The novel components of the method are mainly driven by the need to reconcile traditional AMR techniques, which are typically employed on stationary meshes with cell-centered quantities, with the staggered grids and grid motion employed by Lagrangian methods. Numerical examples are presented which demonstrate the accuracy and efficiency of the method.
NASA Astrophysics Data System (ADS)
Rastigejev, Y.; Semakin, A. N.
2013-12-01
Accurate numerical simulations of global scale three-dimensional atmospheric chemical transport models (CTMs) are essential for studies of many important atmospheric chemistry problems such as adverse effect of air pollutants on human health, ecosystems and the Earth's climate. These simulations usually require large CPU time due to numerical difficulties associated with a wide range of spatial and temporal scales, nonlinearity and large number of reacting species. In our previous work we have shown that in order to achieve adequate convergence rate and accuracy, the mesh spacing in numerical simulation of global synoptic-scale pollution plume transport must be decreased to a few kilometers. This resolution is difficult to achieve for global CTMs on uniform or quasi-uniform grids. To address the described above difficulty we developed a three-dimensional Wavelet-based Adaptive Mesh Refinement (WAMR) algorithm. The method employs a highly non-uniform adaptive grid with fine resolution over the areas of interest without requiring small grid-spacing throughout the entire domain. The method uses multi-grid iterative solver that naturally takes advantage of a multilevel structure of the adaptive grid. In order to represent the multilevel adaptive grid efficiently, a dynamic data structure based on indirect memory addressing has been developed. The data structure allows rapid access to individual points, fast inter-grid operations and re-gridding. The WAMR method has been implemented on parallel computer architectures. The parallel algorithm is based on run-time partitioning and load-balancing scheme for the adaptive grid. The partitioning scheme maintains locality to reduce communications between computing nodes. The parallel scheme was found to be cost-effective. Specifically we obtained an order of magnitude increase in computational speed for numerical simulations performed on a twelve-core single processor workstation. We have applied the WAMR method for numerical
Projections of grounding line retreat in West Antarctica carried out with an adaptive mesh model
NASA Astrophysics Data System (ADS)
Cornford, Stephen; Payne, Antony; Martin, Daniel; Le Brocq, Anne
2013-04-01
Present and future sea level rise associated with mass loss from West Antarctica is typically attributed to marine glaciers retreating in response to a warming ocean. Warmer waters melt the floating ice shelves that restrain some, if not all, marine glaciers, and the glaciers themselves respond by speeding up. That leads to thinning and in turn grounding line retreat. Satellite observations indicate that Amundsen Sea Embayment and, in particular, Pine Island Glacier, are undergoing this kind of dynamic change today. Numerical models, however, struggle to reproduce the observed behavior because either high resolution or some other kind special treatment is required at the grounding line. We present 200-year projections of three major glacier systems of West Antarctica: those that drain into the Amundsen Sea , the Filchner-Ronne Ice Shelf and the Ross Ice shelf. We do so using the newly developed BISICLES ice sheet model, which employs adaptive mesh refinement to maintain sub-kilometer resolution close to the grounding line and coarser resolution elsewhere. Ice accumulation and ice shelf melt-rate are derived from a range of models of the Antarctic atmosphere and ocean forced by the SRES A1B and E1 scenarios. We find that a substantial proportion of the grounding line in West Antarctica retreats, however the total sea level rise is less than 50 mm by 2100, and less than 100 mm by 2200. The lion's share of the mass loss is attributed to Pine Island Glacier, while its immediate neighbor Thwaites Glacier does not retreat until the end of the simulations.
Numerical Modelling of Volcanic Ash Settling in Water Using Adaptive Unstructured Meshes
NASA Astrophysics Data System (ADS)
Jacobs, C. T.; Collins, G. S.; Piggott, M. D.; Kramer, S. C.; Wilson, C. R.
2011-12-01
At the bottom of the world's oceans lies layer after layer of ash deposited from past volcanic eruptions. Correct interpretation of these layers can provide important constraints on the duration and frequency of volcanism, but requires a full understanding of the complex multi-phase settling and deposition process. Analogue experiments of tephra settling through a tank of water demonstrate that small ash particles can either settle individually, or collectively as a gravitationally unstable ash-laden plume. These plumes are generated when the concentration of particles exceeds a certain threshold such that the density of the tephra-water mixture is sufficiently large relative to the underlying particle-free water for a gravitational Rayleigh-Taylor instability to develop. These ash-laden plumes are observed to descend as a vertical density current at a velocity much greater than that of single particles, which has important implications for the emplacement of tephra deposits on the seabed. To extend the results of laboratory experiments to large scales and explore the conditions under which vertical density currents may form and persist, we have developed a multi-phase extension to Fluidity, a combined finite element / control volume CFD code that uses adaptive unstructured meshes. As a model validation, we present two- and three-dimensional simulations of tephra plume formation in a water tank that replicate laboratory experiments (Carey, 1997, doi:10.1130/0091-7613(1997)025<0839:IOCSOT>2.3.CO;2). An inflow boundary condition at the top of the domain allows particles to flux in at a constant rate of 0.472 gm-2s-1, forming a near-surface layer of tephra particles, which initially settle individually at the predicted Stokes velocity of 1.7 mms-1. As more tephra enters the water and the particle concentration increases, the layer eventually becomes unstable and plumes begin to form, descending with velocities more than ten times greater than those of individual
Lee, Won Hee; Kim, Tae-Seong
2012-01-01
This study proposes an advanced finite element (FE) head modeling technique through which high-resolution FE meshes adaptive to the degree of tissue anisotropy can be generated. Our adaptive meshing scheme (called wMesh) uses MRI structural information and fractional anisotropy maps derived from diffusion tensors in the FE mesh generation process, optimally reflecting electrical properties of the human brain. We examined the characteristics of the wMeshes through various qualitative and quantitative comparisons to the conventional FE regular-sized meshes that are non-adaptive to the degree of white matter anisotropy. We investigated numerical differences in the FE forward solutions that include the electrical potential and current density generated by current sources in the brain. The quantitative difference was calculated by two statistical measures of relative difference measure (RDM) and magnification factor (MAG). The results show that the wMeshes are adaptive to the anisotropic density of the WM anisotropy, and they better reflect the density and directionality of tissue conductivity anisotropy. Our comparison results between various anisotropic regular mesh and wMesh models show that there are substantial differences in the EEG forward solutions in the brain (up to RDM=0.48 and MAG=0.63 in the electrical potential, and RDM=0.65 and MAG=0.52 in the current density). Our analysis results indicate that the wMeshes produce different forward solutions that are different from the conventional regular meshes. We present some results that the wMesh head modeling approach enhances the sensitivity and accuracy of the FE solutions at the interfaces or in the regions where the anisotropic conductivities change sharply or their directional changes are complex. The fully automatic wMesh generation technique should be useful for modeling an individual-specific and high-resolution anisotropic FE head model incorporating realistic anisotropic conductivity distributions
NASA Astrophysics Data System (ADS)
Donmez, Orhan
We present a general procedure to solve the General Relativistic Hydrodynamical (GRH) equations with Adaptive-Mesh Refinement (AMR) and model of an accretion disk around a black hole. To do this, the GRH equations are written in a conservative form to exploit their hyperbolic character. The numerical solutions of the general relativistic hydrodynamic equations is done by High Resolution Shock Capturing schemes (HRSC), specifically designed to solve non-linear hyperbolic systems of conservation laws. These schemes depend on the characteristic information of the system. We use Marquina fluxes with MUSCL left and right states to solve GRH equations. First, we carry out different test problems with uniform and AMR grids on the special relativistic hydrodynamics equations to verify the second order convergence of the code in 1D, 2 D and 3D. Second, we solve the GRH equations and use the general relativistic test problems to compare the numerical solutions with analytic ones. In order to this, we couple the flux part of general relativistic hydrodynamic equation with a source part using Strang splitting. The coupling of the GRH equations is carried out in a treatment which gives second order accurate solutions in space and time. The test problems examined include shock tubes, geodesic flows, and circular motion of particle around the black hole. Finally, we apply this code to the accretion disk problems around the black hole using the Schwarzschild metric at the background of the computational domain. We find spiral shocks on the accretion disk. They are observationally expected results. We also examine the star-disk interaction near a massive black hole. We find that when stars are grounded down or a hole is punched on the accretion disk, they create shock waves which destroy the accretion disk.
NASA Astrophysics Data System (ADS)
Viré, Axelle; Xiang, Jiansheng; Milthaler, Frank; Farrell, Patrick Emmet; Piggott, Matthew David; Latham, John-Paul; Pavlidis, Dimitrios; Pain, Christopher Charles
2012-12-01
Fluid-structure interactions are modelled by coupling the finite element fluid/ocean model `Fluidity-ICOM' with a combined finite-discrete element solid model `Y3D'. Because separate meshes are used for the fluids and solids, the present method is flexible in terms of discretisation schemes used for each material. Also, it can tackle multiple solids impacting on one another, without having ill-posed problems in the resolution of the fluid's equations. Importantly, the proposed approach ensures that Newton's third law is satisfied at the discrete level. This is done by first computing the action-reaction force on a supermesh, i.e. a function superspace of the fluid and solid meshes, and then projecting it to both meshes to use it as a source term in the fluid and solid equations. This paper demonstrates the properties of spatial conservation and accuracy of the method for a sphere immersed in a fluid, with prescribed fluid and solid velocities. While spatial conservation is shown to be independent of the mesh resolutions, accuracy requires fine resolutions in both fluid and solid meshes. It is further highlighted that unstructured meshes adapted to the solid concentration field reduce the numerical errors, in comparison with uniformly structured meshes with the same number of elements. The method is verified on flow past a falling sphere. Its potential for ocean applications is further shown through the simulation of vortex-induced vibrations of two cylinders and the flow past two flexible fibres.
NASA Astrophysics Data System (ADS)
Lichtenberg, Tim; Schleicher, Dominik R. G.
2015-07-01
The astonishing diversity in the observed planetary population requires theoretical efforts and advances in planet formation theories. The use of numerical approaches provides a method to tackle the weaknesses of current models and is an important tool to close gaps in poorly constrained areas such as the rapid formation of giant planets in highly evolved systems. So far, most numerical approaches make use of Lagrangian-based smoothed-particle hydrodynamics techniques or grid-based 2D axisymmetric simulations. We present a new global disk setup to model the first stages of giant planet formation via gravitational instabilities (GI) in 3D with the block-structured adaptive mesh refinement (AMR) hydrodynamics code enzo. With this setup, we explore the potential impact of AMR techniques on the fragmentation and clumping due to large-scale instabilities using different AMR configurations. Additionally, we seek to derive general resolution criteria for global simulations of self-gravitating disks of variable extent. We run a grid of simulations with varying AMR settings, including runs with a static grid for comparison. Additionally, we study the effects of varying the disk radius. The physical settings involve disks with Rdisk = 10,100 and 300 AU, with a mass of Mdisk ≈ 0.05 M⊙ and a central object of subsolar mass (M⋆ = 0.646 M⊙). To validate our thermodynamical approach we include a set of simulations with a dynamically stable profile (Qinit = 3) and similar grid parameters. The development of fragmentation and the buildup of distinct clumps in the disk is strongly dependent on the chosen AMR grid settings. By combining our findings from the resolution and parameter studies we find a general lower limit criterion to be able to resolve GI induced fragmentation features and distinct clumps, which induce turbulence in the disk and seed giant planet formation. Irrespective of the physical extension of the disk, topologically disconnected clump features are only
Adaptive mesh refinement in titanium
Colella, Phillip; Wen, Tong
2005-01-21
In this paper, we evaluate Titanium's usability as a high-level parallel programming language through a case study, where we implement a subset of Chombo's functionality in Titanium. Chombo is a software package applying the Adaptive Mesh Refinement methodology to numerical Partial Differential Equations at the production level. In Chombo, the library approach is used to parallel programming (C++ and Fortran, with MPI), whereas Titanium is a Java dialect designed for high-performance scientific computing. The performance of our implementation is studied and compared with that of Chombo in solving Poisson's equation based on two grid configurations from a real application. Also provided are the counts of lines of code from both sides.
Floating shock fitting via Lagrangian adaptive meshes
NASA Technical Reports Server (NTRS)
Vanrosendale, John
1995-01-01
In recent work we have formulated a new approach to compressible flow simulation, combining the advantages of shock-fitting and shock-capturing. Using a cell-centered on Roe scheme discretization on unstructured meshes, we warp the mesh while marching to steady state, so that mesh edges align with shocks and other discontinuities. This new algorithm, the Shock-fitting Lagrangian Adaptive Method (SLAM), is, in effect, a reliable shock-capturing algorithm which yields shock-fitted accuracy at convergence.
Divett, T; Vennell, R; Stevens, C
2013-02-28
At tidal energy sites, large arrays of hundreds of turbines will be required to generate economically significant amounts of energy. Owing to wake effects within the array, the placement of turbines within will be vital to capturing the maximum energy from the resource. This study presents preliminary results using Gerris, an adaptive mesh flow solver, to investigate the flow through four different arrays of 15 turbines each. The goal is to optimize the position of turbines within an array in an idealized channel. The turbines are represented as areas of increased bottom friction in an adaptive mesh model so that the flow and power capture in tidally reversing flow through large arrays can be studied. The effect of oscillating tides is studied, with interesting dynamics generated as the tidal current reverses direction, forcing turbulent flow through the array. The energy removed from the flow by each of the four arrays is compared over a tidal cycle. A staggered array is found to extract 54 per cent more energy than a non-staggered array. Furthermore, an array positioned to one side of the channel is found to remove a similar amount of energy compared with an array in the centre of the channel.
PARAMESH: A Parallel Adaptive Mesh Refinement Community Toolkit
NASA Technical Reports Server (NTRS)
MacNeice, Peter; Olson, Kevin M.; Mobarry, Clark; deFainchtein, Rosalinda; Packer, Charles
1999-01-01
In this paper, we describe a community toolkit which is designed to provide parallel support with adaptive mesh capability for a large and important class of computational models, those using structured, logically cartesian meshes. The package of Fortran 90 subroutines, called PARAMESH, is designed to provide an application developer with an easy route to extend an existing serial code which uses a logically cartesian structured mesh into a parallel code with adaptive mesh refinement. Alternatively, in its simplest use, and with minimal effort, it can operate as a domain decomposition tool for users who want to parallelize their serial codes, but who do not wish to use adaptivity. The package can provide them with an incremental evolutionary path for their code, converting it first to uniformly refined parallel code, and then later if they so desire, adding adaptivity.
Current sheets, reconnection and adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Marliani, Christiane
1998-11-01
Adaptive structured mesh refinement methods have proved to be an appropriate tool for the numerical study of a variety of problems where largely separated length scales are involved, e.g. [R. Grauer, C. Marliani, K. Germaschewski, PRL, 80, 4177, (1998)]. A typical example in plasma physics are the current sheets in magnetohydrodynamic flows. Their dynamics is investigated in the framework of incompressible MHD. We present simulations of the ideal and inviscid dynamics in two and three dimensions. In addition, we show numerical simulations for the resistive case in two dimensions. Specifically, we show simulations for the case of the doubly periodic coalescence instability. At the onset of the reconnection process the kinetic energy rises and drops rapidly and afterwards settles into an oscillatory phase. The timescale of the magnetic reconnection process is not affected by these fast events but consistent with the Sweet-Parker model of stationary reconnection. Taking into account the electron inertia terms in the generalized Ohm's law the electron skin depth is introduced as an additional parameter. The modified equations allow for magnetic reconnection in the collisionless regime. Current density and vorticity concentrate in extremely long and thin sheets. Their dynamics becomes numerically accessible by means of adaptive mesh refinement.
Grid adaption using Chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1993-01-01
The objective of this paper is to perform grid adaptation using composite over-lapping meshes in regions of large gradient to capture the salient features accurately during computation. The Chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using tri-linear interpolation. Applications to the Euler equations for shock reflections and to a shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well resolved.
Grid adaptation using chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1994-01-01
The objective of this paper is to perform grid adaptation using composite overlapping meshes in regions of large gradient to accurately capture the salient features during computation. The chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using trilinear interpolation. Application to the Euler equations for shock reflections and to shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well-resolved.
Grid adaptation using Chimera composite overlapping meshes
NASA Technical Reports Server (NTRS)
Kao, Kai-Hsiung; Liou, Meng-Sing; Chow, Chuen-Yen
1993-01-01
The objective of this paper is to perform grid adaptation using composite over-lapping meshes in regions of large gradient to capture the salient features accurately during computation. The Chimera grid scheme, a multiple overset mesh technique, is used in combination with a Navier-Stokes solver. The numerical solution is first converged to a steady state based on an initial coarse mesh. Solution-adaptive enhancement is then performed by using a secondary fine grid system which oversets on top of the base grid in the high-gradient region, but without requiring the mesh boundaries to join in any special way. Communications through boundary interfaces between those separated grids are carried out using tri-linear interpolation. Applications to the Euler equations for shock reflections and to a shock wave/boundary layer interaction problem are tested. With the present method, the salient features are well resolved.
NASA Astrophysics Data System (ADS)
Guo, Zhikui; Chen, Chao; Tao, Chunhui
2016-04-01
Since 2007, there are four China Da yang cruises (CDCs), which have been carried out to investigate polymetallic sulfides in the southwest Indian ridge (SWIR) and have acquired both gravity data and bathymetry data on the corresponding survey lines(Tao et al., 2014). Sandwell et al. (2014) published a new global marine gravity model including the free air gravity data and its first order vertical gradient (Vzz). Gravity data and its gradient can be used to extract unknown density structure information(e.g. crust thickness) under surface of the earth, but they contain all the mass effect under the observation point. Therefore, how to get accurate gravity and its gradient effect of the existing density structure (e.g. terrain) has been a key issue. Using the bathymetry data or ETOPO1 (http://www.ngdc.noaa.gov/mgg/global/global.html) model at a full resolution to calculate the terrain effect could spend too much computation time. We expect to develop an effective method that takes less time but can still yield the desired accuracy. In this study, a constant-density polyhedral model is used to calculate the gravity field and its vertical gradient, which is based on the work of Tsoulis (2012). According to gravity field attenuation with distance and variance of bathymetry, we present an adaptive mesh refinement and coarsening strategies to merge both global topography data and multi-beam bathymetry data. The local coarsening or size of mesh depends on user-defined accuracy and terrain variation (Davis et al., 2011). To depict terrain better, triangular surface element and rectangular surface element are used in fine and coarse mesh respectively. This strategy can also be applied to spherical coordinate in large region and global scale. Finally, we applied this method to calculate Bouguer gravity anomaly (BGA), mantle Bouguer anomaly(MBA) and their vertical gradient in SWIR. Further, we compared the result with previous results in the literature. Both synthetic model
Vertical Scan (V-SCAN) for 3-D Grid Adaptive Mesh Refinement for an atmospheric Model Dynamical Core
NASA Astrophysics Data System (ADS)
Andronova, N. G.; Vandenberg, D.; Oehmke, R.; Stout, Q. F.; Penner, J. E.
2009-12-01
One of the major building blocks of a rigorous representation of cloud evolution in global atmospheric models is a parallel adaptive grid MPI-based communication library (an Adaptive Blocks for Locally Cartesian Topologies library -- ABLCarT), which manages the block-structured data layout, handles ghost cell updates among neighboring blocks and splits a block as refinements occur. The library has several modules that provide a layer of abstraction for adaptive refinement: blocks, which contain individual cells of user data; shells - the global geometry for the problem, including a sphere, reduced sphere, and now a 3D sphere; a load balancer for placement of blocks onto processors; and a communication support layer which encapsulates all data movement. A major performance concern with adaptive mesh refinement is how to represent calculations that have need to be sequenced in a particular order in a direction, such as calculating integrals along a specific path (e.g. atmospheric pressure or geopotential in the vertical dimension). This concern is compounded if the blocks have varying levels of refinement, or are scattered across different processors, as can be the case in parallel computing. In this paper we describe an implementation in ABLCarT of a vertical scan operation, which allows computing along vertical paths in the correct order across blocks transparent to their resolution and processor location. We test this functionality on a 2D and a 3D advection problem, which tests the performance of the model’s dynamics (transport) and physics (sources and sinks) for different model resolutions needed for inclusion of cloud formation.
An Adaptive Mesh Algorithm: Mapping the Mesh Variables
Scannapieco, Anthony J.
2016-07-25
Both thermodynamic and kinematic variables must be mapped. The kinematic variables are defined on a separate kinematic mesh; it is the duel mesh to the thermodynamic mesh. The map of the kinematic variables is done by calculating the contributions of kinematic variables on the old thermodynamic mesh, mapping the kinematic variable contributions onto the new thermodynamic mesh and then synthesizing the mapped kinematic variables on the new kinematic mesh. In this document the map of the thermodynamic variables will be described.
Mesh saliency with adaptive local patches
NASA Astrophysics Data System (ADS)
Nouri, Anass; Charrier, Christophe; Lézoray, Olivier
2015-03-01
3D object shapes (represented by meshes) include both areas that attract the visual attention of human observers and others less or not attractive at all. This visual attention depends on the degree of saliency exposed by these areas. In this paper, we propose a technique for detecting salient regions in meshes. To do so, we define a local surface descriptor based on local patches of adaptive size and filled with a local height field. The saliency of mesh vertices is then defined as its degree measure with edges weights computed from adaptive patch similarities. Our approach is compared to the state-of-the-art and presents competitive results. A study evaluating the influence of the parameters establishing this approach is also carried out. The strength and the stability of our approach with respect to noise and simplification are also studied.
A hierarchical structure for automatic meshing and adaptive FEM analysis
NASA Technical Reports Server (NTRS)
Kela, Ajay; Saxena, Mukul; Perucchio, Renato
1987-01-01
A new algorithm for generating automatically, from solid models of mechanical parts, finite element meshes that are organized as spatially addressable quaternary trees (for 2-D work) or octal trees (for 3-D work) is discussed. Because such meshes are inherently hierarchical as well as spatially addressable, they permit efficient substructuring techniques to be used for both global analysis and incremental remeshing and reanalysis. The global and incremental techniques are summarized and some results from an experimental closed loop 2-D system in which meshing, analysis, error evaluation, and remeshing and reanalysis are done automatically and adaptively are presented. The implementation of 3-D work is briefly discussed.
Unstructured Adaptive Meshes: Bad for Your Memory?
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Feng, Hui-Yu; VanderWijngaart, Rob
2003-01-01
This viewgraph presentation explores the need for a NASA Advanced Supercomputing (NAS) parallel benchmark for problems with irregular dynamical memory access. This benchmark is important and necessary because: 1) Problems with localized error source benefit from adaptive nonuniform meshes; 2) Certain machines perform poorly on such problems; 3) Parallel implementation may provide further performance improvement but is difficult. Some examples of problems which use irregular dynamical memory access include: 1) Heat transfer problem; 2) Heat source term; 3) Spectral element method; 4) Base functions; 5) Elemental discrete equations; 6) Global discrete equations. Nonconforming Mesh and Mortar Element Method are covered in greater detail in this presentation.
Multigrid solution strategies for adaptive meshing problems
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1995-01-01
This paper discusses the issues which arise when combining multigrid strategies with adaptive meshing techniques for solving steady-state problems on unstructured meshes. A basic strategy is described, and demonstrated by solving several inviscid and viscous flow cases. Potential inefficiencies in this basic strategy are exposed, and various alternate approaches are discussed, some of which are demonstrated with an example. Although each particular approach exhibits certain advantages, all methods have particular drawbacks, and the formulation of a completely optimal strategy is considered to be an open problem.
Cartesian-cell based grid generation and adaptive mesh refinement
NASA Technical Reports Server (NTRS)
Coirier, William J.; Powell, Kenneth G.
1993-01-01
Viewgraphs on Cartesian-cell based grid generation and adaptive mesh refinement are presented. Topics covered include: grid generation; cell cutting; data structures; flow solver formulation; adaptive mesh refinement; and viscous flow.
GRChombo: Numerical relativity with adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Clough, Katy; Figueras, Pau; Finkel, Hal; Kunesch, Markus; Lim, Eugene A.; Tunyasuvunakool, Saran
2015-12-01
In this work, we introduce {\\mathtt{GRChombo}}: a new numerical relativity code which incorporates full adaptive mesh refinement (AMR) using block structured Berger-Rigoutsos grid generation. The code supports non-trivial ‘many-boxes-in-many-boxes’ mesh hierarchies and massive parallelism through the message passing interface. {\\mathtt{GRChombo}} evolves the Einstein equation using the standard BSSN formalism, with an option to turn on CCZ4 constraint damping if required. The AMR capability permits the study of a range of new physics which has previously been computationally infeasible in a full 3 + 1 setting, while also significantly simplifying the process of setting up the mesh for these problems. We show that {\\mathtt{GRChombo}} can stably and accurately evolve standard spacetimes such as binary black hole mergers and scalar collapses into black holes, demonstrate the performance characteristics of our code, and discuss various physics problems which stand to benefit from the AMR technique.
Floating shock fitting via Lagrangian adaptive meshes
NASA Technical Reports Server (NTRS)
Vanrosendale, John
1994-01-01
In recent works we have formulated a new approach to compressible flow simulation, combining the advantages of shock-fitting and shock-capturing. Using a cell-centered Roe scheme discretization on unstructured meshes, we warp the mesh while marching to steady state, so that mesh edges align with shocks and other discontinuities. This new algorithm, the Shock-fitting Lagrangian Adaptive Method (SLAM) is, in effect, a reliable shock-capturing algorithm which yields shock-fitted accuracy at convergence. Shock-capturing algorithms like this, which warp the mesh to yield shock-fitted accuracy, are new and relatively untried. However, their potential is clear. In the context of sonic booms, accurate calculation of near-field sonic boom signatures is critical to the design of the High Speed Civil Transport (HSCT). SLAM should allow computation of accurate N-wave pressure signatures on comparatively coarse meshes, significantly enhancing our ability to design low-boom configurations for high-speed aircraft.
Details of tetrahedral anisotropic mesh adaptation
NASA Astrophysics Data System (ADS)
Jensen, Kristian Ejlebjerg; Gorman, Gerard
2016-04-01
We have implemented tetrahedral anisotropic mesh adaptation using the local operations of coarsening, swapping, refinement and smoothing in MATLAB without the use of any for- N loops, i.e. the script is fully vectorised. In the process of doing so, we have made three observations related to details of the implementation: 1. restricting refinement to a single edge split per element not only simplifies the code, it also improves mesh quality, 2. face to edge swapping is unnecessary, and 3. optimising for the Vassilevski functional tends to give a little higher value for the mean condition number functional than optimising for the condition number functional directly. These observations have been made for a uniform and a radial shock metric field, both starting from a structured mesh in a cube. Finally, we compare two coarsening techniques and demonstrate the importance of applying smoothing in the mesh adaptation loop. The results pertain to a unit cube geometry, but we also show the effect of corners and edges by applying the implementation in a spherical geometry.
PLUM: Parallel Load Balancing for Adaptive Unstructured Meshes
NASA Technical Reports Server (NTRS)
Oliker, Leonid; Biswas, Rupak; Saini, Subhash (Technical Monitor)
1998-01-01
Mesh adaption is a powerful tool for efficient unstructured-grid computations but causes load imbalance among processors on a parallel machine. We present a novel method called PLUM to dynamically balance the processor workloads with a global view. This paper presents the implementation and integration of all major components within our dynamic load balancing strategy for adaptive grid calculations. Mesh adaption, repartitioning, processor assignment, and remapping are critical components of the framework that must be accomplished rapidly and efficiently so as not to cause a significant overhead to the numerical simulation. A data redistribution model is also presented that predicts the remapping cost on the SP2. This model is required to determine whether the gain from a balanced workload distribution offsets the cost of data movement. Results presented in this paper demonstrate that PLUM is an effective dynamic load balancing strategy which remains viable on a large number of processors.
White Dwarf Mergers on Adaptive Meshes
NASA Astrophysics Data System (ADS)
Katz, Maximilian Peter
The mergers of binary white dwarf systems are potential progenitors of astrophysical explosions such as Type Ia supernovae. These white dwarfs can merge either by orbital decay through the emission of gravitational waves or by direct collisions as a result of orbital perturbations. The coalescence of the stars may ignite nuclear fusion, resulting in the destruction of both stars through a thermonuclear runaway and ensuing detonation. The goal of this dissertation is to simulate binary white dwarf systems using the techniques of computational fluid dynamics and therefore to understand what numerical techniques are necessary to obtain accurate dynamical evolution of the system, as well as to learn what conditions are necessary to enable a realistic detonation. For this purpose I have used software that solves the relevant fluid equations, the Poisson equation for self-gravity, and the systems governing nuclear reactions between atomic species. These equations are modeled on a computational domain that uses the technique of adaptive mesh refinement to have the highest spatial resolution in the areas of the domain that are most sensitive to the need for accurate numerical evolution. I have identified that the most important obstacles to accurate evolution are the numerical violation of conservation of energy and angular momentum in the system, and the development of numerically seeded thermonuclear detonations that do not bear resemblance to physically correct detonations. I then developed methods for ameliorating these problems, and determined what metrics can be used for judging whether a given white dwarf merger simulation is trustworthy. This involved the development of a number of algorithmic improvements to the simulation software, which I describe. Finally, I performed high-resolution simulations of typical cases of white dwarf mergers and head-on collisions to demonstrate the impacts of these choices. The results of these simulations and the corresponding
On Adaptive Mesh Generation in Two-Dimensions
D'Azevedo, E.
1999-10-11
This work considers the effectiveness of using anisotropic coordinate transformation in adaptive mesh generation. The anisotropic coordinate transformation is derived by interpreting the Hessian matrix of the data function as a metric tensor that measures the local approximation error. The Hessian matrix contains information about the local curvature of the surface and gives guidance in the aspect ratio and orientation for mesh generation. Since theoretically, an asymptotically optimally efficient mesh can be produced by transforming a regular mesh of optimal shape elements, it would be interesting to compare this approach with existing techniques in solution adaptive meshes. PLTMG , a general elliptic solver, is used to generate solution adapted triangular meshes for comparison. The solver has the capability of performing a posteriori error estimates in performing longest edge refinement, vertex unrefinement and mesh smoothing. Numerical experiments on three simple problems suggest the methodology employed in PLTMG is effective in generating near optimally efficient meshes.
Adaptive Mesh Refinement for ICF Calculations
NASA Astrophysics Data System (ADS)
Fyfe, David
2005-10-01
This paper describes our use of the package PARAMESH to create an Adaptive Mesh Refinement (AMR) version of NRL's FASTRAD3D code. PARAMESH was designed to create an MPI-based AMR code from a block structured serial code such as FASTRAD3D. FASTRAD3D is a compressible hydrodynamics code containing the physical effects relevant for the simulation of high-temperature plasmas including inertial confinement fusion (ICF) Rayleigh-Taylor unstable direct drive laser targets. These effects include inverse bremmstrahlung laser energy absorption, classical flux-limited Spitzer thermal conduction, real (table look-up) equation-of-state with either separate or identical electron and ion temperatures, multi-group variable Eddington radiation transport, and multi-group alpha particle transport and thermonuclear burn. Numerically, this physics requires an elliptic solver and a ray tracing approach on the AMR grid, which is the main subject of this paper. A sample ICF calculation will be presented. MacNeice et al., ``PARAMESH: A parallel adaptive mesh refinement community tool,'' Computer Physics Communications, 126 (2000), pp. 330-354.
An adaptive mesh-moving and refinement procedure for one-dimensional conservation laws
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Flaherty, Joseph E.; Arney, David C.
1993-01-01
We examine the performance of an adaptive mesh-moving and /or local mesh refinement procedure for the finite difference solution of one-dimensional hyperbolic systems of conservation laws. Adaptive motion of a base mesh is designed to isolate spatially distinct phenomena, and recursive local refinement of the time step and cells of the stationary or moving base mesh is performed in regions where a refinement indicator exceeds a prescribed tolerance. These adaptive procedures are incorporated into a computer code that includes a MacCormack finite difference scheme wih Davis' artificial viscosity model and a discretization error estimate based on Richardson's extrapolation. Experiments are conducted on three problems in order to qualify the advantages of adaptive techniques relative to uniform mesh computations and the relative benefits of mesh moving and refinement. Key results indicate that local mesh refinement, with and without mesh moving, can provide reliable solutions at much lower computational cost than possible on uniform meshes; that mesh motion can be used to improve the results of uniform mesh solutions for a modest computational effort; that the cost of managing the tree data structure associated with refinement is small; and that a combination of mesh motion and refinement reliably produces solutions for the least cost per unit accuracy.
Adaptive Meshing Techniques for Viscous Flow Calculations on Mixed Element Unstructured Meshes
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1997-01-01
An adaptive refinement strategy based on hierarchical element subdivision is formulated and implemented for meshes containing arbitrary mixtures of tetrahendra, hexahendra, prisms and pyramids. Special attention is given to keeping memory overheads as low as possible. This procedure is coupled with an algebraic multigrid flow solver which operates on mixed-element meshes. Inviscid flows as well as viscous flows are computed an adaptively refined tetrahedral, hexahedral, and hybrid meshes. The efficiency of the method is demonstrated by generating an adapted hexahedral mesh containing 3 million vertices on a relatively inexpensive workstation.
Visualization of Scalar Adaptive Mesh Refinement Data
VACET; Weber, Gunther; Weber, Gunther H.; Beckner, Vince E.; Childs, Hank; Ligocki, Terry J.; Miller, Mark C.; Van Straalen, Brian; Bethel, E. Wes
2007-12-06
Adaptive Mesh Refinement (AMR) is a highly effective computation method for simulations that span a large range of spatiotemporal scales, such as astrophysical simulations, which must accommodate ranges from interstellar to sub-planetary. Most mainstream visualization tools still lack support for AMR grids as a first class data type and AMR code teams use custom built applications for AMR visualization. The Department of Energy's (DOE's) Science Discovery through Advanced Computing (SciDAC) Visualization and Analytics Center for Enabling Technologies (VACET) is currently working on extending VisIt, which is an open source visualization tool that accommodates AMR as a first-class data type. These efforts will bridge the gap between general-purpose visualization applications and highly specialized AMR visual analysis applications. Here, we give an overview of the state of the art in AMR scalar data visualization research.
Elliptic Solvers for Adaptive Mesh Refinement Grids
Quinlan, D.J.; Dendy, J.E., Jr.; Shapira, Y.
1999-06-03
We are developing multigrid methods that will efficiently solve elliptic problems with anisotropic and discontinuous coefficients on adaptive grids. The final product will be a library that provides for the simplified solution of such problems. This library will directly benefit the efforts of other Laboratory groups. The focus of this work is research on serial and parallel elliptic algorithms and the inclusion of our black-box multigrid techniques into this new setting. The approach applies the Los Alamos object-oriented class libraries that greatly simplify the development of serial and parallel adaptive mesh refinement applications. In the final year of this LDRD, we focused on putting the software together; in particular we completed the final AMR++ library, we wrote tutorials and manuals, and we built example applications. We implemented the Fast Adaptive Composite Grid method as the principal elliptic solver. We presented results at the Overset Grid Conference and other more AMR specific conferences. We worked on optimization of serial and parallel performance and published several papers on the details of this work. Performance remains an important issue and is the subject of continuing research work.
Tetrahedral and Hexahedral Mesh Adaptation for CFD Problems
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Strawn, Roger C.; Chancellor, Marisa K. (Technical Monitor)
1997-01-01
This paper presents two unstructured mesh adaptation schemes for problems in computational fluid dynamics. The procedures allow localized grid refinement and coarsening to efficiently capture aerodynamic flow features of interest. The first procedure is for purely tetrahedral grids; unfortunately, repeated anisotropic adaptation may significantly deteriorate the quality of the mesh. Hexahedral elements, on the other hand, can be subdivided anisotropically without mesh quality problems. Furthermore, hexahedral meshes yield more accurate solutions than their tetrahedral counterparts for the same number of edges. Both the tetrahedral and hexahedral mesh adaptation procedures use edge-based data structures that facilitate efficient subdivision by allowing individual edges to be marked for refinement or coarsening. However, for hexahedral adaptation, pyramids, prisms, and tetrahedra are used as buffer elements between refined and unrefined regions to eliminate hanging vertices. Computational results indicate that the hexahedral adaptation procedure is a viable alternative to adaptive tetrahedral schemes.
Carpet: Adaptive Mesh Refinement for the Cactus Framework
NASA Astrophysics Data System (ADS)
Schnetter, Erik; Hawley, Scott; Hawke, Ian
2016-11-01
Carpet is an adaptive mesh refinement and multi-patch driver for the Cactus Framework (ascl:1102.013). Cactus is a software framework for solving time-dependent partial differential equations on block-structured grids, and Carpet acts as driver layer providing adaptive mesh refinement, multi-patch capability, as well as parallelization and efficient I/O.
A parallel adaptive mesh refinement algorithm
NASA Technical Reports Server (NTRS)
Quirk, James J.; Hanebutte, Ulf R.
1993-01-01
Over recent years, Adaptive Mesh Refinement (AMR) algorithms which dynamically match the local resolution of the computational grid to the numerical solution being sought have emerged as powerful tools for solving problems that contain disparate length and time scales. In particular, several workers have demonstrated the effectiveness of employing an adaptive, block-structured hierarchical grid system for simulations of complex shock wave phenomena. Unfortunately, from the parallel algorithm developer's viewpoint, this class of scheme is quite involved; these schemes cannot be distilled down to a small kernel upon which various parallelizing strategies may be tested. However, because of their block-structured nature such schemes are inherently parallel, so all is not lost. In this paper we describe the method by which Quirk's AMR algorithm has been parallelized. This method is built upon just a few simple message passing routines and so it may be implemented across a broad class of MIMD machines. Moreover, the method of parallelization is such that the original serial code is left virtually intact, and so we are left with just a single product to support. The importance of this fact should not be underestimated given the size and complexity of the original algorithm.
Zhang, S.; Yuen, D.A.; Zhu, A.; Song, S.; George, D.L.
2011-01-01
We parallelized the GeoClaw code on one-level grid using OpenMP in March, 2011 to meet the urgent need of simulating tsunami waves at near-shore from Tohoku 2011 and achieved over 75% of the potential speed-up on an eight core Dell Precision T7500 workstation [1]. After submitting that work to SC11 - the International Conference for High Performance Computing, we obtained an unreleased OpenMP version of GeoClaw from David George, who developed the GeoClaw code as part of his PH.D thesis. In this paper, we will show the complementary characteristics of the two approaches used in parallelizing GeoClaw and the speed-up obtained by combining the advantage of each of the two individual approaches with adaptive mesh refinement (AMR), demonstrating the capabilities of running GeoClaw efficiently on many-core systems. We will also show a novel simulation of the Tohoku 2011 Tsunami waves inundating the Sendai airport and Fukushima Nuclear Power Plants, over which the finest grid distance of 20 meters is achieved through a 4-level AMR. This simulation yields quite good predictions about the wave-heights and travel time of the tsunami waves. ?? 2011 IEEE.
AMR++: Object-Oriented Parallel Adaptive Mesh Refinement
Quinlan, D.; Philip, B.
2000-02-02
Adaptive mesh refinement (AMR) computations are complicated by their dynamic nature. The development of solvers for realistic applications is complicated by both the complexity of the AMR and the geometry of realistic problem domains. The additional complexity of distributed memory parallelism within such AMR applications most commonly exceeds the level of complexity that can be reasonable maintained with traditional approaches toward software development. This paper will present the details of our object-oriented work on the simplification of the use of adaptive mesh refinement on applications with complex geometries for both serial and distributed memory parallel computation. We will present an independent set of object-oriented abstractions (C++ libraries) well suited to the development of such seemingly intractable scientific computations. As an example of the use of this object-oriented approach we will present recent results of an application modeling fluid flow in the eye. Within this example, the geometry is too complicated for a single curvilinear coordinate grid and so a set of overlapping curvilinear coordinate grids' are used. Adaptive mesh refinement and the required grid generation work to support the refinement process is coupled together in the solution of essentially elliptic equations within this domain. This paper will focus on the management of complexity within development of the AMR++ library which forms a part of the Overture object-oriented framework for the solution of partial differential equations within scientific computing.
Elliptic Solvers with Adaptive Mesh Refinement on Complex Geometries
Phillip, B.
2000-07-24
Adaptive Mesh Refinement (AMR) is a numerical technique for locally tailoring the resolution computational grids. Multilevel algorithms for solving elliptic problems on adaptive grids include the Fast Adaptive Composite grid method (FAC) and its parallel variants (AFAC and AFACx). Theory that confirms the independence of the convergence rates of FAC and AFAC on the number of refinement levels exists under certain ellipticity and approximation property conditions. Similar theory needs to be developed for AFACx. The effectiveness of multigrid-based elliptic solvers such as FAC, AFAC, and AFACx on adaptively refined overlapping grids is not clearly understood. Finally, a non-trivial eye model problem will be solved by combining the power of using overlapping grids for complex moving geometries, AMR, and multilevel elliptic solvers.
Adaptive mesh fluid simulations on GPU
NASA Astrophysics Data System (ADS)
Wang, Peng; Abel, Tom; Kaehler, Ralf
2010-10-01
We describe an implementation of compressible inviscid fluid solvers with block-structured adaptive mesh refinement on Graphics Processing Units using NVIDIA's CUDA. We show that a class of high resolution shock capturing schemes can be mapped naturally on this architecture. Using the method of lines approach with the second order total variation diminishing Runge-Kutta time integration scheme, piecewise linear reconstruction, and a Harten-Lax-van Leer Riemann solver, we achieve an overall speedup of approximately 10 times faster execution on one graphics card as compared to a single core on the host computer. We attain this speedup in uniform grid runs as well as in problems with deep AMR hierarchies. Our framework can readily be applied to more general systems of conservation laws and extended to higher order shock capturing schemes. This is shown directly by an implementation of a magneto-hydrodynamic solver and comparing its performance to the pure hydrodynamic case. Finally, we also combined our CUDA parallel scheme with MPI to make the code run on GPU clusters. Close to ideal speedup is observed on up to four GPUs.
Procedure for Adapting Direct Simulation Monte Carlo Meshes
NASA Technical Reports Server (NTRS)
Woronowicz, Michael S.; Wilmoth, Richard G.; Carlson, Ann B.; Rault, Didier F. G.
1992-01-01
A technique is presented for adapting computational meshes used in the G2 version of the direct simulation Monte Carlo method. The physical ideas underlying the technique are discussed, and adaptation formulas are developed for use on solutions generated from an initial mesh. The effect of statistical scatter on adaptation is addressed, and results demonstrate the ability of this technique to achieve more accurate results without increasing necessary computational resources.
Adaptive mesh refinement for stochastic reaction-diffusion processes
Bayati, Basil; Chatelain, Philippe; Koumoutsakos, Petros
2011-01-01
We present an algorithm for adaptive mesh refinement applied to mesoscopic stochastic simulations of spatially evolving reaction-diffusion processes. The transition rates for the diffusion process are derived on adaptive, locally refined structured meshes. Convergence of the diffusion process is presented and the fluctuations of the stochastic process are verified. Furthermore, a refinement criterion is proposed for the evolution of the adaptive mesh. The method is validated in simulations of reaction-diffusion processes as described by the Fisher-Kolmogorov and Gray-Scott equations.
GAMER: GPU-accelerated Adaptive MEsh Refinement code
NASA Astrophysics Data System (ADS)
Schive, Hsi-Yu; Tsai, Yu-Chih; Chiueh, Tzihong
2016-12-01
GAMER (GPU-accelerated Adaptive MEsh Refinement) serves as a general-purpose adaptive mesh refinement + GPU framework and solves hydrodynamics with self-gravity. The code supports adaptive mesh refinement (AMR), hydrodynamics with self-gravity, and a variety of GPU-accelerated hydrodynamic and Poisson solvers. It also supports hybrid OpenMP/MPI/GPU parallelization, concurrent CPU/GPU execution for performance optimization, and Hilbert space-filling curve for load balance. Although the code is designed for simulating galaxy formation, it can be easily modified to solve a variety of applications with different governing equations. All optimization strategies implemented in the code can be inherited straightforwardly.
Serial and parallel dynamic adaptation of general hybrid meshes
NASA Astrophysics Data System (ADS)
Kavouklis, Christos
The Navier-Stokes equations are a standard mathematical representation of viscous fluid flow. Their numerical solution in three dimensions remains a computationally intensive and challenging task, despite recent advances in computer speed and memory. A strategy to increase accuracy of Navier-Stokes simulations, while maintaining computing resources to a minimum, is local refinement of the associated computational mesh in regions of large solution gradients and coarsening in regions where the solution does not vary appreciably. In this work we consider adaptation of general hybrid meshes for Computational Fluid Dynamics (CFD) applications. Hybrid meshes are composed of four types of elements; hexahedra, prisms, pyramids and tetrahedra, and have been proven a promising technology in accurately resolving fluid flow for complex geometries. The first part of this dissertation is concerned with the design and implementation of a serial scheme for the adaptation of general three dimensional hybrid meshes. We have defined 29 refinement types, for all four kinds of elements. The core of the present adaptation scheme is an iterative algorithm that flags mesh edges for refinement, so that the adapted mesh is conformal. Of primary importance is considered the design of a suitable dynamic data structure that facilitates refinement and coarsening operations and furthermore minimizes memory requirements. A special dynamic list is defined for mesh elements, in contrast with the usual tree structures. It contains only elements of the current adaptation step and minimal information that is utilized to reconstruct parent elements when the mesh is coarsened. In the second part of this work, a new parallel dynamic mesh adaptation and load balancing algorithm for general hybrid meshes is presented. Partitioning of a hybrid mesh reduces to partitioning of the corresponding dual graph. Communication among processors is based on the faces of the interpartition boundary. The distributed
Lee, W; Kim, T-S; Cho, M; Lee, S
2005-01-01
In studying bioelectromagnetic problems, finite element method offers several advantages over other conventional methods such as boundary element method. It allows truly volumetric analysis and incorporation of material properties such as anisotropy. Mesh generation is the first requirement in the finite element analysis and there are many different approaches in mesh generation. However conventional approaches offered by commercial packages and various algorithms do not generate content-adaptive meshes, resulting in numerous elements in the smaller volume regions, thereby increasing computational load and demand. In this work, we present an improved content-adaptive mesh generation scheme that is efficient and fast along with options to change the contents of meshes. For demonstration, mesh models of the head from a volume MRI are presented in 2-D and 3-D.
A two-dimensional adaptive mesh generation method
NASA Astrophysics Data System (ADS)
Altas, Irfan; Stephenson, John W.
1991-05-01
The present, two-dimensional adaptive mesh-generation method allows selective modification of a small portion of the mesh without affecting large areas of adjacent mesh-points, and is applicable with or without boundary-fitted coordinate-generation procedures. The cases of differential equation discretization by, on the one hand, classical difference formulas designed for uniform meshes, and on the other the present difference formulas, are illustrated through the application of the method to the Hiemenz flow for which the Navier-Stokes equation's exact solution is known, as well as to a two-dimensional viscous internal flow problem.
Adaptive Mesh and Algorithm Refinement Using Direct Simulation Monte Carlo
NASA Astrophysics Data System (ADS)
Garcia, Alejandro L.; Bell, John B.; Crutchfield, William Y.; Alder, Berni J.
1999-09-01
Adaptive mesh and algorithm refinement (AMAR) embeds a particle method within a continuum method at the finest level of an adaptive mesh refinement (AMR) hierarchy. The coupling between the particle region and the overlaying continuum grid is algorithmically equivalent to that between the fine and coarse levels of AMR. Direct simulation Monte Carlo (DSMC) is used as the particle algorithm embedded within a Godunov-type compressible Navier-Stokes solver. Several examples are presented and compared with purely continuum calculations.
Parallel adaptive mesh refinement for electronic structure calculations
Kohn, S.; Weare, J.; Ong, E.; Baden, S.
1996-12-01
We have applied structured adaptive mesh refinement techniques to the solution of the LDA equations for electronic structure calculations. Local spatial refinement concentrates memory resources and numerical effort where it is most needed, near the atomic centers and in regions of rapidly varying charge density. The structured grid representation enables us to employ efficient iterative solver techniques such as conjugate gradients with multigrid preconditioning. We have parallelized our solver using an object-oriented adaptive mesh refinement framework.
Friedman, A.; Miller, W. Jr.
1993-12-31
The program was divided into segments: (Week 1) geometric modeling and mesh generation (Weeks 2 and 3) error estimation and adaptive strategies. Participants in the program came from a wide variety of disciplines dealing with remarkably analogous problems in this area. Ideas were exchanged and interdisciplinary collaboration was initiated in informal contexts as well as in the talks and question periods. In the talks, a number of algorithms were described along with specific applications to problems of great current interest in various scientific and engineering disciplines. In this emerging field, participants developed criteria for evaluation of algorithms and established guidelines for selection of algorithms appropriate to any specific problem. Special features of a problem may include curved surfaces, complicated boundaries, evolving interfaces (such as occur in coating flows), and/or criticality of error estimation.
Adaptive superposition of finite element meshes in linear and nonlinear dynamic analysis
NASA Astrophysics Data System (ADS)
Yue, Zhihua
2005-11-01
The numerical analysis of transient phenomena in solids, for instance, wave propagation and structural dynamics, is a very important and active area of study in engineering. Despite the current evolutionary state of modern computer hardware, practical analysis of large scale, nonlinear transient problems requires the use of adaptive methods where computational resources are locally allocated according to the interpolation requirements of the solution form. Adaptive analysis of transient problems involves obtaining solutions at many different time steps, each of which requires a sequence of adaptive meshes. Therefore, the execution speed of the adaptive algorithm is of paramount importance. In addition, transient problems require that the solution must be passed from one adaptive mesh to the next adaptive mesh with a bare minimum of solution-transfer error since this form of error compromises the initial conditions used for the next time step. A new adaptive finite element procedure (s-adaptive) is developed in this study for modeling transient phenomena in both linear elastic solids and nonlinear elastic solids caused by progressive damage. The adaptive procedure automatically updates the time step size and the spatial mesh discretization in transient analysis, achieving the accuracy and the efficiency requirements simultaneously. The novel feature of the s-adaptive procedure is the original use of finite element mesh superposition to produce spatial refinement in transient problems. The use of mesh superposition enables the s-adaptive procedure to completely avoid the need for cumbersome multipoint constraint algorithms and mesh generators, which makes the s-adaptive procedure extremely fast. Moreover, the use of mesh superposition enables the s-adaptive procedure to minimize the solution-transfer error. In a series of different solid mechanics problem types including 2-D and 3-D linear elastic quasi-static problems, 2-D material nonlinear quasi-static problems
Lee, W H; Kim, T-S; Cho, M H; Ahn, Y B; Lee, S Y
2006-12-07
In studying bioelectromagnetic problems, finite element analysis (FEA) offers several advantages over conventional methods such as the boundary element method. It allows truly volumetric analysis and incorporation of material properties such as anisotropic conductivity. For FEA, mesh generation is the first critical requirement and there exist many different approaches. However, conventional approaches offered by commercial packages and various algorithms do not generate content-adaptive meshes (cMeshes), resulting in numerous nodes and elements in modelling the conducting domain, and thereby increasing computational load and demand. In this work, we present efficient content-adaptive mesh generation schemes for complex biological volumes of MR images. The presented methodology is fully automatic and generates FE meshes that are adaptive to the geometrical contents of MR images, allowing optimal representation of conducting domain for FEA. We have also evaluated the effect of cMeshes on FEA in three dimensions by comparing the forward solutions from various cMesh head models to the solutions from the reference FE head model in which fine and equidistant FEs constitute the model. The results show that there is a significant gain in computation time with minor loss in numerical accuracy. We believe that cMeshes should be useful in the FEA of bioelectromagnetic problems.
NASA Astrophysics Data System (ADS)
Lee, W. H.; Kim, T.-S.; Cho, M. H.; Ahn, Y. B.; Lee, S. Y.
2006-12-01
In studying bioelectromagnetic problems, finite element analysis (FEA) offers several advantages over conventional methods such as the boundary element method. It allows truly volumetric analysis and incorporation of material properties such as anisotropic conductivity. For FEA, mesh generation is the first critical requirement and there exist many different approaches. However, conventional approaches offered by commercial packages and various algorithms do not generate content-adaptive meshes (cMeshes), resulting in numerous nodes and elements in modelling the conducting domain, and thereby increasing computational load and demand. In this work, we present efficient content-adaptive mesh generation schemes for complex biological volumes of MR images. The presented methodology is fully automatic and generates FE meshes that are adaptive to the geometrical contents of MR images, allowing optimal representation of conducting domain for FEA. We have also evaluated the effect of cMeshes on FEA in three dimensions by comparing the forward solutions from various cMesh head models to the solutions from the reference FE head model in which fine and equidistant FEs constitute the model. The results show that there is a significant gain in computation time with minor loss in numerical accuracy. We believe that cMeshes should be useful in the FEA of bioelectromagnetic problems.
Adaptive Mesh Refinement Algorithms for Parallel Unstructured Finite Element Codes
Parsons, I D; Solberg, J M
2006-02-03
This project produced algorithms for and software implementations of adaptive mesh refinement (AMR) methods for solving practical solid and thermal mechanics problems on multiprocessor parallel computers using unstructured finite element meshes. The overall goal is to provide computational solutions that are accurate to some prescribed tolerance, and adaptivity is the correct path toward this goal. These new tools will enable analysts to conduct more reliable simulations at reduced cost, both in terms of analyst and computer time. Previous academic research in the field of adaptive mesh refinement has produced a voluminous literature focused on error estimators and demonstration problems; relatively little progress has been made on producing efficient implementations suitable for large-scale problem solving on state-of-the-art computer systems. Research issues that were considered include: effective error estimators for nonlinear structural mechanics; local meshing at irregular geometric boundaries; and constructing efficient software for parallel computing environments.
Simulation of nonpoint source contamination based on adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Kourakos, G.; Harter, T.
2014-12-01
Contamination of groundwater aquifers from nonpoint sources is a worldwide problem. Typical agricultural groundwater basins receive contamination from a large array (in the order of ~10^5-6) of spatially and temporally heterogeneous sources such as fields, crops, dairies etc, while the received contaminants emerge at significantly uncertain time lags to a large array of discharge surfaces such as public supply, domestic and irrigation wells and streams. To support decision making in such complex regimes several approaches have been developed, which can be grouped into 3 categories: i) Index methods, ii)regression methods and iii) physically based methods. Among the three, physically based methods are considered more accurate, but at the cost of computational demand. In this work we present a physically based simulation framework which exploits the latest hardware and software developments to simulate large (>>1,000 km2) groundwater basins. First we simulate groundwater flow using a sufficiently detailed mesh to capture the spatial heterogeneity. To achieve optimal mesh quality we combine adaptive mesh refinement with the nonlinear solution for unconfined flow. Starting from a coarse grid the mesh is refined iteratively in the parts of the domain where the flow heterogeneity appears higher resulting in optimal grid. Secondly we simulate the nonpoint source pollution based on the detailed velocity field computed from the previous step. In our approach we use the streamline model where the 3D transport problem is decomposed into multiple 1D transport problems. The proposed framework is applied to simulate nonpoint source pollution in the Central Valley aquifer system, California.
Development and Verification of Unstructured Adaptive Mesh Technique with Edge Compatibility
NASA Astrophysics Data System (ADS)
Ito, Kei; Kunugi, Tomoaki; Ohshima, Hiroyuki
In the design study of the large-sized sodium-cooled fast reactor (JSFR), one key issue is suppression of gas entrainment (GE) phenomena at a gas-liquid interface. Therefore, the authors have been developed a high-precision CFD algorithm to evaluate the GE phenomena accurately. The CFD algorithm has been developed on unstructured meshes to establish an accurate modeling of JSFR system. For two-phase interfacial flow simulations, a high-precision volume-of-fluid algorithm is employed. It was confirmed that the developed CFD algorithm could reproduce the GE phenomena in a simple GE experiment. Recently, the authors have been developed an important technique for the simulation of the GE phenomena in JSFR. That is an unstructured adaptive mesh technique which can apply fine cells dynamically to the region where the GE occurs in JSFR. In this paper, as a part of the development, a two-dimensional unstructured adaptive mesh technique is discussed. In the two-dimensional adaptive mesh technique, each cell is refined isotropically to reduce distortions of the mesh. In addition, connection cells are formed to eliminate the edge incompatibility between refined and non-refined cells. The two-dimensional unstructured adaptive mesh technique is verified by solving well-known lid-driven cavity flow problem. As a result, the two-dimensional unstructured adaptive mesh technique succeeds in providing a high-precision solution, even though poor-quality distorted initial mesh is employed. In addition, the simulation error on the two-dimensional unstructured adaptive mesh is much less than the error on the structured mesh with a larger number of cells.
Adaptive upscaling with the dual mesh method
Guerillot, D.; Verdiere, S.
1997-08-01
The objective of this paper is to demonstrate that upscaling should be calculated during the flow simulation instead of trying to enhance the a priori upscaling methods. Hence, counter-examples are given to motivate our approach, the so-called Dual Mesh Method. The main steps of this numerical algorithm are recalled. Applications illustrate the necessity to consider different average relative permeability values depending on the direction in space. Moreover, these values could be different for the same average saturation. This proves that an a priori upscaling cannot be the answer even in homogeneous cases because of the {open_quotes}dynamical heterogeneity{close_quotes} created by the saturation profile. Other examples show the efficiency of the Dual Mesh Method applied to heterogeneous medium and to an actual field case in South America.
Lober, R.R.; Tautges, T.J.; Vaughan, C.T.
1997-03-01
Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.
Huang, W.; Zheng, Lingyun; Zhan, X.
2002-01-01
Accurate modelling of groundwater flow and transport with sharp moving fronts often involves high computational cost, when a fixed/uniform mesh is used. In this paper, we investigate the modelling of groundwater problems using a particular adaptive mesh method called the moving mesh partial differential equation approach. With this approach, the mesh is dynamically relocated through a partial differential equation to capture the evolving sharp fronts with a relatively small number of grid points. The mesh movement and physical system modelling are realized by solving the mesh movement and physical partial differential equations alternately. The method is applied to the modelling of a range of groundwater problems, including advection dominated chemical transport and reaction, non-linear infiltration in soil, and the coupling of density dependent flow and transport. Numerical results demonstrate that sharp moving fronts can be accurately and efficiently captured by the moving mesh approach. Also addressed are important implementation strategies, e.g. the construction of the monitor function based on the interpolation error, control of mesh concentration, and two-layer mesh movement. Copyright ?? 2002 John Wiley and Sons, Ltd.
Parallel tetrahedral mesh adaptation with dynamic load balancing
Oliker, Leonid; Biswas, Rupak; Gabow, Harold N.
2000-06-28
The ability to dynamically adapt an unstructured grid is a powerful tool for efficiently solving computational problems with evolving physical features. In this paper, we report on our experience parallelizing an edge-based adaptation scheme, called 3D-TAG, using message passing. Results show excellent speedup when a realistic helicopter rotor mesh is randomly refined. However, performance deteriorates when the mesh is refined using a solution-based error indicator since mesh adaptation for practical problems occurs in a localized region, creating a severe load imbalance. To address this problem, we have developed PLUM, a global dynamic load balancing framework for adaptive numerical computations. Even though PLUM primarily balances processor workloads for the solution phase, it reduces the load imbalance problem within mesh adaptation by repartitioning the mesh after targeting edges for refinement but before the actual subdivision. This dramatically improves the performance of parallel 3D-TAG since refinement occurs in a more load balanced fashion. We also present optimal and heuristic algorithms that, when applied to the default mapping of a parallel repartitioner, significantly reduce the data redistribution overhead. Finally, portability is examined by comparing performance on three state-of-the-art parallel machines.
Parallel Tetrahedral Mesh Adaptation with Dynamic Load Balancing
NASA Technical Reports Server (NTRS)
Oliker, Leonid; Biswas, Rupak; Gabow, Harold N.
1999-01-01
The ability to dynamically adapt an unstructured grid is a powerful tool for efficiently solving computational problems with evolving physical features. In this paper, we report on our experience parallelizing an edge-based adaptation scheme, called 3D_TAG. using message passing. Results show excellent speedup when a realistic helicopter rotor mesh is randomly refined. However. performance deteriorates when the mesh is refined using a solution-based error indicator since mesh adaptation for practical problems occurs in a localized region., creating a severe load imbalance. To address this problem, we have developed PLUM, a global dynamic load balancing framework for adaptive numerical computations. Even though PLUM primarily balances processor workloads for the solution phase, it reduces the load imbalance problem within mesh adaptation by repartitioning the mesh after targeting edges for refinement but before the actual subdivision. This dramatically improves the performance of parallel 3D_TAG since refinement occurs in a more load balanced fashion. We also present optimal and heuristic algorithms that, when applied to the default mapping of a parallel repartitioner, significantly reduce the data redistribution overhead. Finally, portability is examined by comparing performance on three state-of-the-art parallel machines.
Parallel adaptation of general three-dimensional hybrid meshes
Kavouklis, Christos Kallinderis, Yannis
2010-05-01
A new parallel dynamic mesh adaptation and load balancing algorithm for general hybrid grids has been developed. The meshes considered in this work are composed of four kinds of elements; tetrahedra, prisms, hexahedra and pyramids, which poses a challenge to parallel mesh adaptation. Additional complexity imposed by the presence of multiple types of elements affects especially data migration, updates of local data structures and interpartition data structures. Efficient partition of hybrid meshes has been accomplished by transforming them to suitable graphs and using serial graph partitioning algorithms. Communication among processors is based on the faces of the interpartition boundary and the termination detection algorithm of Dijkstra is employed to ensure proper flagging of edges for refinement. An inexpensive dynamic load balancing strategy is introduced to redistribute work load among processors after adaptation. In particular, only the initial coarse mesh, with proper weighting, is balanced which yields savings in computation time and relatively simple implementation of mesh quality preservation rules, while facilitating coarsening of refined elements. Special algorithms are employed for (i) data migration and dynamic updates of the local data structures, (ii) determination of the resulting interpartition boundary and (iii) identification of the communication pattern of processors. Several representative applications are included to evaluate the method.
Adaptive meshing technique applied to an orthopaedic finite element contact problem.
Roarty, Colleen M; Grosland, Nicole M
2004-01-01
Finite element methods have been applied extensively and with much success in the analysis of orthopaedic implants. Recently a growing interest has developed, in the orthopaedic biomechanics community, in how numerical models can be constructed for the optimal solution of problems in contact mechanics. New developments in this area are of paramount importance in the design of improved implants for orthopaedic surgery. Finite element and other computational techniques are widely applied in the analysis and design of hip and knee implants, with additional joints (ankle, shoulder, wrist) attracting increased attention. The objective of this investigation was to develop a simplified adaptive meshing scheme to facilitate the finite element analysis of a dual-curvature total wrist implant. Using currently available software, the analyst has great flexibility in mesh generation, but must prescribe element sizes and refinement schemes throughout the domain of interest. Unfortunately, it is often difficult to predict in advance a mesh spacing that will give acceptable results. Adaptive finite-element mesh capabilities operate to continuously refine the mesh to improve accuracy where it is required, with minimal intervention by the analyst. Such mesh adaptation generally means that in certain areas of the analysis domain, the size of the elements is decreased (or increased) and/or the order of the elements may be increased (or decreased). In concept, mesh adaptation is very appealing. Although there have been several previous applications of adaptive meshing for in-house FE codes, we have coupled an adaptive mesh formulation with the pre-existing commercial programs PATRAN (MacNeal-Schwendler Corp., USA) and ABAQUS (Hibbit Karlson and Sorensen, Pawtucket, RI). In doing so, we have retained several attributes of the commercial software, which are very attractive for orthopaedic implant applications.
A structured multi-block solution-adaptive mesh algorithm with mesh quality assessment
NASA Technical Reports Server (NTRS)
Ingram, Clint L.; Laflin, Kelly R.; Mcrae, D. Scott
1995-01-01
The dynamic solution adaptive grid algorithm, DSAGA3D, is extended to automatically adapt 2-D structured multi-block grids, including adaption of the block boundaries. The extension is general, requiring only input data concerning block structure, connectivity, and boundary conditions. Imbedded grid singular points are permitted, but must be prevented from moving in space. Solutions for workshop cases 1 and 2 are obtained on multi-block grids and illustrate both increased resolution of and alignment with the solution. A mesh quality assessment criteria is proposed to determine how well a given mesh resolves and aligns with the solution obtained upon it. The criteria is used to evaluate the grid quality for solutions of workshop case 6 obtained on both static and dynamically adapted grids. The results indicate that this criteria shows promise as a means of evaluating resolution.
Adaptive mesh refinement and adjoint methods in geophysics simulations
NASA Astrophysics Data System (ADS)
Burstedde, Carsten
2013-04-01
It is an ongoing challenge to increase the resolution that can be achieved by numerical geophysics simulations. This applies to considering sub-kilometer mesh spacings in global-scale mantle convection simulations as well as to using frequencies up to 1 Hz in seismic wave propagation simulations. One central issue is the numerical cost, since for three-dimensional space discretizations, possibly combined with time stepping schemes, a doubling of resolution can lead to an increase in storage requirements and run time by factors between 8 and 16. A related challenge lies in the fact that an increase in resolution also increases the dimensionality of the model space that is needed to fully parametrize the physical properties of the simulated object (a.k.a. earth). Systems that exhibit a multiscale structure in space are candidates for employing adaptive mesh refinement, which varies the resolution locally. An example that we found well suited is the mantle, where plate boundaries and fault zones require a resolution on the km scale, while deeper area can be treated with 50 or 100 km mesh spacings. This approach effectively reduces the number of computational variables by several orders of magnitude. While in this case it is possible to derive the local adaptation pattern from known physical parameters, it is often unclear what are the most suitable criteria for adaptation. We will present the goal-oriented error estimation procedure, where such criteria are derived from an objective functional that represents the observables to be computed most accurately. Even though this approach is well studied, it is rarely used in the geophysics community. A related strategy to make finer resolution manageable is to design methods that automate the inference of model parameters. Tweaking more than a handful of numbers and judging the quality of the simulation by adhoc comparisons to known facts and observations is a tedious task and fundamentally limited by the turnaround times
Multigrid solution of internal flows using unstructured solution adaptive meshes
NASA Technical Reports Server (NTRS)
Smith, Wayne A.; Blake, Kenneth R.
1992-01-01
This is the final report of the NASA Lewis SBIR Phase 2 Contract Number NAS3-25785, Multigrid Solution of Internal Flows Using Unstructured Solution Adaptive Meshes. The objective of this project, as described in the Statement of Work, is to develop and deliver to NASA a general three-dimensional Navier-Stokes code using unstructured solution-adaptive meshes for accuracy and multigrid techniques for convergence acceleration. The code will primarily be applied, but not necessarily limited, to high speed internal flows in turbomachinery.
NASA Astrophysics Data System (ADS)
Sarkis, C.; Silva, L.; Gandin, Ch-A.; Plapp, M.
2016-03-01
Dendritic growth is computed with automatic adaptation of an anisotropic and unstructured finite element mesh. The energy conservation equation is formulated for solid and liquid phases considering an interface balance that includes the Gibbs-Thomson effect. An equation for a diffuse interface is also developed by considering a phase field function with constant negative value in the liquid and constant positive value in the solid. Unknowns are the phase field function and a dimensionless temperature, as proposed by [1]. Linear finite element interpolation is used for both variables, and discretization stabilization techniques ensure convergence towards a correct non-oscillating solution. In order to perform quantitative computations of dendritic growth on a large domain, two additional numerical ingredients are necessary: automatic anisotropic unstructured adaptive meshing [2,[3] and parallel implementations [4], both made available with the numerical platform used (CimLib) based on C++ developments. Mesh adaptation is found to greatly reduce the number of degrees of freedom. Results of phase field simulations for dendritic solidification of a pure material in two and three dimensions are shown and compared with reference work [1]. Discussion on algorithm details and the CPU time will be outlined.
Adaptive mesh refinement for shocks and material interfaces
Dai, William Wenlong
2010-01-01
There are three kinds of adaptive mesh refinement (AMR) in structured meshes. Block-based AMR sometimes over refines meshes. Cell-based AMR treats cells cell by cell and thus loses the advantage of the nature of structured meshes. Patch-based AMR is intended to combine advantages of block- and cell-based AMR, i.e., the nature of structured meshes and sharp regions of refinement. But, patch-based AMR has its own difficulties. For example, patch-based AMR typically cannot preserve symmetries of physics problems. In this paper, we will present an approach for a patch-based AMR for hydrodynamics simulations. The approach consists of clustering, symmetry preserving, mesh continuity, flux correction, communications, management of patches, and load balance. The special features of this patch-based AMR include symmetry preserving, efficiency of refinement across shock fronts and material interfaces, special implementation of flux correction, and patch management in parallel computing environments. To demonstrate the capability of the AMR framework, we will show both two- and three-dimensional hydrodynamics simulations with many levels of refinement.
Numerical simulation of immiscible viscous fingering using adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Adam, A.; Salinas, P.; Percival, J. R.; Pavlidis, D.; Pain, C.; Muggeridge, A. H.; Jackson, M.
2015-12-01
Displacement of one fluid by another in porous media occurs in various settings including hydrocarbon recovery, CO2 storage and water purification. When the invading fluid is of lower viscosity than the resident fluid, the displacement front is subject to a Saffman-Taylor instability and is unstable to transverse perturbations. These instabilities can grow, leading to fingering of the invading fluid. Numerical simulation of viscous fingering is challenging. The physics is controlled by a complex interplay of viscous and diffusive forces and it is necessary to ensure physical diffusion dominates numerical diffusion to obtain converged solutions. This typically requires the use of high mesh resolution and high order numerical methods. This is computationally expensive. We demonstrate here the use of a novel control volume - finite element (CVFE) method along with dynamic unstructured mesh adaptivity to simulate viscous fingering with higher accuracy and lower computational cost than conventional methods. Our CVFE method employs a discontinuous representation for both pressure and velocity, allowing the use of smaller control volumes (CVs). This yields higher resolution of the saturation field which is represented CV-wise. Moreover, dynamic mesh adaptivity allows high mesh resolution to be employed where it is required to resolve the fingers and lower resolution elsewhere. We use our results to re-examine the existing criteria that have been proposed to govern the onset of instability.Mesh adaptivity requires the mapping of data from one mesh to another. Conventional methods such as consistent interpolation do not readily generalise to discontinuous fields and are non-conservative. We further contribute a general framework for interpolation of CV fields by Galerkin projection. The method is conservative, higher order and yields improved results, particularly with higher order or discontinuous elements where existing approaches are often excessively diffusive.
Adaptive mesh strategies for the spectral element method
NASA Technical Reports Server (NTRS)
Mavriplis, Catherine
1992-01-01
An adaptive spectral method was developed for the efficient solution of time dependent partial differential equations. Adaptive mesh strategies that include resolution refinement and coarsening by three different methods are illustrated on solutions to the 1-D viscous Burger equation and the 2-D Navier-Stokes equations for driven flow in a cavity. Sharp gradients, singularities, and regions of poor resolution are resolved optimally as they develop in time using error estimators which indicate the choice of refinement to be used. The adaptive formulation presents significant increases in efficiency, flexibility, and general capabilities for high order spectral methods.
Parallel Block Structured Adaptive Mesh Refinement on Graphics Processing Units
Beckingsale, D. A.; Gaudin, W. P.; Hornung, R. D.; Gunney, B. T.; Gamblin, T.; Herdman, J. A.; Jarvis, S. A.
2014-11-17
Block-structured adaptive mesh refinement is a technique that can be used when solving partial differential equations to reduce the number of zones necessary to achieve the required accuracy in areas of interest. These areas (shock fronts, material interfaces, etc.) are recursively covered with finer mesh patches that are grouped into a hierarchy of refinement levels. Despite the potential for large savings in computational requirements and memory usage without a corresponding reduction in accuracy, AMR adds overhead in managing the mesh hierarchy, adding complex communication and data movement requirements to a simulation. In this paper, we describe the design and implementation of a native GPU-based AMR library, including: the classes used to manage data on a mesh patch, the routines used for transferring data between GPUs on different nodes, and the data-parallel operators developed to coarsen and refine mesh data. We validate the performance and accuracy of our implementation using three test problems and two architectures: an eight-node cluster, and over four thousand nodes of Oak Ridge National Laboratory’s Titan supercomputer. Our GPU-based AMR hydrodynamics code performs up to 4.87× faster than the CPU-based implementation, and has been scaled to over four thousand GPUs using a combination of MPI and CUDA.
Parallel 3D Mortar Element Method for Adaptive Nonconforming Meshes
NASA Technical Reports Server (NTRS)
Feng, Huiyu; Mavriplis, Catherine; VanderWijngaart, Rob; Biswas, Rupak
2004-01-01
High order methods are frequently used in computational simulation for their high accuracy. An efficient way to avoid unnecessary computation in smooth regions of the solution is to use adaptive meshes which employ fine grids only in areas where they are needed. Nonconforming spectral elements allow the grid to be flexibly adjusted to satisfy the computational accuracy requirements. The method is suitable for computational simulations of unsteady problems with very disparate length scales or unsteady moving features, such as heat transfer, fluid dynamics or flame combustion. In this work, we select the Mark Element Method (MEM) to handle the non-conforming interfaces between elements. A new technique is introduced to efficiently implement MEM in 3-D nonconforming meshes. By introducing an "intermediate mortar", the proposed method decomposes the projection between 3-D elements and mortars into two steps. In each step, projection matrices derived in 2-D are used. The two-step method avoids explicitly forming/deriving large projection matrices for 3-D meshes, and also helps to simplify the implementation. This new technique can be used for both h- and p-type adaptation. This method is applied to an unsteady 3-D moving heat source problem. With our new MEM implementation, mesh adaptation is able to efficiently refine the grid near the heat source and coarsen the grid once the heat source passes. The savings in computational work resulting from the dynamic mesh adaptation is demonstrated by the reduction of the the number of elements used and CPU time spent. MEM and mesh adaptation, respectively, bring irregularity and dynamics to the computer memory access pattern. Hence, they provide a good way to gauge the performance of computer systems when running scientific applications whose memory access patterns are irregular and unpredictable. We select a 3-D moving heat source problem as the Unstructured Adaptive (UA) grid benchmark, a new component of the NAS Parallel
Anisotropic norm-oriented mesh adaptation for a Poisson problem
NASA Astrophysics Data System (ADS)
Brèthes, Gautier; Dervieux, Alain
2016-10-01
We present a novel formulation for the mesh adaptation of the approximation of a Partial Differential Equation (PDE). The discussion is restricted to a Poisson problem. The proposed norm-oriented formulation extends the goal-oriented formulation since it is equation-based and uses an adjoint. At the same time, the norm-oriented formulation somewhat supersedes the goal-oriented one since it is basically a solution-convergent method. Indeed, goal-oriented methods rely on the reduction of the error in evaluating a chosen scalar output with the consequence that, as mesh size is increased (more degrees of freedom), only this output is proven to tend to its continuous analog while the solution field itself may not converge. A remarkable quality of goal-oriented metric-based adaptation is the mathematical formulation of the mesh adaptation problem under the form of the optimization, in the well-identified set of metrics, of a well-defined functional. In the new proposed formulation, we amplify this advantage. We search, in the same well-identified set of metrics, the minimum of a norm of the approximation error. The norm is prescribed by the user and the method allows addressing the case of multi-objective adaptation like, for example in aerodynamics, adaptating the mesh for drag, lift and moment in one shot. In this work, we consider the basic linear finite-element approximation and restrict our study to L2 norm in order to enjoy second-order convergence. Numerical examples for the Poisson problem are computed.
AN ADAPTIVE PARTICLE-MESH GRAVITY SOLVER FOR ENZO
Passy, Jean-Claude; Bryan, Greg L.
2014-11-01
We describe and implement an adaptive particle-mesh algorithm to solve the Poisson equation for grid-based hydrodynamics codes with nested grids. The algorithm is implemented and extensively tested within the astrophysical code Enzo against the multigrid solver available by default. We find that while both algorithms show similar accuracy for smooth mass distributions, the adaptive particle-mesh algorithm is more accurate for the case of point masses, and is generally less noisy. We also demonstrate that the two-body problem can be solved accurately in a configuration with nested grids. In addition, we discuss the effect of subcycling, and demonstrate that evolving all the levels with the same timestep yields even greater precision.
Parallel adaptive mesh refinement techniques for plasticity problems
Barry, W.J.; Jones, M.T. |; Plassmann, P.E.
1997-12-31
The accurate modeling of the nonlinear properties of materials can be computationally expensive. Parallel computing offers an attractive way for solving such problems; however, the efficient use of these systems requires the vertical integration of a number of very different software components, we explore the solution of two- and three-dimensional, small-strain plasticity problems. We consider a finite-element formulation of the problem with adaptive refinement of an unstructured mesh to accurately model plastic transition zones. We present a framework for the parallel implementation of such complex algorithms. This framework, using libraries from the SUMAA3d project, allows a user to build a parallel finite-element application without writing any parallel code. To demonstrate the effectiveness of this approach on widely varying parallel architectures, we present experimental results from an IBM SP parallel computer and an ATM-connected network of Sun UltraSparc workstations. The results detail the parallel performance of the computational phases of the application during the process while the material is incrementally loaded.
Parallel adaptive mesh refinement techniques for plasticity problems
NASA Technical Reports Server (NTRS)
Barry, W. J.; Jones, M. T.; Plassmann, P. E.
1997-01-01
The accurate modeling of the nonlinear properties of materials can be computationally expensive. Parallel computing offers an attractive way for solving such problems; however, the efficient use of these systems requires the vertical integration of a number of very different software components, we explore the solution of two- and three-dimensional, small-strain plasticity problems. We consider a finite-element formulation of the problem with adaptive refinement of an unstructured mesh to accurately model plastic transition zones. We present a framework for the parallel implementation of such complex algorithms. This framework, using libraries from the SUMAA3d project, allows a user to build a parallel finite-element application without writing any parallel code. To demonstrate the effectiveness of this approach on widely varying parallel architectures, we present experimental results from an IBM SP parallel computer and an ATM-connected network of Sun UltraSparc workstations. The results detail the parallel performance of the computational phases of the application during the process while the material is incrementally loaded.
Boltzmann Solver with Adaptive Mesh in Velocity Space
Kolobov, Vladimir I.; Arslanbekov, Robert R.; Frolova, Anna A.
2011-05-20
We describe the implementation of direct Boltzmann solver with Adaptive Mesh in Velocity Space (AMVS) using quad/octree data structure. The benefits of the AMVS technique are demonstrated for the charged particle transport in weakly ionized plasmas where the collision integral is linear. We also describe the implementation of AMVS for the nonlinear Boltzmann collision integral. Test computations demonstrate both advantages and deficiencies of the current method for calculations of narrow-kernel distributions.
Advances in Patch-Based Adaptive Mesh Refinement Scalability
Gunney, Brian T.N.; Anderson, Robert W.
2015-12-18
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extensionmore » of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.« less
Advances in Patch-Based Adaptive Mesh Refinement Scalability
Gunney, Brian T.N.; Anderson, Robert W.
2015-12-18
Patch-based structured adaptive mesh refinement (SAMR) is widely used for high-resolution simu- lations. Combined with modern supercomputers, it could provide simulations of unprecedented size and resolution. A persistent challenge for this com- bination has been managing dynamically adaptive meshes on more and more MPI tasks. The dis- tributed mesh management scheme in SAMRAI has made some progress SAMR scalability, but early al- gorithms still had trouble scaling past the regime of 105 MPI tasks. This work provides two critical SAMR regridding algorithms, which are integrated into that scheme to ensure efficiency of the whole. The clustering algorithm is an extension of the tile- clustering approach, making it more flexible and efficient in both clustering and parallelism. The partitioner is a new algorithm designed to prevent the network congestion experienced by its prede- cessor. We evaluated performance using weak- and strong-scaling benchmarks designed to be difficult for dynamic adaptivity. Results show good scaling on up to 1.5M cores and 2M MPI tasks. Detailed timing diagnostics suggest scaling would continue well past that.
Adaptive radial basis function mesh deformation using data reduction
NASA Astrophysics Data System (ADS)
Gillebaart, T.; Blom, D. S.; van Zuijlen, A. H.; Bijl, H.
2016-09-01
Radial Basis Function (RBF) mesh deformation is one of the most robust mesh deformation methods available. Using the greedy (data reduction) method in combination with an explicit boundary correction, results in an efficient method as shown in literature. However, to ensure the method remains robust, two issues are addressed: 1) how to ensure that the set of control points remains an accurate representation of the geometry in time and 2) how to use/automate the explicit boundary correction, while ensuring a high mesh quality. In this paper, we propose an adaptive RBF mesh deformation method, which ensures the set of control points always represents the geometry/displacement up to a certain (user-specified) criteria, by keeping track of the boundary error throughout the simulation and re-selecting when needed. Opposed to the unit displacement and prescribed displacement selection methods, the adaptive method is more robust, user-independent and efficient, for the cases considered. Secondly, the analysis of a single high aspect ratio cell is used to formulate an equation for the correction radius needed, depending on the characteristics of the correction function used, maximum aspect ratio, minimum first cell height and boundary error. Based on the analysis two new radial basis correction functions are derived and proposed. This proposed automated procedure is verified while varying the correction function, Reynolds number (and thus first cell height and aspect ratio) and boundary error. Finally, the parallel efficiency is studied for the two adaptive methods, unit displacement and prescribed displacement for both the CPU as well as the memory formulation with a 2D oscillating and translating airfoil with oscillating flap, a 3D flexible locally deforming tube and deforming wind turbine blade. Generally, the memory formulation requires less work (due to the large amount of work required for evaluating RBF's), but the parallel efficiency reduces due to the limited
Block-structured adaptive mesh refinement - theory, implementation and application
Deiterding, Ralf
2011-01-01
Structured adaptive mesh refinement (SAMR) techniques can enable cutting-edge simulations of problems governed by conservation laws. Focusing on the strictly hyperbolic case, these notes explain all algorithmic and mathematical details of a technically relevant implementation tailored for distributed memory computers. An overview of the background of commonly used finite volume discretizations for gas dynamics is included and typical benchmarks to quantify accuracy and performance of the dynamically adaptive code are discussed. Large-scale simulations of shock-induced realistic combustion in non-Cartesian geometry and shock-driven fluid-structure interaction with fully coupled dynamic boundary motion demonstrate the applicability of the discussed techniques for complex scenarios.
Numerical study of Taylor bubbles with adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Xie, Zhihua; Pavlidis, Dimitrios; Percival, James; Pain, Chris; Matar, Omar; Hasan, Abbas; Azzopardi, Barry
2014-11-01
The Taylor bubble is a single long bubble which nearly fills the entire cross section of a liquid-filled circular tube. This type of bubble flow regime often occurs in gas-liquid slug flows in many industrial applications, including oil-and-gas production, chemical and nuclear reactors, and heat exchangers. The objective of this study is to investigate the fluid dynamics of Taylor bubbles rising in a vertical pipe filled with oils of extremely high viscosity (mimicking the ``heavy oils'' found in the oil-and-gas industry). A modelling and simulation framework is presented here which can modify and adapt anisotropic unstructured meshes to better represent the underlying physics of bubble rise and reduce the computational effort without sacrificing accuracy. The numerical framework consists of a mixed control-volume and finite-element formulation, a ``volume of fluid''-type method for the interface capturing based on a compressive control volume advection method, and a force-balanced algorithm for the surface tension implementation. Numerical examples of some benchmark tests and the dynamics of Taylor bubbles are presented to show the capability of this method. EPSRC Programme Grant, MEMPHIS, EP/K0039761/1.
Multilevel Error Estimation and Adaptive h-Refinement for Cartesian Meshes with Embedded Boundaries
NASA Technical Reports Server (NTRS)
Aftosmis, M. J.; Berger, M. J.; Kwak, Dochan (Technical Monitor)
2002-01-01
This paper presents the development of a mesh adaptation module for a multilevel Cartesian solver. While the module allows mesh refinement to be driven by a variety of different refinement parameters, a central feature in its design is the incorporation of a multilevel error estimator based upon direct estimates of the local truncation error using tau-extrapolation. This error indicator exploits the fact that in regions of uniform Cartesian mesh, the spatial operator is exactly the same on the fine and coarse grids, and local truncation error estimates can be constructed by evaluating the residual on the coarse grid of the restricted solution from the fine grid. A new strategy for adaptive h-refinement is also developed to prevent errors in smooth regions of the flow from being masked by shocks and other discontinuous features. For certain classes of error histograms, this strategy is optimal for achieving equidistribution of the refinement parameters on hierarchical meshes, and therefore ensures grid converged solutions will be achieved for appropriately chosen refinement parameters. The robustness and accuracy of the adaptation module is demonstrated using both simple model problems and complex three dimensional examples using meshes with from 10(exp 6), to 10(exp 7) cells.
NASA Astrophysics Data System (ADS)
Danilov, A. A.; Salamatova, V. Yu; Vassilevski, Yu V.
2012-12-01
Here, a workflow for high-resolution efficient numerical modeling of bioimpedance measurements is suggested that includes 3D image segmentation, adaptive mesh generation, finite-element discretization, and the analysis of simulation results. Using the adaptive unstructured tetrahedral meshes enables to decrease significantly a number of mesh elements while keeping model accuracy. The numerical results illustrate current, potential, and sensitivity field distributions for a conventional Kubicek-like scheme of bioimpedance measurements using segmented geometric model of human torso based on Visible Human Project data. The whole body VHP man computational mesh is constructed that contains 574 thousand vertices and 3.3 million tetrahedrons.
Experiences with an adaptive mesh refinement algorithm in numerical relativity.
NASA Astrophysics Data System (ADS)
Choptuik, M. W.
An implementation of the Berger/Oliger mesh refinement algorithm for a model problem in numerical relativity is described. The principles of operation of the method are reviewed and its use in conjunction with leap-frog schemes is considered. The performance of the algorithm is illustrated with results from a study of the Einstein/massless scalar field equations in spherical symmetry.
An unstructured-mesh atmospheric model for nonhydrostatic dynamics: Towards optimal mesh resolution
NASA Astrophysics Data System (ADS)
Szmelter, Joanna; Zhang, Zhao; Smolarkiewicz, Piotr K.
2015-08-01
The paper advances the limited-area anelastic model (Smolarkiewicz et al. (2013) [45]) for investigation of nonhydrostatic dynamics in mesoscale atmospheric flows. New developments include the extension to a tetrahedral-based median-dual option for unstructured meshes and a static mesh adaptivity technique using an error indicator based on inherent properties of the Multidimensional Positive Definite Advection Transport Algorithm (MPDATA). The model employs semi-implicit nonoscillatory forward-in-time integrators for soundproof PDEs, built on MPDATA and a robust non-symmetric Krylov-subspace elliptic solver. Finite-volume spatial discretisation adopts an edge-based data structure. Simulations of stratified orographic flows and the associated gravity-wave phenomena in media with uniform and variable dispersive properties verify the advancement and demonstrate the potential of heterogeneous anisotropic discretisation with large variation in spatial resolution for study of complex stratified flows that can be computationally unattainable with regular grids.
NASA Astrophysics Data System (ADS)
Yang, Dikun; Oldenburg, Douglas W.; Haber, Eldad
2014-03-01
Airborne electromagnetic (AEM) methods are highly efficient tools for assessing the Earth's conductivity structures in a large area at low cost. However, the configuration of AEM measurements, which typically have widely distributed transmitter-receiver pairs, makes the rigorous modelling and interpretation extremely time-consuming in 3-D. Excessive overcomputing can occur when working on a large mesh covering the entire survey area and inverting all soundings in the data set. We propose two improvements. The first is to use a locally optimized mesh for each AEM sounding for the forward modelling and calculation of sensitivity. This dedicated local mesh is small with fine cells near the sounding location and coarse cells far away in accordance with EM diffusion and the geometric decay of the signals. Once the forward problem is solved on the local meshes, the sensitivity for the inversion on the global mesh is available through quick interpolation. Using local meshes for AEM forward modelling avoids unnecessary computing on fine cells on a global mesh that are far away from the sounding location. Since local meshes are highly independent, the forward modelling can be efficiently parallelized over an array of processors. The second improvement is random and dynamic down-sampling of the soundings. Each inversion iteration only uses a random subset of the soundings, and the subset is reselected for every iteration. The number of soundings in the random subset, determined by an adaptive algorithm, is tied to the degree of model regularization. This minimizes the overcomputing caused by working with redundant soundings. Our methods are compared against conventional methods and tested with a synthetic example. We also invert a field data set that was previously considered to be too large to be practically inverted in 3-D. These examples show that our methodology can dramatically reduce the processing time of 3-D inversion to a practical level without losing resolution
Dynamic Load Balancing for Adaptive Meshes using Symmetric Broadcast Networks
NASA Technical Reports Server (NTRS)
Das, Sajal K.; Harvey, Daniel J.; Biswas, Rupak; Saini, Subhash (Technical Monitor)
1998-01-01
Many scientific applications involve grids that lack a uniform underlying structure. These applications are often dynamic in the sense that the grid structure significantly changes between successive phases of execution. In parallel computing environments, mesh adaptation of grids through selective refinement/coarsening has proven to be an effective approach. However, achieving load balance while minimizing inter-processor communication and redistribution costs is a difficult problem. Traditional dynamic load balancers are mostly inadequate because they lack a global view across processors. In this paper, we compare a novel load balancer that utilizes symmetric broadcast networks (SBN) to a successful global load balancing environment (PLUM) created to handle adaptive unstructured applications. Our experimental results on the IBM SP2 demonstrate that performance of the proposed SBN load balancer is comparable to results achieved under PLUM.
Parallel Processing of Adaptive Meshes with Load Balancing
NASA Technical Reports Server (NTRS)
Das, Sajal K.; Harvey, Daniel J.; Biswas, Rupak; Biegel, Bryan (Technical Monitor)
2001-01-01
Many scientific applications involve grids that lack a uniform underlying structure. These applications are often also dynamic in nature in that the grid structure significantly changes between successive phases of execution. In parallel computing environments, mesh adaptation of unstructured grids through selective refinement/coarsening has proven to be an effective approach. However, achieving load balance while minimizing interprocessor communication and redistribution costs is a difficult problem. Traditional dynamic load balancers are mostly inadequate because they lack a global view of system loads across processors. In this paper, we propose a novel and general-purpose load balancer that utilizes symmetric broadcast networks (SBN) as the underlying communication topology, and compare its performance with a successful global load balancing environment, called PLUM, specifically created to handle adaptive unstructured applications. Our experimental results on an IBM SP2 demonstrate that the SBN-based load balancer achieves lower redistribution costs than that under PLUM by overlapping processing and data migration.
Spatially adaptive bases in wavelet-based coding of semi-regular meshes
NASA Astrophysics Data System (ADS)
Denis, Leon; Florea, Ruxandra; Munteanu, Adrian; Schelkens, Peter
2010-05-01
In this paper we present a wavelet-based coding approach for semi-regular meshes, which spatially adapts the employed wavelet basis in the wavelet transformation of the mesh. The spatially-adaptive nature of the transform requires additional information to be stored in the bit-stream in order to allow the reconstruction of the transformed mesh at the decoder side. In order to limit this overhead, the mesh is first segmented into regions of approximately equal size. For each spatial region, a predictor is selected in a rate-distortion optimal manner by using a Lagrangian rate-distortion optimization technique. When compared against the classical wavelet transform employing the butterfly subdivision filter, experiments reveal that the proposed spatially-adaptive wavelet transform significantly decreases the energy of the wavelet coefficients for all subbands. Preliminary results show also that employing the proposed transform for the lowest-resolution subband systematically yields improved compression performance at low-to-medium bit-rates. For the Venus and Rabbit test models the compression improvements add up to 1.47 dB and 0.95 dB, respectively.
Mesh adaption for efficient multiscale implementation of one-dimensional turbulence
NASA Astrophysics Data System (ADS)
Lignell, D. O.; Kerstein, A. R.; Sun, G.; Monson, E. I.
2013-06-01
One-Dimensional Turbulence (ODT) is a stochastic model for turbulent flow simulation. In an atmospheric context, it is analogous to single-column modeling (SCM) in that it lives on a 1D spatial domain, but different in that it time advances individual flow realizations rather than ensemble-averaged quantities. The lack of averaging enables a physically sound multiscale treatment, which is useful for resolving sporadic localized phenomena, as seen in stably stratified regimes, and sharp interfaces, as observed where a convective layer encounters a stable overlying zone. In such flows, the relevant scale range is so large that it is beneficial to enhance model performance by introducing an adaptive mesh. An adaptive-mesh algorithm that provides the desired performance characteristics is described and demonstrated, and its implications for the ODT advancement scheme are explained.
Visualization of Octree Adaptive Mesh Refinement (AMR) in Astrophysical Simulations
NASA Astrophysics Data System (ADS)
Labadens, M.; Chapon, D.; Pomaréde, D.; Teyssier, R.
2012-09-01
Computer simulations are important in current cosmological research. Those simulations run in parallel on thousands of processors, and produce huge amount of data. Adaptive mesh refinement is used to reduce the computing cost while keeping good numerical accuracy in regions of interest. RAMSES is a cosmological code developed by the Commissariat à l'énergie atomique et aux énergies alternatives (English: Atomic Energy and Alternative Energies Commission) which uses Octree adaptive mesh refinement. Compared to grid based AMR, the Octree AMR has the advantage to fit very precisely the adaptive resolution of the grid to the local problem complexity. However, this specific octree data type need some specific software to be visualized, as generic visualization tools works on Cartesian grid data type. This is why the PYMSES software has been also developed by our team. It relies on the python scripting language to ensure a modular and easy access to explore those specific data. In order to take advantage of the High Performance Computer which runs the RAMSES simulation, it also uses MPI and multiprocessing to run some parallel code. We would like to present with more details our PYMSES software with some performance benchmarks. PYMSES has currently two visualization techniques which work directly on the AMR. The first one is a splatting technique, and the second one is a custom ray tracing technique. Both have their own advantages and drawbacks. We have also compared two parallel programming techniques with the python multiprocessing library versus the use of MPI run. The load balancing strategy has to be smartly defined in order to achieve a good speed up in our computation. Results obtained with this software are illustrated in the context of a massive, 9000-processor parallel simulation of a Milky Way-like galaxy.
Ray, Jaideep; Lefantzi, Sophia; Najm, Habib N.; Kennedy, Christopher A.
2006-01-01
Block-structured adaptively refined meshes (SAMR) strive for efficient resolution of partial differential equations (PDEs) solved on large computational domains by clustering mesh points only where required by large gradients. Previous work has indicated that fourth-order convergence can be achieved on such meshes by using a suitable combination of high-order discretizations, interpolations, and filters and can deliver significant computational savings over conventional second-order methods at engineering error tolerances. In this paper, we explore the interactions between the errors introduced by discretizations, interpolations and filters. We develop general expressions for high-order discretizations, interpolations, and filters, in multiple dimensions, using a Fourier approach, facilitating the high-order SAMR implementation. We derive a formulation for the necessary interpolation order for given discretization and derivative orders. We also illustrate this order relationship empirically using one and two-dimensional model problems on refined meshes. We study the observed increase in accuracy with increasing interpolation order. We also examine the empirically observed order of convergence, as the effective resolution of the mesh is increased by successively adding levels of refinement, with different orders of discretization, interpolation, or filtering.
TRIM: A finite-volume MHD algorithm for an unstructured adaptive mesh
Schnack, D.D.; Lottati, I.; Mikic, Z.
1995-07-01
The authors describe TRIM, a MHD code which uses finite volume discretization of the MHD equations on an unstructured adaptive grid of triangles in the poloidal plane. They apply it to problems related to modeling tokamak toroidal plasmas. The toroidal direction is treated by a pseudospectral method. Care was taken to center variables appropriately on the mesh and to construct a self adjoint diffusion operator for cell centered variables.
Production-quality Tools for Adaptive Mesh RefinementVisualization
Weber, Gunther H.; Childs, Hank; Bonnell, Kathleen; Meredith,Jeremy; Miller, Mark; Whitlock, Brad; Bethel, E. Wes
2007-10-25
Adaptive Mesh Refinement (AMR) is a highly effectivesimulation method for spanning a large range of spatiotemporal scales,such as astrophysical simulations that must accommodate ranges frominterstellar to sub-planetary. Most mainstream visualization tools stilllack support for AMR as a first class data type and AMR code teams usecustom built applications for AMR visualization. The Department ofEnergy's (DOE's) Science Discovery through Advanced Computing (SciDAC)Visualization and Analytics Center for Enabling Technologies (VACET) isextending and deploying VisIt, an open source visualization tool thataccommodates AMR as a first-class data type, for use asproduction-quality, parallel-capable AMR visual data analysisinfrastructure. This effort will help science teams that use AMR-basedsimulations and who develop their own AMR visual data analysis softwareto realize cost and labor savings.
NASA Astrophysics Data System (ADS)
Combet, F.; Gelman, L.
2011-04-01
In this paper, a novel adaptive demodulation technique including a new diagnostic feature is proposed for gear diagnosis in conditions of variable amplitudes of the mesh harmonics. This vibration technique employs the time synchronous average (TSA) of vibration signals. The new adaptive diagnostic feature is defined as the ratio of the sum of the sideband components of the envelope spectrum of a mesh harmonic to the measured power of the mesh harmonic. The proposed adaptation of the technique is justified theoretically and experimentally by the high level of the positive covariance between amplitudes of the mesh harmonics and the sidebands in conditions of variable amplitudes of the mesh harmonics. It is shown that the adaptive demodulation technique preserves effectiveness of local fault detection of gears operating in conditions of variable mesh amplitudes.
Unstructured and adaptive mesh generation for high Reynolds number viscous flows
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1991-01-01
A method for generating and adaptively refining a highly stretched unstructured mesh suitable for the computation of high-Reynolds-number viscous flows about arbitrary two-dimensional geometries was developed. The method is based on the Delaunay triangulation of a predetermined set of points and employs a local mapping in order to achieve the high stretching rates required in the boundary-layer and wake regions. The initial mesh-point distribution is determined in a geometry-adaptive manner which clusters points in regions of high curvature and sharp corners. Adaptive mesh refinement is achieved by adding new points in regions of large flow gradients, and locally retriangulating; thus, obviating the need for global mesh regeneration. Initial and adapted meshes about complex multi-element airfoil geometries are shown and compressible flow solutions are computed on these meshes.
NASA Astrophysics Data System (ADS)
Bespalov, Alex; Haberl, Alexander; Praetorius, Dirk
2017-04-01
We prove that for compactly perturbed elliptic problems, where the corresponding bilinear form satisfies a Garding inequality, adaptive mesh-refinement is capable of overcoming the preasymptotic behavior and eventually leads to convergence with optimal algebraic rates. As an important consequence of our analysis, one does not have to deal with the a-priori assumption that the underlying meshes are sufficiently fine. Hence, the overall conclusion of our results is that adaptivity has stabilizing effects and can overcome possibly pessimistic restrictions on the meshes. In particular, our analysis covers adaptive mesh-refinement for the finite element discretization of the Helmholtz equation from where our interest originated.
Mesh Convergence Requirements for Composite Damage Models
NASA Technical Reports Server (NTRS)
Davila, Carlos G.
2016-01-01
The ability of the finite element method to accurately represent the response of objects with intricate geometry and loading renders the finite element method as an extremely versatile analysis technique for structural analysis. Finite element analysis is routinely used in industry to calculate deflections, stress concentrations, natural frequencies, buckling loads, and much more. The method works by discretizing complex problems into smaller, simpler approximations that are valid over small uniform domains. For common analyses, the maximum size of the elements that can be used is often be determined by experience. However, to verify the quality of a solution, analyses with several levels of mesh refinement should be performed to ensure that the solution has converged. In recent years, the finite element method has been used to calculate the resistance of structures, and in particular that of composite structures. A number of techniques such as cohesive zone modeling, the virtual crack closure technique, and continuum damage modeling have emerged that can be used to predict cracking, delaminations, fiber failure, and other composite damage modes that lead to structural collapse. However, damage models present mesh refinement requirements that are not well understood. In this presentation, we examine different mesh refinement issues related to the representation of damage in composite materials. Damage process zone sizes and their corresponding mesh requirements will be discussed. The difficulties of modeling discontinuities and the associated need for regularization techniques will be illustrated, and some unexpected element size constraints will be presented. Finally, some of the difficulties in constructing models of composite structures capable of predicting transverse matrix cracking will be discussed. It will be shown that to predict the initiation and propagation of transverse matrix cracks, their density, and their saturation may require models that are
Lung lobe modeling and segmentation with individualized surface meshes
NASA Astrophysics Data System (ADS)
Blaffert, Thomas; Barschdorf, Hans; von Berg, Jens; Dries, Sebastian; Franz, Astrid; Klinder, Tobias; Lorenz, Cristian; Renisch, Steffen; Wiemker, Rafael
2008-03-01
An automated segmentation of lung lobes in thoracic CT images is of interest for various diagnostic purposes like the quantification of emphysema or the localization of tumors within the lung. Although the separating lung fissures are visible in modern multi-slice CT-scanners, their contrast in the CT-image often does not separate the lobes completely. This makes it impossible to build a reliable segmentation algorithm without additional information. Our approach uses general anatomical knowledge represented in a geometrical mesh model to construct a robust lobe segmentation, which even gives reasonable estimates of lobe volumes if fissures are not visible at all. The paper describes the generation of the lung model mesh including lobes by an average volume model, its adaptation to individual patient data using a special fissure feature image, and a performance evaluation over a test data set showing an average segmentation accuracy of 1 to 3 mm.
Fluidity: a fully-unstructured adaptive mesh computational framework for geodynamics
NASA Astrophysics Data System (ADS)
Kramer, S. C.; Davies, D.; Wilson, C. R.
2010-12-01
Fluidity is a finite element, finite volume fluid dynamics model developed by the Applied Modelling and Computation Group at Imperial College London. Several features of the model make it attractive for use in geodynamics. A core finite element library enables the rapid implementation and investigation of new numerical schemes. For example, the function spaces used for each variable can be changed allowing properties of the discretisation, such as stability, conservation and balance, to be easily varied and investigated. Furthermore, unstructured, simplex meshes allow the underlying resolution to vary rapidly across the computational domain. Combined with dynamic mesh adaptivity, where the mesh is periodically optimised to the current conditions, this allows significant savings in computational cost over traditional chessboard-like structured mesh simulations [1]. In this study we extend Fluidity (using the Portable, Extensible Toolkit for Scientific Computation [PETSc, 2]) to Stokes flow problems relevant to geodynamics. However, due to the assumptions inherent in all models, it is necessary to properly verify and validate the code before applying it to any large-scale problems. In recent years this has been made easier by the publication of a series of ‘community benchmarks’ for geodynamic modelling. We discuss the use of several of these to help validate Fluidity [e.g. 3, 4]. The experimental results of Vatteville et al. [5] are then used to validate Fluidity against laboratory measurements. This test case is also used to highlight the computational advantages of using adaptive, unstructured meshes - significantly reducing the number of nodes and total CPU time required to match a fixed mesh simulation. References: 1. C. C. Pain et al. Comput. Meth. Appl. M, 190:3771-3796, 2001. doi:10.1016/S0045-7825(00)00294-2. 2. B. Satish et al. http://www.mcs.anl.gov/petsc/petsc-2/, 2001. 3. Blankenbach et al. Geophys. J. Int., 98:23-28, 1989. 4. Busse et al. Geophys
Xia, Kelin; Zhan, Meng; Wan, Decheng; Wei, Guo-Wei
2011-01-01
Mesh deformation methods are a versatile strategy for solving partial differential equations (PDEs) with a vast variety of practical applications. However, these methods break down for elliptic PDEs with discontinuous coefficients, namely, elliptic interface problems. For this class of problems, the additional interface jump conditions are required to maintain the well-posedness of the governing equation. Consequently, in order to achieve high accuracy and high order convergence, additional numerical algorithms are required to enforce the interface jump conditions in solving elliptic interface problems. The present work introduces an interface technique based adaptively deformed mesh strategy for resolving elliptic interface problems. We take the advantages of the high accuracy, flexibility and robustness of the matched interface and boundary (MIB) method to construct an adaptively deformed mesh based interface method for elliptic equations with discontinuous coefficients. The proposed method generates deformed meshes in the physical domain and solves the transformed governed equations in the computational domain, which maintains regular Cartesian meshes. The mesh deformation is realized by a mesh transformation PDE, which controls the mesh redistribution by a source term. The source term consists of a monitor function, which builds in mesh contraction rules. Both interface geometry based deformed meshes and solution gradient based deformed meshes are constructed to reduce the L∞ and L2 errors in solving elliptic interface problems. The proposed adaptively deformed mesh based interface method is extensively validated by many numerical experiments. Numerical results indicate that the adaptively deformed mesh based interface method outperforms the original MIB method for dealing with elliptic interface problems. PMID:22586356
An object-oriented approach for parallel self adaptive mesh refinement on block structured grids
NASA Technical Reports Server (NTRS)
Lemke, Max; Witsch, Kristian; Quinlan, Daniel
1993-01-01
Self-adaptive mesh refinement dynamically matches the computational demands of a solver for partial differential equations to the activity in the application's domain. In this paper we present two C++ class libraries, P++ and AMR++, which significantly simplify the development of sophisticated adaptive mesh refinement codes on (massively) parallel distributed memory architectures. The development is based on our previous research in this area. The C++ class libraries provide abstractions to separate the issues of developing parallel adaptive mesh refinement applications into those of parallelism, abstracted by P++, and adaptive mesh refinement, abstracted by AMR++. P++ is a parallel array class library to permit efficient development of architecture independent codes for structured grid applications, and AMR++ provides support for self-adaptive mesh refinement on block-structured grids of rectangular non-overlapping blocks. Using these libraries, the application programmers' work is greatly simplified to primarily specifying the serial single grid application and obtaining the parallel and self-adaptive mesh refinement code with minimal effort. Initial results for simple singular perturbation problems solved by self-adaptive multilevel techniques (FAC, AFAC), being implemented on the basis of prototypes of the P++/AMR++ environment, are presented. Singular perturbation problems frequently arise in large applications, e.g. in the area of computational fluid dynamics. They usually have solutions with layers which require adaptive mesh refinement and fast basic solvers in order to be resolved efficiently.
An adaptive mesh magneto-hydrodynamic analysis of interstellar clouds
NASA Astrophysics Data System (ADS)
Kominsky, Paul J.
Interstellar clouds play a key role in many astrophysical events. The interactions of dense interstellar clouds with shock waves and interstellar wind were investigated using an adaptive three-dimensional Cartesian mesh approach to the magneto-hydrodynamic equations. The mixing of the cloud material with the post-shock material results in complex layers of current density. In both the shock and wind interactions, a tail develops similar to the tail found with comets due to the solar wind. The orientation of this tail structure changes with the direction of the magnetic field, and may be useful to observationally determining the orientation of magnetic fields in the interstellar medium. The octree data structure was analyzed in regard to parallel work units. Larger block sizes have a higher volume to surface ratio and support a higher percentage of computational cells to non-computational cells, but require more cells at the finest grid resolution. Keeping the minimum resolution of the grid fixed, and averaging over all possible grids, the analysis confirms experience that block sizes larger than 8 × 8 × 8 cells do not improve storage efficiency. A novel algorithm was developed to implement rotationally periodic boundary conditions on quadtree and octree data, structures. Astrophysical flows wit h symmetric circulation, such as accretion disks, or periodic instabilities, such supernova remnants, may be able to take advantage of such boundary conditions while maintaining the other benefits of a Cartesian grid.
CONSTRAINED-TRANSPORT MAGNETOHYDRODYNAMICS WITH ADAPTIVE MESH REFINEMENT IN CHARM
Miniati, Francesco; Martin, Daniel F. E-mail: DFMartin@lbl.gov
2011-07-01
We present the implementation of a three-dimensional, second-order accurate Godunov-type algorithm for magnetohydrodynamics (MHD) in the adaptive-mesh-refinement (AMR) cosmological code CHARM. The algorithm is based on the full 12-solve spatially unsplit corner-transport-upwind (CTU) scheme. The fluid quantities are cell-centered and are updated using the piecewise-parabolic method (PPM), while the magnetic field variables are face-centered and are evolved through application of the Stokes theorem on cell edges via a constrained-transport (CT) method. The so-called multidimensional MHD source terms required in the predictor step for high-order accuracy are applied in a simplified form which reduces their complexity in three dimensions without loss of accuracy or robustness. The algorithm is implemented on an AMR framework which requires specific synchronization steps across refinement levels. These include face-centered restriction and prolongation operations and a reflux-curl operation, which maintains a solenoidal magnetic field across refinement boundaries. The code is tested against a large suite of test problems, including convergence tests in smooth flows, shock-tube tests, classical two- and three-dimensional MHD tests, a three-dimensional shock-cloud interaction problem, and the formation of a cluster of galaxies in a fully cosmological context. The magnetic field divergence is shown to remain negligible throughout.
Adaptive Mesh Refinement for Hyperbolic Partial Differential Equations
1983-03-01
grids. We use either the Coarse Mesh Approximation fethod ( Ciment , [1971]) or interpolation from a coarser grid to get the boundary values. In Berger...Problems, Math. Conp. 31 (1977), 333-390. M. Ciment , Stable Difference Schemes with Uneven Mesh Spacings, Math. Comp. 25 (1971), 219-227. H. Cramr
A Mass Conservation Algorithm for Adaptive Unrefinement Meshes Used by Finite Element Methods
2012-01-01
dimensional mesh generation. In: Proc. 4th ACM-SIAM Symp. on Disc. Algorithms. (1993) 83–92 [9] Weatherill, N., Hassan, O., Marcum, D., Marchant, M.: Grid ...Conference on Computational Science, ICCS 2012 A Mass Conservation Algorithm For Adaptive Unrefinement Meshes Used By Finite Element Methods Hung V. Nguyen...velocity fields, and chemical distribution, as well as conserve mass, especially for water quality applications. Solution accuracy depends highly on mesh
Global Load Balancing with Parallel Mesh Adaption on Distributed-Memory Systems
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Oliker, Leonid; Sohn, Andrew
1996-01-01
Dynamic mesh adaption on unstructured grids is a powerful tool for efficiently computing unsteady problems to resolve solution features of interest. Unfortunately, this causes load imbalance among processors on a parallel machine. This paper describes the parallel implementation of a tetrahedral mesh adaption scheme and a new global load balancing method. A heuristic remapping algorithm is presented that assigns partitions to processors such that the redistribution cost is minimized. Results indicate that the parallel performance of the mesh adaption code depends on the nature of the adaption region and show a 35.5X speedup on 64 processors of an SP2 when 35% of the mesh is randomly adapted. For large-scale scientific computations, our load balancing strategy gives almost a sixfold reduction in solver execution times over non-balanced loads. Furthermore, our heuristic remapper yields processor assignments that are less than 3% off the optimal solutions but requires only 1% of the computational time.
Global Load Balancing with Parallel Mesh Adaption on Distributed-Memory Systems
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Oliker, Leonid; Sohn, Andrew
1996-01-01
Dynamic mesh adaptation on unstructured grids is a powerful tool for efficiently computing unsteady problems to resolve solution features of interest. Unfortunately, this causes load inbalances among processors on a parallel machine. This paper described the parallel implementation of a tetrahedral mesh adaption scheme and a new global load balancing method. A heuristic remapping algorithm is presented that assigns partitions to processors such that the redistribution coast is minimized. Results indicate that the parallel performance of the mesh adaption code depends on the nature of the adaption region and show a 35.5X speedup on 64 processors of an SP2 when 35 percent of the mesh is randomly adapted. For large scale scientific computations, our load balancing strategy gives an almost sixfold reduction in solver execution times over non-balanced loads. Furthermore, our heuristic remappier yields processor assignments that are less than 3 percent of the optimal solutions, but requires only 1 percent of the computational time.
A Robust and Scalable Software Library for Parallel Adaptive Refinement on Unstructured Meshes
NASA Technical Reports Server (NTRS)
Lou, John Z.; Norton, Charles D.; Cwik, Thomas A.
1999-01-01
The design and implementation of Pyramid, a software library for performing parallel adaptive mesh refinement (PAMR) on unstructured meshes, is described. This software library can be easily used in a variety of unstructured parallel computational applications, including parallel finite element, parallel finite volume, and parallel visualization applications using triangular or tetrahedral meshes. The library contains a suite of well-designed and efficiently implemented modules that perform operations in a typical PAMR process. Among these are mesh quality control during successive parallel adaptive refinement (typically guided by a local-error estimator), parallel load-balancing, and parallel mesh partitioning using the ParMeTiS partitioner. The Pyramid library is implemented in Fortran 90 with an interface to the Message-Passing Interface (MPI) library, supporting code efficiency, modularity, and portability. An EM waveguide filter application, adaptively refined using the Pyramid library, is illustrated.
Adaptive Meshing of Ship Air-Wake Flowfields
2014-03-03
this code are currently generated using Pointwise .[2] This code also uses a second order spatial finite-volume scheme with first order explicit...simulated with the two codes and is shown below. The surface mesh from the 3D mesh generated by Pointwise serves as the geometry for the OctFlow code. A...Geometries", AIAA- 2000-1006,2000. 2. " Pointwise ." Pointwise , Inc., http://www.pointwise.com. 3. O’Connell, M., and Karman, S., "Mesh Rupturing: A
Using Adaptive Mesh Refinment to Simulate Storm Surge
NASA Astrophysics Data System (ADS)
Mandli, K. T.; Dawson, C.
2012-12-01
Coastal hazards related to strong storms such as hurricanes and typhoons are one of the most frequently recurring and wide spread hazards to coastal communities. Storm surges are among the most devastating effects of these storms, and their prediction and mitigation through numerical simulations is of great interest to coastal communities that need to plan for the subsequent rise in sea level during these storms. Unfortunately these simulations require a large amount of resolution in regions of interest to capture relevant effects resulting in a computational cost that may be intractable. This problem is exacerbated in situations where a large number of similar runs is needed such as in design of infrastructure or forecasting with ensembles of probable storms. One solution to address the problem of computational cost is to employ adaptive mesh refinement (AMR) algorithms. AMR functions by decomposing the computational domain into regions which may vary in resolution as time proceeds. Decomposing the domain as the flow evolves makes this class of methods effective at ensuring that computational effort is spent only where it is needed. AMR also allows for placement of computational resolution independent of user interaction and expectation of the dynamics of the flow as well as particular regions of interest such as harbors. The simulation of many different applications have only been made possible by using AMR-type algorithms, which have allowed otherwise impractical simulations to be performed for much less computational expense. Our work involves studying how storm surge simulations can be improved with AMR algorithms. We have implemented relevant storm surge physics in the GeoClaw package and tested how Hurricane Ike's surge into Galveston Bay and up the Houston Ship Channel compares to available tide gauge data. We will also discuss issues dealing with refinement criteria, optimal resolution and refinement ratios, and inundation.
RAM: a Relativistic Adaptive Mesh Refinement Hydrodynamics Code
Zhang, Wei-Qun; MacFadyen, Andrew I.; /Princeton, Inst. Advanced Study
2005-06-06
The authors have developed a new computer code, RAM, to solve the conservative equations of special relativistic hydrodynamics (SRHD) using adaptive mesh refinement (AMR) on parallel computers. They have implemented a characteristic-wise, finite difference, weighted essentially non-oscillatory (WENO) scheme using the full characteristic decomposition of the SRHD equations to achieve fifth-order accuracy in space. For time integration they use the method of lines with a third-order total variation diminishing (TVD) Runge-Kutta scheme. They have also implemented fourth and fifth order Runge-Kutta time integration schemes for comparison. The implementation of AMR and parallelization is based on the FLASH code. RAM is modular and includes the capability to easily swap hydrodynamics solvers, reconstruction methods and physics modules. In addition to WENO they have implemented a finite volume module with the piecewise parabolic method (PPM) for reconstruction and the modified Marquina approximate Riemann solver to work with TVD Runge-Kutta time integration. They examine the difficulty of accurately simulating shear flows in numerical relativistic hydrodynamics codes. They show that under-resolved simulations of simple test problems with transverse velocity components produce incorrect results and demonstrate the ability of RAM to correctly solve these problems. RAM has been tested in one, two and three dimensions and in Cartesian, cylindrical and spherical coordinates. they have demonstrated fifth-order accuracy for WENO in one and two dimensions and performed detailed comparison with other schemes for which they show significantly lower convergence rates. Extensive testing is presented demonstrating the ability of RAM to address challenging open questions in relativistic astrophysics.
NASA Technical Reports Server (NTRS)
Ashford, Gregory A.; Powell, Kenneth G.
1995-01-01
A method for generating high quality unstructured triangular grids for high Reynolds number Navier-Stokes calculations about complex geometries is described. Careful attention is paid in the mesh generation process to resolving efficiently the disparate length scales which arise in these flows. First the surface mesh is constructed in a way which ensures that the geometry is faithfully represented. The volume mesh generation then proceeds in two phases thus allowing the viscous and inviscid regions of the flow to be meshed optimally. A solution-adaptive remeshing procedure which allows the mesh to adapt itself to flow features is also described. The procedure for tracking wakes and refinement criteria appropriate for shock detection are described. Although at present it has only been implemented in two dimensions, the grid generation process has been designed with the extension to three dimensions in mind. An implicit, higher-order, upwind method is also presented for computing compressible turbulent flows on these meshes. Two recently developed one-equation turbulence models have been implemented to simulate the effects of the fluid turbulence. Results for flow about a RAE 2822 airfoil and a Douglas three-element airfoil are presented which clearly show the improved resolution obtainable.
Adaptive Meshing of Ship Air-Wake Flowfields
2014-10-21
resolve gradients of the adaptation function. The third method is a meshless method that uses a physics-based force model to move nodes around to...method that uses a physics-based force model to move nodes around to resolve the geometry and flowfield. The initial phase of the research conducted...three codes all solve the unsteady Euler equations, but use different discretization strategies. The target application is an aircraft in a landing
NASA Technical Reports Server (NTRS)
Barnard, Stephen T.; Simon, Horst; Lasinski, T. A. (Technical Monitor)
1994-01-01
The design of a parallel implementation of multilevel recursive spectral bisection is described. The goal is to implement a code that is fast enough to enable dynamic repartitioning of adaptive meshes.
Greene, Patrick T.; Schofield, Samuel P.; Nourgaliev, Robert
2016-06-21
A new mesh smoothing method designed to cluster mesh cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function being computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered elds, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well for the weight function as the actual level set. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Dynamic cases for moving interfaces are presented to demonstrate the method's potential usefulness to arbitrary Lagrangian Eulerian (ALE) methods.
Improved Simulation of Electrodiffusion in the Node of Ranvier by Mesh Adaptation.
Dione, Ibrahima; Deteix, Jean; Briffard, Thomas; Chamberland, Eric; Doyon, Nicolas
2016-01-01
In neural structures with complex geometries, numerical resolution of the Poisson-Nernst-Planck (PNP) equations is necessary to accurately model electrodiffusion. This formalism allows one to describe ionic concentrations and the electric field (even away from the membrane) with arbitrary spatial and temporal resolution which is impossible to achieve with models relying on cable theory. However, solving the PNP equations on complex geometries involves handling intricate numerical difficulties related either to the spatial discretization, temporal discretization or the resolution of the linearized systems, often requiring large computational resources which have limited the use of this approach. In the present paper, we investigate the best ways to use the finite elements method (FEM) to solve the PNP equations on domains with discontinuous properties (such as occur at the membrane-cytoplasm interface). 1) Using a simple 2D geometry to allow comparison with analytical solution, we show that mesh adaptation is a very (if not the most) efficient way to obtain accurate solutions while limiting the computational efforts, 2) We use mesh adaptation in a 3D model of a node of Ranvier to reveal details of the solution which are nearly impossible to resolve with other modelling techniques. For instance, we exhibit a non linear distribution of the electric potential within the membrane due to the non uniform width of the myelin and investigate its impact on the spatial profile of the electric field in the Debye layer.
Improved Simulation of Electrodiffusion in the Node of Ranvier by Mesh Adaptation
Dione, Ibrahima; Briffard, Thomas; Doyon, Nicolas
2016-01-01
In neural structures with complex geometries, numerical resolution of the Poisson-Nernst-Planck (PNP) equations is necessary to accurately model electrodiffusion. This formalism allows one to describe ionic concentrations and the electric field (even away from the membrane) with arbitrary spatial and temporal resolution which is impossible to achieve with models relying on cable theory. However, solving the PNP equations on complex geometries involves handling intricate numerical difficulties related either to the spatial discretization, temporal discretization or the resolution of the linearized systems, often requiring large computational resources which have limited the use of this approach. In the present paper, we investigate the best ways to use the finite elements method (FEM) to solve the PNP equations on domains with discontinuous properties (such as occur at the membrane-cytoplasm interface). 1) Using a simple 2D geometry to allow comparison with analytical solution, we show that mesh adaptation is a very (if not the most) efficient way to obtain accurate solutions while limiting the computational efforts, 2) We use mesh adaptation in a 3D model of a node of Ranvier to reveal details of the solution which are nearly impossible to resolve with other modelling techniques. For instance, we exhibit a non linear distribution of the electric potential within the membrane due to the non uniform width of the myelin and investigate its impact on the spatial profile of the electric field in the Debye layer. PMID:27548674
Composite-Grid Techniques and Adaptive Mesh Refinement in Computational Fluid Dynamics
1990-01-01
the equations govern- ing the flow. The patched adaptive mesh refinement technique, devised at Stanford by Oliger, et al ., copes with these sources of...patched adaptive mesh refinement technique, devised at Stanford by Oliger et al . [OL184], copes with these sources of error efficiently by refining...differential equation, as in the numerical grid generation methods proposed by Thompson et al . [THO85], or simply a list of pairs of points in
Kohn, S.; Weare, J.; Ong, E.; Baden, S.
1997-05-01
We have applied structured adaptive mesh refinement techniques to the solution of the LDA equations for electronic structure calculations. Local spatial refinement concentrates memory resources and numerical effort where it is most needed, near the atomic centers and in regions of rapidly varying charge density. The structured grid representation enables us to employ efficient iterative solver techniques such as conjugate gradient with FAC multigrid preconditioning. We have parallelized our solver using an object- oriented adaptive mesh refinement framework.
A tetrahedral mesh generation approach for 3D marine controlled-source electromagnetic modeling
NASA Astrophysics Data System (ADS)
Um, Evan Schankee; Kim, Seung-Sep; Fu, Haohuan
2017-03-01
3D finite-element (FE) mesh generation is a major hurdle for marine controlled-source electromagnetic (CSEM) modeling. In this paper, we present a FE discretization operator (FEDO) that automatically converts a 3D finite-difference (FD) model into reliable and efficient tetrahedral FE meshes for CSEM modeling. FEDO sets up wireframes of a background seabed model that precisely honors the seafloor topography. The wireframes are then partitioned into multiple regions. Outer regions of the wireframes are discretized with coarse tetrahedral elements whose maximum size is as large as a skin depth of the regions. We demonstrate that such coarse meshes can produce accurate FE solutions because numerical dispersion errors of tetrahedral meshes do not accumulate but oscillates. In contrast, central regions of the wireframes are discretized with fine tetrahedral elements to describe complex geology in detail. The conductivity distribution is mapped from FD to FE meshes in a volume-averaged sense. To avoid excessive mesh refinement around receivers, we introduce an effective receiver size. Major advantages of FEDO are summarized as follow. First, FEDO automatically generates reliable and economic tetrahedral FE meshes without adaptive meshing or interactive CAD workflows. Second, FEDO produces FE meshes that precisely honor the boundaries of the seafloor topography. Third, FEDO derives multiple sets of FE meshes from a given FD model. Each FE mesh is optimized for a different set of sources and receivers and is fed to a subgroup of processors on a parallel computer. This divide and conquer approach improves the parallel scalability of the FE solution. Both accuracy and effectiveness of FEDO are demonstrated with various CSEM examples.
A User's Guide to AMR1D: An Instructional Adaptive Mesh Refinement Code for Unstructured Grids
NASA Technical Reports Server (NTRS)
deFainchtein, Rosalinda
1996-01-01
This report documents the code AMR1D, which is currently posted on the World Wide Web (http://sdcd.gsfc.nasa.gov/ESS/exchange/contrib/de-fainchtein/adaptive _mesh_refinement.html). AMR1D is a one-dimensional finite element fluid-dynamics solver, capable of adaptive mesh refinement (AMR). It was written as an instructional tool for AMR on unstructured mesh codes. It is meant to illustrate the minimum requirements for AMR on more than one dimension. For that purpose, it uses the same type of data structure that would be necessary on a two-dimensional AMR code (loosely following the algorithm described by Lohner).
2013-01-01
Analysis of Adaptive Mesh Refinement for IMEX Discontinuous Galerkin Solutions of the Compressible Euler Equations with Application to Atmospheric...order discontinuous Galerkin method on quadrilateral grids with non-conforming elements. We perform a detailed analysis of the cost of AMR by comparing...adaptive mesh refinement, discontinuous Galerkin method, non-conforming mesh, IMEX, compressible Euler equations, atmospheric simulations 1. Introduction
Zonal multigrid solution of compressible flow problems on unstructured and adaptive meshes
NASA Technical Reports Server (NTRS)
Mavriplis, Dimitri J.
1989-01-01
The simultaneous use of adaptive meshing techniques with a multigrid strategy for solving the 2-D Euler equations in the context of unstructured meshes is studied. To obtain optimal efficiency, methods capable of computing locally improved solutions without recourse to global recalculations are pursued. A method for locally refining an existing unstructured mesh, without regenerating a new global mesh is employed, and the domain is automatically partitioned into refined and unrefined regions. Two multigrid strategies are developed. In the first, time-stepping is performed on a global fine mesh covering the entire domain, and convergence acceleration is achieved through the use of zonal coarse grid accelerator meshes, which lie under the adaptively refined regions of the global fine mesh. Both schemes are shown to produce similar convergence rates to each other, and also with respect to a previously developed global multigrid algorithm, which performs time-stepping throughout the entire domain, on each mesh level. However, the present schemes exhibit higher computational efficiency due to the smaller number of operations on each level.
Wang Yaqi; Ragusa, Jean C.
2011-02-01
Standard and goal-oriented adaptive mesh refinement (AMR) techniques are presented for the linear Boltzmann transport equation. A posteriori error estimates are employed to drive the AMR process and are based on angular-moment information rather than on directional information, leading to direction-independent adapted meshes. An error estimate based on a two-mesh approach and a jump-based error indicator are compared for various test problems. In addition to the standard AMR approach, where the global error in the solution is diminished, a goal-oriented AMR procedure is devised and aims at reducing the error in user-specified quantities of interest. The quantities of interest are functionals of the solution and may include, for instance, point-wise flux values or average reaction rates in a subdomain. A high-order (up to order 4) Discontinuous Galerkin technique with standard upwinding is employed for the spatial discretization; the discrete ordinates method is used to treat the angular variable.
NASA Astrophysics Data System (ADS)
Schwing, Alan Michael
For computational fluid dynamics, the governing equations are solved on a discretized domain of nodes, faces, and cells. The quality of the grid or mesh can be a driving source for error in the results. While refinement studies can help guide the creation of a mesh, grid quality is largely determined by user expertise and understanding of the flow physics. Adaptive mesh refinement is a technique for enriching the mesh during a simulation based on metrics for error, impact on important parameters, or location of important flow features. This can offload from the user some of the difficult and ambiguous decisions necessary when discretizing the domain. This work explores the implementation of adaptive mesh refinement in an implicit, unstructured, finite-volume solver. Consideration is made for applying modern computational techniques in the presence of hanging nodes and refined cells. The approach is developed to be independent of the flow solver in order to provide a path for augmenting existing codes. It is designed to be applicable for unsteady simulations and refinement and coarsening of the grid does not impact the conservatism of the underlying numerics. The effect on high-order numerical fluxes of fourth- and sixth-order are explored. Provided the criteria for refinement is appropriately selected, solutions obtained using adapted meshes have no additional error when compared to results obtained on traditional, unadapted meshes. In order to leverage large-scale computational resources common today, the methods are parallelized using MPI. Parallel performance is considered for several test problems in order to assess scalability of both adapted and unadapted grids. Dynamic repartitioning of the mesh during refinement is crucial for load balancing an evolving grid. Development of the methods outlined here depend on a dual-memory approach that is described in detail. Validation of the solver developed here against a number of motivating problems shows favorable
Adaptive mesh refinement strategies in isogeometric analysis— A computational comparison
NASA Astrophysics Data System (ADS)
Hennig, Paul; Kästner, Markus; Morgenstern, Philipp; Peterseim, Daniel
2017-04-01
We explain four variants of an adaptive finite element method with cubic splines and compare their performance in simple elliptic model problems. The methods in comparison are Truncated Hierarchical B-splines with two different refinement strategies, T-splines with the refinement strategy introduced by Scott et al. in 2012, and T-splines with an alternative refinement strategy introduced by some of the authors. In four examples, including singular and non-singular problems of linear elasticity and the Poisson problem, the H1-errors of the discrete solutions, the number of degrees of freedom as well as sparsity patterns and condition numbers of the discretized problem are compared.
Numerical study of three-dimensional liquid jet breakup with adaptive unstructured meshes
NASA Astrophysics Data System (ADS)
Xie, Zhihua; Pavlidis, Dimitrios; Salinas, Pablo; Pain, Christopher; Matar, Omar
2016-11-01
Liquid jet breakup is an important fundamental multiphase flow, often found in many industrial engineering applications. The breakup process is very complex, involving jets, liquid films, ligaments, and small droplets, featuring tremendous complexity in interfacial topology and a large range of spatial scales. The objective of this study is to investigate the fluid dynamics of three-dimensional liquid jet breakup problems, such as liquid jet primary breakup and gas-sheared liquid jet breakup. An adaptive unstructured mesh modelling framework is employed here, which can modify and adapt unstructured meshes to optimally represent the underlying physics of multiphase problems and reduce computational effort without sacrificing accuracy. The numerical framework consists of a mixed control volume and finite element formulation, a 'volume of fluid' type method for the interface capturing based on a compressive control volume advection method and second-order finite element methods, and a force-balanced algorithm for the surface tension implementation. Numerical examples of some benchmark tests and the dynamics of liquid jet breakup with and without ambient gas are presented to demonstrate the capability of this method.
Output-based mesh adaptation for high order Navier-Stokes simulations on deformable domains
NASA Astrophysics Data System (ADS)
Kast, Steven M.; Fidkowski, Krzysztof J.
2013-11-01
We present an output-based mesh adaptation strategy for Navier-Stokes simulations on deforming domains. The equations are solved with an arbitrary Lagrangian-Eulerian (ALE) approach, using a discontinuous Galerkin finite-element discretization in both space and time. Discrete unsteady adjoint solutions, derived for both the state and the geometric conservation law, provide output error estimates and drive adaptation of the space-time mesh. Spatial adaptation consists of dynamic order increment or decrement on a fixed tessellation of the domain, while a combination of coarsening and refinement is used to provide an efficient time step distribution. Results from compressible Navier-Stokes simulations in both two and three dimensions demonstrate the accuracy and efficiency of the proposed approach. In particular, the method is shown to outperform other common adaptation strategies, which, while sometimes adequate for static problems, struggle in the presence of mesh motion.
NASA Astrophysics Data System (ADS)
Owens, A. R.; Kópházi, J.; Welch, J. A.; Eaton, M. D.
2017-04-01
In this paper a hanging-node, discontinuous Galerkin, isogeometric discretisation of the multigroup, discrete ordinates (SN) equations is presented in which each energy group has its own mesh. The equations are discretised using Non-Uniform Rational B-Splines (NURBS), which allows the coarsest mesh to exactly represent the geometry for a wide range of engineering problems of interest; this would not be the case using straight-sided finite elements. Information is transferred between meshes via the construction of a supermesh. This is a non-trivial task for two arbitrary meshes, but is significantly simplified here by deriving every mesh from a common coarsest initial mesh. In order to take full advantage of this flexible discretisation, goal-based error estimators are derived for the multigroup, discrete ordinates equations with both fixed (extraneous) and fission sources, and these estimators are used to drive an adaptive mesh refinement (AMR) procedure. The method is applied to a variety of test cases for both fixed and fission source problems. The error estimators are found to be extremely accurate for linear NURBS discretisations, with degraded performance for quadratic discretisations owing to a reduction in relative accuracy of the ;exact; adjoint solution required to calculate the estimators. Nevertheless, the method seems to produce optimal meshes in the AMR process for both linear and quadratic discretisations, and is ≈×100 more accurate than uniform refinement for the same amount of computational effort for a 67 group deep penetration shielding problem.
Fast animation of lightning using an adaptive mesh.
Kim, Theodore; Lin, Ming C
2007-01-01
We present a fast method for simulating, animating, and rendering lightning using adaptive grids. The "dielectric breakdown model" is an elegant algorithm for electrical pattern formation that we extend to enable animation of lightning. The simulation can be slow, particularly in 3D, because it involves solving a large Poisson problem. Losasso et al. recently proposed an octree data structure for simulating water and smoke, and we show that this discretization can be applied to the problem of lightning simulation as well. However, implementing the incomplete Cholesky conjugate gradient (ICCG) solver for this problem can be daunting, so we provide an extensive discussion of implementation issues. ICCG solvers can usually be accelerated using "Eisenstat's trick," but the trick cannot be directly applied to the adaptive case. Fortunately, we show that an "almost incomplete Cholesky" factorization can be computed so that Eisenstat's trick can still be used. We then present a fast rendering method based on convolution that is competitive with Monte Carlo ray tracing but orders of magnitude faster, and we also show how to further improve the visual results using jittering.
A Numerical Study of Mesh Adaptivity in Multiphase Flows with Non-Newtonian Fluids
NASA Astrophysics Data System (ADS)
Percival, James; Pavlidis, Dimitrios; Xie, Zhihua; Alberini, Federico; Simmons, Mark; Pain, Christopher; Matar, Omar
2014-11-01
We present an investigation into the computational efficiency benefits of dynamic mesh adaptivity in the numerical simulation of transient multiphase fluid flow problems involving Non-Newtonian fluids. Such fluids appear in a range of industrial applications, from printing inks to toothpastes and introduce new challenges for mesh adaptivity due to the additional ``memory'' of viscoelastic fluids. Nevertheless, the multiscale nature of these flows implies huge potential benefits for a successful implementation. The study is performed using the open source package Fluidity, which couples an unstructured mesh control volume finite element solver for the multiphase Navier-Stokes equations to a dynamic anisotropic mesh adaptivity algorithm, based on estimated solution interpolation error criteria, and conservative mesh-to-mesh interpolation routine. The code is applied to problems involving rheologies ranging from simple Newtonian to shear-thinning to viscoelastic materials and verified against experimental data for various industrial and microfluidic flows. This work was undertaken as part of the EPSRC MEMPHIS programme grant EP/K003976/1.
3D Adaptive Mesh Refinement Simulations of Pellet Injection in Tokamaks
R. Samtaney; S.C. Jardin; P. Colella; D.F. Martin
2003-10-20
We present results of Adaptive Mesh Refinement (AMR) simulations of the pellet injection process, a proven method of refueling tokamaks. AMR is a computationally efficient way to provide the resolution required to simulate realistic pellet sizes relative to device dimensions. The mathematical model comprises of single-fluid MHD equations with source terms in the continuity equation along with a pellet ablation rate model. The numerical method developed is an explicit unsplit upwinding treatment of the 8-wave formulation, coupled with a MAC projection method to enforce the solenoidal property of the magnetic field. The Chombo framework is used for AMR. The role of the E x B drift in mass redistribution during inside and outside pellet injections is emphasized.
WHITE DWARF MERGERS ON ADAPTIVE MESHES. I. METHODOLOGY AND CODE VERIFICATION
Katz, Max P.; Zingale, Michael; Calder, Alan C.; Swesty, F. Douglas; Almgren, Ann S.; Zhang, Weiqun
2016-03-10
The Type Ia supernova (SN Ia) progenitor problem is one of the most perplexing and exciting problems in astrophysics, requiring detailed numerical modeling to complement observations of these explosions. One possible progenitor that has merited recent theoretical attention is the white dwarf (WD) merger scenario, which has the potential to naturally explain many of the observed characteristics of SNe Ia. To date there have been relatively few self-consistent simulations of merging WD systems using mesh-based hydrodynamics. This is the first paper in a series describing simulations of these systems using a hydrodynamics code with adaptive mesh refinement. In this paper we describe our numerical methodology and discuss our implementation in the compressible hydrodynamics code CASTRO, which solves the Euler equations, and the Poisson equation for self-gravity, and couples the gravitational and rotation forces to the hydrodynamics. Standard techniques for coupling gravitation and rotation forces to the hydrodynamics do not adequately conserve the total energy of the system for our problem, but recent advances in the literature allow progress and we discuss our implementation here. We present a set of test problems demonstrating the extent to which our software sufficiently models a system where large amounts of mass are advected on the computational domain over long timescales. Future papers in this series will describe our treatment of the initial conditions of these systems and will examine the early phases of the merger to determine its viability for triggering a thermonuclear detonation.
White Dwarf Mergers on Adaptive Meshes. I. Methodology and Code Verification
NASA Astrophysics Data System (ADS)
Katz, Max P.; Zingale, Michael; Calder, Alan C.; Swesty, F. Douglas; Almgren, Ann S.; Zhang, Weiqun
2016-03-01
The Type Ia supernova (SN Ia) progenitor problem is one of the most perplexing and exciting problems in astrophysics, requiring detailed numerical modeling to complement observations of these explosions. One possible progenitor that has merited recent theoretical attention is the white dwarf (WD) merger scenario, which has the potential to naturally explain many of the observed characteristics of SNe Ia. To date there have been relatively few self-consistent simulations of merging WD systems using mesh-based hydrodynamics. This is the first paper in a series describing simulations of these systems using a hydrodynamics code with adaptive mesh refinement. In this paper we describe our numerical methodology and discuss our implementation in the compressible hydrodynamics code CASTRO, which solves the Euler equations, and the Poisson equation for self-gravity, and couples the gravitational and rotation forces to the hydrodynamics. Standard techniques for coupling gravitation and rotation forces to the hydrodynamics do not adequately conserve the total energy of the system for our problem, but recent advances in the literature allow progress and we discuss our implementation here. We present a set of test problems demonstrating the extent to which our software sufficiently models a system where large amounts of mass are advected on the computational domain over long timescales. Future papers in this series will describe our treatment of the initial conditions of these systems and will examine the early phases of the merger to determine its viability for triggering a thermonuclear detonation.
Automatic off-body overset adaptive Cartesian mesh method based on an octree approach
NASA Astrophysics Data System (ADS)
Péron, Stéphanie; Benoit, Christophe
2013-01-01
This paper describes a method for generating adaptive structured Cartesian grids within a near-body/off-body mesh partitioning framework for the flow simulation around complex geometries. The off-body Cartesian mesh generation derives from an octree structure, assuming each octree leaf node defines a structured Cartesian block. This enables one to take into account the large scale discrepancies in terms of resolution between the different bodies involved in the simulation, with minimum memory requirements. Two different conversions from the octree to Cartesian grids are proposed: the first one generates Adaptive Mesh Refinement (AMR) type grid systems, and the second one generates abutting or minimally overlapping Cartesian grid set. We also introduce an algorithm to control the number of points at each adaptation, that automatically determines relevant values of the refinement indicator driving the grid refinement and coarsening. An application to a wing tip vortex computation assesses the capability of the method to capture accurately the flow features.
Development of a scalable gas-dynamics solver with adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Korkut, Burak
There are various computational physics areas in which Direct Simulation Monte Carlo (DSMC) and Particle in Cell (PIC) methods are being employed. The accuracy of results from such simulations depend on the fidelity of the physical models being used. The computationally demanding nature of these problems make them ideal candidates to make use of modern supercomputers. The software developed to run such simulations also needs special attention so that the maintainability and extendability is considered with the recent numerical methods and programming paradigms. Suited for gas-dynamics problems, a software called SUGAR (Scalable Unstructured Gas dynamics with Adaptive mesh Refinement) has recently been developed and written in C++ and MPI. Physical and numerical models were added to this framework to simulate ion thruster plumes. SUGAR is used to model the charge-exchange (CEX) reactions occurring between the neutral and ion species as well as the induced electric field effect due to ions. Multiple adaptive mesh refinement (AMR) meshes were used in order to capture different physical length scales present in the flow. A multiple-thruster configuration was run to extend the studies to cases for which there is no axial or radial symmetry present that could only be modeled with a three-dimensional simulation capability. The combined plume structure showed interactions between individual thrusters where AMR capability captured this in an automated way. The back flow for ions was found to occur when CEX and momentum-exchange (MEX) collisions are present and strongly enhanced when the induced electric field is considered. The ion energy distributions in the back flow region were obtained and it was found that the inclusion of the electric field modeling is the most important factor in determining its shape. The plume back flow structure was also examined for a triple-thruster, 3-D geometry case and it was found that the ion velocity in the back flow region appears to be
Adaptive unstructured meshing for thermal stress analysis of built-up structures
NASA Technical Reports Server (NTRS)
Dechaumphai, Pramote
1992-01-01
An adaptive unstructured meshing technique for mechanical and thermal stress analysis of built-up structures has been developed. A triangular membrane finite element and a new plate bending element are evaluated on a panel with a circular cutout and a frame stiffened panel. The adaptive unstructured meshing technique, without a priori knowledge of the solution to the problem, generates clustered elements only where needed. An improved solution accuracy is obtained at a reduced problem size and analysis computational time as compared to the results produced by the standard finite element procedure.
Parallelization of Unsteady Adaptive Mesh Refinement for Unstructured Navier-Stokes Solvers
NASA Technical Reports Server (NTRS)
Schwing, Alan M.; Nompelis, Ioannis; Candler, Graham V.
2014-01-01
This paper explores the implementation of the MPI parallelization in a Navier-Stokes solver using adaptive mesh re nement. Viscous and inviscid test problems are considered for the purpose of benchmarking, as are implicit and explicit time advancement methods. The main test problem for comparison includes e ects from boundary layers and other viscous features and requires a large number of grid points for accurate computation. Ex- perimental validation against double cone experiments in hypersonic ow are shown. The adaptive mesh re nement shows promise for a staple test problem in the hypersonic com- munity. Extension to more advanced techniques for more complicated ows is described.
Using Multi-threading for the Automatic Load Balancing of 2D Adaptive Finite Element Meshes
NASA Technical Reports Server (NTRS)
Heber, Gerd; Biswas, Rupak; Thulasiraman, Parimala; Gao, Guang R.; Saini, Subhash (Technical Monitor)
1998-01-01
In this paper, we present a multi-threaded approach for the automatic load balancing of adaptive finite element (FE) meshes The platform of our choice is the EARTH multi-threaded system which offers sufficient capabilities to tackle this problem. We implement the adaption phase of FE applications oil triangular meshes and exploit the EARTH token mechanism to automatically balance the resulting irregular and highly nonuniform workload. We discuss the results of our experiments oil EARTH-SP2, on implementation of EARTH on the IBM SP2 with different load balancing strategies that are built into the runtime system.
Thickness-based adaptive mesh refinement methods for multi-phase flow simulations with thin regions
Chen, Xiaodong; Yang, Vigor
2014-07-15
In numerical simulations of multi-scale, multi-phase flows, grid refinement is required to resolve regions with small scales. A notable example is liquid-jet atomization and subsequent droplet dynamics. It is essential to characterize the detailed flow physics with variable length scales with high fidelity, in order to elucidate the underlying mechanisms. In this paper, two thickness-based mesh refinement schemes are developed based on distance- and topology-oriented criteria for thin regions with confining wall/plane of symmetry and in any situation, respectively. Both techniques are implemented in a general framework with a volume-of-fluid formulation and an adaptive-mesh-refinement capability. The distance-oriented technique compares against a critical value, the ratio of an interfacial cell size to the distance between the mass center of the cell and a reference plane. The topology-oriented technique is developed from digital topology theories to handle more general conditions. The requirement for interfacial mesh refinement can be detected swiftly, without the need of thickness information, equation solving, variable averaging or mesh repairing. The mesh refinement level increases smoothly on demand in thin regions. The schemes have been verified and validated against several benchmark cases to demonstrate their effectiveness and robustness. These include the dynamics of colliding droplets, droplet motions in a microchannel, and atomization of liquid impinging jets. Overall, the thickness-based refinement technique provides highly adaptive meshes for problems with thin regions in an efficient and fully automatic manner.
2015-04-09
numerical weather prediction (NWP) models is that28 they be unified, or that one nonhydrostatic dynamical core has the capability of simulating a29 wide...spectrum atmospheric spatial and temporal scales of motion, from microscale to global,30 and weather to climate. Severe and high-impact weather can...often take the form of localized31 weather systems, such as severe thunderstorms, tornadoes, fronts, and tropical cyclones.32 With limited computational
Detached Eddy Simulation of the UH-60 Rotor Wake Using Adaptive Mesh Refinement
NASA Technical Reports Server (NTRS)
Chaderjian, Neal M.; Ahmad, Jasim U.
2012-01-01
Time-dependent Navier-Stokes flow simulations have been carried out for a UH-60 rotor with simplified hub in forward flight and hover flight conditions. Flexible rotor blades and flight trim conditions are modeled and established by loosely coupling the OVERFLOW Computational Fluid Dynamics (CFD) code with the CAMRAD II helicopter comprehensive code. High order spatial differences, Adaptive Mesh Refinement (AMR), and Detached Eddy Simulation (DES) are used to obtain highly resolved vortex wakes, where the largest turbulent structures are captured. Special attention is directed towards ensuring the dual time accuracy is within the asymptotic range, and verifying the loose coupling convergence process using AMR. The AMR/DES simulation produced vortical worms for forward flight and hover conditions, similar to previous results obtained for the TRAM rotor in hover. AMR proved to be an efficient means to capture a rotor wake without a priori knowledge of the wake shape.
On the Computation of Integral Curves in Adaptive Mesh Refinement Vector Fields
Deines, Eduard; Weber, Gunther H.; Garth, Christoph; Van Straalen, Brian; Borovikov, Sergey; Martin, Daniel F.; Joy, Kenneth I.
2011-06-27
Integral curves, such as streamlines, streaklines, pathlines, and timelines, are an essential tool in the analysis of vector field structures, offering straightforward and intuitive interpretation of visualization results. While such curves have a long-standing tradition in vector field visualization, their application to Adaptive Mesh Refinement (AMR) simulation results poses unique problems. AMR is a highly effective discretization method for a variety of physical simulation problems and has recently been applied to the study of vector fields in flow and magnetohydrodynamic applications. The cell-centered nature of AMR data and discontinuities in the vector field representation arising from AMR level boundaries complicate the application of numerical integration methods to compute integral curves. In this paper, we propose a novel approach to alleviate these problems and show its application to streamline visualization in an AMR model of the magnetic field of the solar system as well as to a simulation of two incompressible viscous vortex rings merging.
A GPU implementation of adaptive mesh refinement to simulate tsunamis generated by landslides
NASA Astrophysics Data System (ADS)
de la Asunción, Marc; Castro, Manuel J.
2016-04-01
In this work we propose a CUDA implementation for the simulation of landslide-generated tsunamis using a two-layer Savage-Hutter type model and adaptive mesh refinement (AMR). The AMR method consists of dynamically increasing the spatial resolution of the regions of interest of the domain while keeping the rest of the domain at low resolution, thus obtaining better runtimes and similar results compared to increasing the spatial resolution of the entire domain. Our AMR implementation uses a patch-based approach, it supports up to three levels, power-of-two ratios of refinement, different refinement criteria and also several user parameters to control the refinement and clustering behaviour. A strategy based on the variation of the cell values during the simulation is used to interpolate and propagate the values of the fine cells. Several numerical experiments using artificial and realistic scenarios are presented.
SU-D-207-04: GPU-Based 4D Cone-Beam CT Reconstruction Using Adaptive Meshing Method
Zhong, Z; Gu, X; Iyengar, P; Mao, W; Wang, J; Guo, X
2015-06-15
Purpose: Due to the limited number of projections at each phase, the image quality of a four-dimensional cone-beam CT (4D-CBCT) is often degraded, which decreases the accuracy of subsequent motion modeling. One of the promising methods is the simultaneous motion estimation and image reconstruction (SMEIR) approach. The objective of this work is to enhance the computational speed of the SMEIR algorithm using adaptive feature-based tetrahedral meshing and GPU-based parallelization. Methods: The first step is to generate the tetrahedral mesh based on the features of a reference phase 4D-CBCT, so that the deformation can be well captured and accurately diffused from the mesh vertices to voxels of the image volume. After the mesh generation, the updated motion model and other phases of 4D-CBCT can be obtained by matching the 4D-CBCT projection images at each phase with the corresponding forward projections of the deformed reference phase of 4D-CBCT. The entire process of this 4D-CBCT reconstruction method is implemented on GPU, resulting in significantly increasing the computational efficiency due to its tremendous parallel computing ability. Results: A 4D XCAT digital phantom was used to test the proposed mesh-based image reconstruction algorithm. The image Result shows both bone structures and inside of the lung are well-preserved and the tumor position can be well captured. Compared to the previous voxel-based CPU implementation of SMEIR, the proposed method is about 157 times faster for reconstructing a 10 -phase 4D-CBCT with dimension 256×256×150. Conclusion: The GPU-based parallel 4D CBCT reconstruction method uses the feature-based mesh for estimating motion model and demonstrates equivalent image Result with previous voxel-based SMEIR approach, with significantly improved computational speed.
Hornung, R.D.
1996-12-31
An adaptive local mesh refinement (AMR) algorithm originally developed for unsteady gas dynamics is extended to multi-phase flow in porous media. Within the AMR framework, we combine specialized numerical methods to treat the different aspects of the partial differential equations. Multi-level iteration and domain decomposition techniques are incorporated to accommodate elliptic/parabolic behavior. High-resolution shock capturing schemes are used in the time integration of the hyperbolic mass conservation equations. When combined with AMR, these numerical schemes provide high resolution locally in a more efficient manner than if they were applied on a uniformly fine computational mesh. We will discuss the interplay of physical, mathematical, and numerical concerns in the application of adaptive mesh refinement to flow in porous media problems of practical interest.
Adaptive Mesh Refinement With Spectral Accuracy for Magnetohydrodynamics in Two Space Dimensions
NASA Astrophysics Data System (ADS)
Rosenberg, D.; Pouquet, A.; Mininni, P.
2006-12-01
We examine the effect of accuracy of high-order adaptive mesh refinement (AMR) in the context of a classical configuration of magnetic reconnection in two space dimensions, the so-called Orszag-Tang vortex made up of a magnetic X-point centered on a stagnation point of the velocity. A recently developed spectral-element adaptive refinement incompressible magnetohydrodynamic (MHD) code is applied to simulate this problem. The MHD solver is explicit, and uses the Elsasser formulation on high-order elements. It automatically takes advantage of the adaptive grid mechanics that have been described elsewhere [Rosenberg, Fournier, Fischer, Pouquet, J. Comp. Phys. 215, 59-80 (2006)] in the fluid context, allowing both statically refined and dynamically refined grids. Comparisons with pseudo-spectral computations are performed. Refinement and coarsening criteria are examined, and several tests are described. We show that low-order truncation--even with a comparable number of global degrees of freedom--fails to correctly model some strong (inf-norm) quantities in this problem, even though it satisfies adequately the weak (integrated) balance diagnostics.
Parallel, Gradient-Based Anisotropic Mesh Adaptation for Re-entry Vehicle Configurations
NASA Technical Reports Server (NTRS)
Bibb, Karen L.; Gnoffo, Peter A.; Park, Michael A.; Jones, William T.
2006-01-01
Two gradient-based adaptation methodologies have been implemented into the Fun3d refine GridEx infrastructure. A spring-analogy adaptation which provides for nodal movement to cluster mesh nodes in the vicinity of strong shocks has been extended for general use within Fun3d, and is demonstrated for a 70 sphere cone at Mach 2. A more general feature-based adaptation metric has been developed for use with the adaptation mechanics available in Fun3d, and is applicable to any unstructured, tetrahedral, flow solver. The basic functionality of general adaptation is explored through a case of flow over the forebody of a 70 sphere cone at Mach 6. A practical application of Mach 10 flow over an Apollo capsule, computed with the Felisa flow solver, is given to compare the adaptive mesh refinement with uniform mesh refinement. The examples of the paper demonstrate that the gradient-based adaptation capability as implemented can give an improvement in solution quality.
Failure of Anisotropic Unstructured Mesh Adaption Based on Multidimensional Residual Minimization
NASA Technical Reports Server (NTRS)
Wood, William A.; Kleb, William L.
2003-01-01
An automated anisotropic unstructured mesh adaptation strategy is proposed, implemented, and assessed for the discretization of viscous flows. The adaption criteria is based upon the minimization of the residual fluctuations of a multidimensional upwind viscous flow solver. For scalar advection, this adaption strategy has been shown to use fewer grid points than gradient based adaption, naturally aligning mesh edges with discontinuities and characteristic lines. The adaption utilizes a compact stencil and is local in scope, with four fundamental operations: point insertion, point deletion, edge swapping, and nodal displacement. Evaluation of the solution-adaptive strategy is performed for a two-dimensional blunt body laminar wind tunnel case at Mach 10. The results demonstrate that the strategy suffers from a lack of robustness, particularly with regard to alignment of the bow shock in the vicinity of the stagnation streamline. In general, constraining the adaption to such a degree as to maintain robustness results in negligible improvement to the solution. Because the present method fails to consistently or significantly improve the flow solution, it is rejected in favor of simple uniform mesh refinement.
NASA Astrophysics Data System (ADS)
Barral, N.; Olivier, G.; Alauzet, F.
2017-02-01
Anisotropic metric-based mesh adaptation has proved its efficiency to reduce the CPU time of steady and unsteady simulations while improving their accuracy. However, its extension to time-dependent problems with body-fitted moving geometries is far from straightforward. This paper establishes a well-founded framework for multiscale mesh adaptation of unsteady problems with moving boundaries. This framework is based on a novel space-time analysis of the interpolation error, within the continuous mesh theory. An optimal metric field, called ALE metric field, is derived, which takes into account the movement of the mesh during the adaptation. Based on this analysis, the global fixed-point adaptation algorithm for time-dependent simulations is extended to moving boundary problems, within the range of body-fitted moving meshes and ALE simulations. Finally, three dimensional adaptive simulations with moving boundaries are presented to validate the proposed approach.
CRASH: A BLOCK-ADAPTIVE-MESH CODE FOR RADIATIVE SHOCK HYDRODYNAMICS-IMPLEMENTATION AND VERIFICATION
Van der Holst, B.; Toth, G.; Sokolov, I. V.; Myra, E. S.; Fryxell, B.; Drake, R. P.; Powell, K. G.; Holloway, J. P.; Stout, Q.; Adams, M. L.; Morel, J. E.; Karni, S.
2011-06-01
We describe the Center for Radiative Shock Hydrodynamics (CRASH) code, a block-adaptive-mesh code for multi-material radiation hydrodynamics. The implementation solves the radiation diffusion model with a gray or multi-group method and uses a flux-limited diffusion approximation to recover the free-streaming limit. Electrons and ions are allowed to have different temperatures and we include flux-limited electron heat conduction. The radiation hydrodynamic equations are solved in the Eulerian frame by means of a conservative finite-volume discretization in either one-, two-, or three-dimensional slab geometry or in two-dimensional cylindrical symmetry. An operator-split method is used to solve these equations in three substeps: (1) an explicit step of a shock-capturing hydrodynamic solver; (2) a linear advection of the radiation in frequency-logarithm space; and (3) an implicit solution of the stiff radiation diffusion, heat conduction, and energy exchange. We present a suite of verification test problems to demonstrate the accuracy and performance of the algorithms. The applications are for astrophysics and laboratory astrophysics. The CRASH code is an extension of the Block-Adaptive Tree Solarwind Roe Upwind Scheme (BATS-R-US) code with a new radiation transfer and heat conduction library and equation-of-state and multi-group opacity solvers. Both CRASH and BATS-R-US are part of the publicly available Space Weather Modeling Framework.
NASA Astrophysics Data System (ADS)
Kopera, Michal A.; Giraldo, Francis X.
2014-10-01
The resolutions of interests in atmospheric simulations require prohibitively large computational resources. Adaptive mesh refinement (AMR) tries to mitigate this problem by putting high resolution in crucial areas of the domain. We investigate the performance of a tree-based AMR algorithm for the high order discontinuous Galerkin method on quadrilateral grids with non-conforming elements. We perform a detailed analysis of the cost of AMR by comparing this to uniform reference simulations of two standard atmospheric test cases: density current and rising thermal bubble. The analysis shows up to 15 times speed-up of the AMR simulations with the cost of mesh adaptation below 1% of the total runtime. We pay particular attention to the implicit-explicit (IMEX) time integration methods and show that the ARK2 method is more robust with respect to dynamically adapting meshes than BDF2. Preliminary analysis of preconditioning reveals that it can be an important factor in the AMR overhead. The compiler optimizations provide significant runtime reduction and positively affect the effectiveness of AMR allowing for speed-ups greater than it would follow from the simple performance model.
Time-dependent grid adaptation for meshes of triangles and tetrahedra
NASA Technical Reports Server (NTRS)
Rausch, Russ D.
1993-01-01
This paper presents in viewgraph form a method of optimizing grid generation for unsteady CFD flow calculations that distributes the numerical error evenly throughout the mesh. Adaptive meshing is used to locally enrich in regions of relatively large errors and to locally coarsen in regions of relatively small errors. The enrichment/coarsening procedures are robust for isotropic cells; however, enrichment of high aspect ratio cells may fail near boundary surfaces with relatively large curvature. The enrichment indicator worked well for the cases shown, but in general requires user supervision for a more efficient solution.
Adaptive Mesh Refinement in Curvilinear Body-Fitted Grid Systems
NASA Technical Reports Server (NTRS)
Steinthorsson, Erlendur; Modiano, David; Colella, Phillip
1995-01-01
To be truly compatible with structured grids, an AMR algorithm should employ a block structure for the refined grids to allow flow solvers to take advantage of the strengths of unstructured grid systems, such as efficient solution algorithms for implicit discretizations and multigrid schemes. One such algorithm, the AMR algorithm of Berger and Colella, has been applied to and adapted for use with body-fitted structured grid systems. Results are presented for a transonic flow over a NACA0012 airfoil (AGARD-03 test case) and a reflection of a shock over a double wedge.
PLUM: Parallel Load Balancing for Unstructured Adaptive Meshes. Degree awarded by Colorado Univ.
NASA Technical Reports Server (NTRS)
Oliker, Leonid
1998-01-01
Dynamic mesh adaption on unstructured grids is a powerful tool for computing large-scale problems that require grid modifications to efficiently resolve solution features. By locally refining and coarsening the mesh to capture physical phenomena of interest, such procedures make standard computational methods more cost effective. Unfortunately, an efficient parallel implementation of these adaptive methods is rather difficult to achieve, primarily due to the load imbalance created by the dynamically-changing nonuniform grid. This requires significant communication at runtime, leading to idle processors and adversely affecting the total execution time. Nonetheless, it is generally thought that unstructured adaptive- grid techniques will constitute a significant fraction of future high-performance supercomputing. Various dynamic load balancing methods have been reported to date; however, most of them either lack a global view of loads across processors or do not apply their techniques to realistic large-scale applications.
An adaptive embedded mesh procedure for leading-edge vortex flows
NASA Technical Reports Server (NTRS)
Powell, Kenneth G.; Beer, Michael A.; Law, Glenn W.
1989-01-01
A procedure for solving the conical Euler equations on an adaptively refined mesh is presented, along with a method for determining which cells to refine. The solution procedure is a central-difference cell-vertex scheme. The adaptation procedure is made up of a parameter on which the refinement decision is based, and a method for choosing a threshold value of the parameter. The refinement parameter is a measure of mesh-convergence, constructed by comparison of locally coarse- and fine-grid solutions. The threshold for the refinement parameter is based on the curvature of the curve relating the number of cells flagged for refinement to the value of the refinement threshold. Results for three test cases are presented. The test problem is that of a delta wing at angle of attack in a supersonic free-stream. The resulting vortices and shocks are captured efficiently by the adaptive code.
An Adaptive Discontinuous Galerkin Method for Modeling Atmospheric Convection (Preprint)
2011-04-13
Giraldo and Volkmar Wirth 5 SENSITIVITY STUDIES One important question for each adaptive numerical model is: how accurate is the adaptive method? For...this criterion that is used later for some sensitivity studies . These studies include a comparison between a simulation on an adaptive mesh with a...simulation on a uniform mesh and a sensitivity study concerning the size of the refinement region. 5.1 Comparison Criterion For comparing different
NASA Astrophysics Data System (ADS)
Adam, A.; Pavlidis, D.; Percival, J. R.; Salinas, P.; Xie, Z.; Fang, F.; Pain, C. C.; Muggeridge, A. H.; Jackson, M. D.
2016-09-01
A general, higher-order, conservative and bounded interpolation for the dynamic and adaptive meshing of control-volume fields dual to continuous and discontinuous finite element representations is presented. Existing techniques such as node-wise interpolation are not conservative and do not readily generalise to discontinuous fields, whilst conservative methods such as Grandy interpolation are often too diffusive. The new method uses control-volume Galerkin projection to interpolate between control-volume fields. Bounded solutions are ensured by using a post-interpolation diffusive correction. Example applications of the method to interface capturing during advection and also to the modelling of multiphase porous media flow are presented to demonstrate the generality and robustness of the approach.
Implementation of Implicit Adaptive Mesh Refinement in an Unstructured Finite-Volume Flow Solver
NASA Technical Reports Server (NTRS)
Schwing, Alan M.; Nompelis, Ioannis; Candler, Graham V.
2013-01-01
This paper explores the implementation of adaptive mesh refinement in an unstructured, finite-volume solver. Unsteady and steady problems are considered. The effect on the recovery of high-order numerics is explored and the results are favorable. Important to this work is the ability to provide a path for efficient, implicit time advancement. A method using a simple refinement sensor based on undivided differences is discussed and applied to a practical problem: a shock-shock interaction on a hypersonic, inviscid double-wedge. Cases are compared to uniform grids without the use of adapted meshes in order to assess error and computational expense. Discussion of difficulties, advances, and future work prepare this method for additional research. The potential for this method in more complicated flows is described.
RCS and antenna modeling with MOM using hybrid meshes
Putnam, J.M.; Kotulski, J.D.
1997-02-01
In this presentation, the authors will investigate the use of hybrid meshes for modeling RCS and antenna problems in three dimensions. They will consider two classes of hybrid basis functions. These include combinations of quadrilateral and triangular meshes for arbitrary 3D geometries, and combinations of axisymmetric body-of-revolution (BOR) basis functions and triangular facets. In particular, they will focus on the problem of enforcing current continuity between two surfaces which are represented by different types of surface discretizations and unknown basis function representations. They will illustrate the use of an operator-based code architecture for the implementation of these formulations, and how it facilitates the incorporation of the various types of boundary conditions in the code. Both serial and parallel code implementation issues for the formulations will be discussed. Results will be presented for both scattering and antenna problems. The emphasis will be on accuracy, and robustness of the techniques. Comparisons of accuracy between triangular meshed and quadrilateral meshed geometries will be shown. The use of hybrid meshes for modeling BORs with attached appendages will also be presented.
Turner, C. David; Kotulski, Joseph Daniel; Pasik, Michael Francis
2005-12-01
This report investigates the feasibility of applying Adaptive Mesh Refinement (AMR) techniques to a vector finite element formulation for the wave equation in three dimensions. Possible error estimators are considered first. Next, approaches for refining tetrahedral elements are reviewed. AMR capabilities within the Nevada framework are then evaluated. We summarize our conclusions on the feasibility of AMR for time-domain vector finite elements and identify a path forward.
4D cone-beam CT reconstruction using multi-organ meshes for sliding motion modeling
NASA Astrophysics Data System (ADS)
Zhong, Zichun; Gu, Xuejun; Mao, Weihua; Wang, Jing
2016-02-01
A simultaneous motion estimation and image reconstruction (SMEIR) strategy was proposed for 4D cone-beam CT (4D-CBCT) reconstruction and showed excellent results in both phantom and lung cancer patient studies. In the original SMEIR algorithm, the deformation vector field (DVF) was defined on voxel grid and estimated by enforcing a global smoothness regularization term on the motion fields. The objective of this work is to improve the computation efficiency and motion estimation accuracy of SMEIR for 4D-CBCT through developing a multi-organ meshing model. Feature-based adaptive meshes were generated to reduce the number of unknowns in the DVF estimation and accurately capture the organ shapes and motion. Additionally, the discontinuity in the motion fields between different organs during respiration was explicitly considered in the multi-organ mesh model. This will help with the accurate visualization and motion estimation of the tumor on the organ boundaries in 4D-CBCT. To further improve the computational efficiency, a GPU-based parallel implementation was designed. The performance of the proposed algorithm was evaluated on a synthetic sliding motion phantom, a 4D NCAT phantom, and four lung cancer patients. The proposed multi-organ mesh based strategy outperformed the conventional Feldkamp-Davis-Kress, iterative total variation minimization, original SMEIR and single meshing method based on both qualitative and quantitative evaluations.
4D cone-beam CT reconstruction using multi-organ meshes for sliding motion modeling
Zhong, Zichun; Gu, Xuejun; Mao, Weihua; Wang, Jing
2016-01-01
A simultaneous motion estimation and image reconstruction (SMEIR) strategy was proposed for 4D cone-beam CT (4D-CBCT) reconstruction and showed excellent results in both phantom and lung cancer patient studies. In the original SMEIR algorithm, the deformation vector field (DVF) was defined on voxel grid and estimated by enforcing a global smoothness regularization term on the motion fields. The objective of this work is to improve the computation efficiency and motion estimation accuracy of SMEIR for 4D-CBCT through developing a multi-organ meshing model. Feature-based adaptive meshes were generated to reduce the number of unknowns in the DVF estimation and accurately capture the organ shapes and motion. Additionally, the discontinuity in the motion fields between different organs during respiration was explicitly considered in the multi-organ mesh model. This will help with the accurate visualization and motion estimation of the tumor on the organ boundaries in 4D-CBCT. To further improve the computational efficiency, a GPU-based parallel implementation was designed. The performance of the proposed algorithm was evaluated on a synthetic sliding motion phantom, a 4D NCAT phantom, and four lung cancer patients. The proposed multi-organ mesh based strategy outperformed the conventional Feldkamp–Davis–Kress, iterative total variation minimization, original SMEIR and single meshing method based on both qualitative and quantitative evaluations. PMID:26758496
ADER-WENO finite volume schemes with space-time adaptive mesh refinement
NASA Astrophysics Data System (ADS)
Dumbser, Michael; Zanotti, Olindo; Hidalgo, Arturo; Balsara, Dinshaw S.
2013-09-01
We present the first high order one-step ADER-WENO finite volume scheme with adaptive mesh refinement (AMR) in multiple space dimensions. High order spatial accuracy is obtained through a WENO reconstruction, while a high order one-step time discretization is achieved using a local space-time discontinuous Galerkin predictor method. Due to the one-step nature of the underlying scheme, the resulting algorithm is particularly well suited for an AMR strategy on space-time adaptive meshes, i.e. with time-accurate local time stepping. The AMR property has been implemented 'cell-by-cell', with a standard tree-type algorithm, while the scheme has been parallelized via the message passing interface (MPI) paradigm. The new scheme has been tested over a wide range of examples for nonlinear systems of hyperbolic conservation laws, including the classical Euler equations of compressible gas dynamics and the equations of magnetohydrodynamics (MHD). High order in space and time have been confirmed via a numerical convergence study and a detailed analysis of the computational speed-up with respect to highly refined uniform meshes is also presented. We also show test problems where the presented high order AMR scheme behaves clearly better than traditional second order AMR methods. The proposed scheme that combines for the first time high order ADER methods with space-time adaptive grids in two and three space dimensions is likely to become a useful tool in several fields of computational physics, applied mathematics and mechanics.
Shape adjustment of cable mesh reflector antennas considering modeling uncertainties
NASA Astrophysics Data System (ADS)
Du, Jingli; Bao, Hong; Cui, Chuanzhen
2014-04-01
Cable mesh antennas are the most important implement to construct large space antennas nowadays. Reflector surface of cable mesh antennas has to be carefully adjusted to achieve required accuracy, which is an effective way to compensate manufacturing and assembly errors or other imperfections. In this paper shape adjustment of cable mesh antennas is addressed. The required displacement of the reflector surface is determined with respect to a modified paraboloid whose axial vertex offset is also considered as a variable. Then the adjustment problem is solved by minimizing the RMS error with respect to the desired paraboloid using minimal norm least squares method. To deal with the modeling uncertainties, the adjustment is achieved by solving a simple worst-case optimization problem instead of directly using the least squares method. A numerical example demonstrates the worst-case method is of good convergence and accuracy, and is robust to perturbations.
Multigroup radiation hydrodynamics with flux-limited diffusion and adaptive mesh refinement
NASA Astrophysics Data System (ADS)
González, M.; Vaytet, N.; Commerçon, B.; Masson, J.
2015-06-01
Context. Radiative transfer plays a crucial role in the star formation process. Because of the high computational cost, radiation-hydrodynamics simulations performed up to now have mainly been carried out in the grey approximation. In recent years, multifrequency radiation-hydrodynamics models have started to be developed in an attempt to better account for the large variations in opacities as a function of frequency. Aims: We wish to develop an efficient multigroup algorithm for the adaptive mesh refinement code RAMSES which is suited to heavy proto-stellar collapse calculations. Methods: Because of the prohibitive timestep constraints of an explicit radiative transfer method, we constructed a time-implicit solver based on a stabilized bi-conjugate gradient algorithm, and implemented it in RAMSES under the flux-limited diffusion approximation. Results: We present a series of tests that demonstrate the high performance of our scheme in dealing with frequency-dependent radiation-hydrodynamic flows. We also present a preliminary simulation of a 3D proto-stellar collapse using 20 frequency groups. Differences between grey and multigroup results are briefly discussed, and the large amount of information this new method brings us is also illustrated. Conclusions: We have implemented a multigroup flux-limited diffusion algorithm in the RAMSES code. The method performed well against standard radiation-hydrodynamics tests, and was also shown to be ripe for exploitation in the computational star formation context.
Effenberger, Frederic; Thust, Kay; Grauer, Rainer; Dreher, Juergen; Arnold, Lukas
2011-03-15
The formation of a thin current sheet in a magnetic quasiseparatrix layer (QSL) is investigated by means of numerical simulation using a simplified ideal, low-{beta}, MHD model. The initial configuration and driving boundary conditions are relevant to phenomena observed in the solar corona and were studied earlier by Aulanier et al. [Astron. Astrophys. 444, 961 (2005)]. In extension to that work, we use the technique of adaptive mesh refinement (AMR) to significantly enhance the local spatial resolution of the current sheet during its formation, which enables us to follow the evolution into a later stage. Our simulations are in good agreement with the results of Aulanier et al. up to the calculated time in that work. In a later phase, we observe a basically unarrested collapse of the sheet to length scales that are more than one order of magnitude smaller than those reported earlier. The current density attains correspondingly larger maximum values within the sheet. During this thinning process, which is finally limited by lack of resolution even in the AMR studies, the current sheet moves upward, following a global expansion of the magnetic structure during the quasistatic evolution. The sheet is locally one-dimensional and the plasma flow in its vicinity, when transformed into a comoving frame, qualitatively resembles a stagnation point flow. In conclusion, our simulations support the idea that extremely high current densities are generated in the vicinities of QSLs as a response to external perturbations, with no sign of saturation.
Patched based methods for adaptive mesh refinement solutions of partial differential equations
Saltzman, J.
1997-09-02
This manuscript contains the lecture notes for a course taught from July 7th through July 11th at the 1997 Numerical Analysis Summer School sponsored by C.E.A., I.N.R.I.A., and E.D.F. The subject area was chosen to support the general theme of that year`s school which is ``Multiscale Methods and Wavelets in Numerical Simulation.`` The first topic covered in these notes is a description of the problem domain. This coverage is limited to classical PDEs with a heavier emphasis on hyperbolic systems and constrained hyperbolic systems. The next topic is difference schemes. These schemes are the foundation for the adaptive methods. After the background material is covered, attention is focused on a simple patched based adaptive algorithm and its associated data structures for square grids and hyperbolic conservation laws. Embellishments include curvilinear meshes, embedded boundary and overset meshes. Next, several strategies for parallel implementations are examined. The remainder of the notes contains descriptions of elliptic solutions on the mesh hierarchy, elliptically constrained flow solution methods and elliptically constrained flow solution methods with diffusion.
Error estimation and adaptive mesh refinement for parallel analysis of shell structures
NASA Technical Reports Server (NTRS)
Keating, Scott C.; Felippa, Carlos A.; Park, K. C.
1994-01-01
The formulation and application of element-level, element-independent error indicators is investigated. This research culminates in the development of an error indicator formulation which is derived based on the projection of element deformation onto the intrinsic element displacement modes. The qualifier 'element-level' means that no information from adjacent elements is used for error estimation. This property is ideally suited for obtaining error values and driving adaptive mesh refinements on parallel computers where access to neighboring elements residing on different processors may incur significant overhead. In addition such estimators are insensitive to the presence of physical interfaces and junctures. An error indicator qualifies as 'element-independent' when only visible quantities such as element stiffness and nodal displacements are used to quantify error. Error evaluation at the element level and element independence for the error indicator are highly desired properties for computing error in production-level finite element codes. Four element-level error indicators have been constructed. Two of the indicators are based on variational formulation of the element stiffness and are element-dependent. Their derivations are retained for developmental purposes. The second two indicators mimic and exceed the first two in performance but require no special formulation of the element stiffness mesh refinement which we demonstrate for two dimensional plane stress problems. The parallelizing of substructures and adaptive mesh refinement is discussed and the final error indicator using two-dimensional plane-stress and three-dimensional shell problems is demonstrated.
Adaptive Mesh Euler Equation Computation of Vortex Breakdown in Delta Wing Flow.
NASA Astrophysics Data System (ADS)
Modiano, David Laurence
A solution method for the three-dimensional Euler equations is formulated and implemented. The solver uses an unstructured mesh of tetrahedral cells and performs adaptive refinement by mesh-point embedding to increase mesh resolution in regions of interesting flow features. The fourth-difference artificial dissipation is increased to a higher order of accuracy using the method of Holmes and Connell. A new method of temporal integration is developed to accelerate the explicit computation of unsteady flows. The solver is applied to the solution of the flow around a sharp edged delta wing, with emphasis on the behavior of the leading edge vortex above the leeside of the wing at high angle of attack, under which conditions the vortex suffers from vortex breakdown. Large deviations in entropy, which indicate vortical regions of the flow, specify the region in which adaptation is performed. Adaptive flow calculations are performed at ten different angles of attack, at seven of which vortex breakdown occurs. The aerodynamic normal force coefficients show excellent agreement with wind tunnel data measured by Jarrah, which demonstrates the importance of adaptation in obtaining an accurate solution. The pitching moment coefficient and the location of vortex breakdown are compared with experimental data measured by Hummel and Srinivasan, with which fairly good agreement is seen in cases in which the location of breakdown is over the wing. A series of unsteady calculations involving a pitching delta wing were performed. The use of the acceleration technique is validated. A hysteresis in the normal force is observed, as in experiments, and a lag in the breakdown position is demonstrated. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617 -253-5668; Fax 617-253-1690.).
Biomechanical analyses of prosthetic mesh repair in a hiatal hernia model.
Alizai, Patrick Hamid; Schmid, Sofie; Otto, Jens; Klink, Christian Daniel; Roeth, Anjali; Nolting, Jochen; Neumann, Ulf Peter; Klinge, Uwe
2014-10-01
Recurrence rate of hiatal hernia can be reduced with prosthetic mesh repair; however, type and shape of the mesh are still a matter of controversy. The purpose of this study was to investigate the biomechanical properties of four conventional meshes: pure polypropylene mesh (PP-P), polypropylene/poliglecaprone mesh (PP-U), polyvinylidenefluoride/polypropylene mesh (PVDF-I), and pure polyvinylidenefluoride mesh (PVDF-S). Meshes were tested either in warp direction (parallel to production direction) or perpendicular to the warp direction. A Zwick testing machine was used to measure elasticity and effective porosity of the textile probes. Stretching of the meshes in warp direction required forces that were up to 85-fold higher than the same elongation in perpendicular direction. Stretch stress led to loss of effective porosity in most meshes, except for PVDF-S. Biomechanical impact of the mesh was additionally evaluated in a hiatal hernia model. The different meshes were used either as rectangular patches or as circular meshes. Circular meshes led to a significant reinforcement of the hiatus, largely unaffected by the orientation of the warp fibers. In contrast, rectangular meshes provided a significant reinforcement only when warp fibers ran perpendicular to the crura. Anisotropic elasticity of prosthetic meshes should therefore be considered in hiatal closure with rectangular patches.
Signal Delay in Leaky RC Mesh Models for Bipolar Interconnect,
1985-10-01
Mesh Networks," IEEE Trans. Circuits and Systems, vol. CAS-32, no. 5, pp. 507-510, May 1985. (3] Desoer , Charles A., and Ernest S. Kuh, Basic Circuit ...is appropriate for * modelling interconnect in digital bipolar circuits . This paper is intended to serve as a tutorial as well as a research report...class of networks that is appropriate for modelling interconnect in digital bipolar circuits . This paper is intended *" to serve as a tutorial as well
Adaptive-mesh-based algorithm for fluorescence molecular tomography using an analytical solution
NASA Astrophysics Data System (ADS)
Wang, Daifa; Song, Xiaolei; Bai, Jing
2007-07-01
Fluorescence molecular tomography (FMT) has become an important method for in-vivo imaging of small animals. It has been widely used for tumor genesis, cancer detection, metastasis, drug discovery, and gene therapy. In this study, an algorithm for FMT is proposed to obtain accurate and fast reconstruction by combining an adaptive mesh refinement technique and an analytical solution of diffusion equation. Numerical studies have been performed on a parallel plate FMT system with matching fluid. The reconstructions obtained show that the algorithm is efficient in computation time, and they also maintain image quality.
Compact integration factor methods for complex domains and adaptive mesh refinement.
Liu, Xinfeng; Nie, Qing
2010-08-10
Implicit integration factor (IIF) method, a class of efficient semi-implicit temporal scheme, was introduced recently for stiff reaction-diffusion equations. To reduce cost of IIF, compact implicit integration factor (cIIF) method was later developed for efficient storage and calculation of exponential matrices associated with the diffusion operators in two and three spatial dimensions for Cartesian coordinates with regular meshes. Unlike IIF, cIIF cannot be directly extended to other curvilinear coordinates, such as polar and spherical coordinate, due to the compact representation for the diffusion terms in cIIF. In this paper, we present a method to generalize cIIF for other curvilinear coordinates through examples of polar and spherical coordinates. The new cIIF method in polar and spherical coordinates has similar computational efficiency and stability properties as the cIIF in Cartesian coordinate. In addition, we present a method for integrating cIIF with adaptive mesh refinement (AMR) to take advantage of the excellent stability condition for cIIF. Because the second order cIIF is unconditionally stable, it allows large time steps for AMR, unlike a typical explicit temporal scheme whose time step is severely restricted by the smallest mesh size in the entire spatial domain. Finally, we apply those methods to simulating a cell signaling system described by a system of stiff reaction-diffusion equations in both two and three spatial dimensions using AMR, curvilinear and Cartesian coordinates. Excellent performance of the new methods is observed.
Sensitivity and initialisation of a moving mesh ice sheet model
NASA Astrophysics Data System (ADS)
Bonan, Bertrand; Baines, Michael J.; Nichols, Nancy K.; Partridge, Dale
2014-05-01
One of the most important issues in marine ice sheet modelling is tracking accurately the evolution of the snout and the grounding line. Here we present a moving mesh method that is well-suited to tracking moving phenomena accurately. We study the behaviour of the method for a flow line version using the Shallow Ice Approximation. The solution procedure uses the conservation of mass fractions to define a deformation velocity that generates movement of the support of the ice. The ice thickness is recovered using the conservation principle. In order to initialise the moving mesh model, we apply advanced inverse data assimilation techniques to the system. We develop particularly an Ensemble Kalman Filter (EnKF) approach in this context. EnKF is an efficient Monte-Carlo method based on Gaussian assumptions. Contrary to variational methods, it does not require the development of the adjoint of the model. The data assimilation procedure treats both the mesh point positions and the ice sheet thickness as unknown state variables and updates both of these at each assimilation step. The advantage of the ensemble approach is that it enables the sensitivity of the system to be understood and, more importantly, provides information on the correlations between the variables, in particular between the grid and the ice thickness. We demonstrate the success of the technique for noisy, infrequent, partial measurements of ice thickness, both with and without noisy measurements of the terminus position. The moving mesh and EnKF methods can be extended to coupled models of grounded ice sheets and floating ice shelves. The covariances between the states of the system at the grounding line derived by the method will provide valuable information on the coupled system. The techniques presented here can also be extended to include the estimation of parameters, such as the basal sliding coefficient and the bedrock topography, and to two dimensional ice sheets.
NASA Astrophysics Data System (ADS)
Bajc, Iztok; Hecht, Frédéric; Žumer, Slobodan
2016-09-01
This paper presents a 3D mesh adaptivity strategy on unstructured tetrahedral meshes by a posteriori error estimates based on metrics derived from the Hessian of a solution. The study is made on the case of a nonlinear finite element minimization scheme for the Landau-de Gennes free energy functional of nematic liquid crystals. Newton's iteration for tensor fields is employed with steepest descent method possibly stepping in. Aspects relating the driving of mesh adaptivity within the nonlinear scheme are considered. The algorithmic performance is found to depend on at least two factors: when to trigger each single mesh adaptation, and the precision of the correlated remeshing. Each factor is represented by a parameter, with its values possibly varying for every new mesh adaptation. We empirically show that the time of the overall algorithm convergence can vary considerably when different sequences of parameters are used, thus posing a question about optimality. The extensive testings and debugging done within this work on the simulation of systems of nematic colloids substantially contributed to the upgrade of an open source finite element-oriented programming language to its 3D meshing possibilities, as also to an outer 3D remeshing module.
De Colle, Fabio; Ramirez-Ruiz, Enrico; Granot, Jonathan; Lopez-Camara, Diego
2012-02-20
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in gamma-ray burst sources. The SRHD equations are solved using finite-volume conservative solvers, with second-order interpolation in space and time. The correct implementation of the algorithms is verified by one-dimensional (1D) and multi-dimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with {rho}{proportional_to}r{sup -k}, bridging between the relativistic and Newtonian phases (which are described by the Blandford-McKee and Sedov-Taylor self-similar solutions, respectively), as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to nonrelativistic speeds in one dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, together with the scaling of position, Lorentz factor, and the shock velocity as a function of time and shock radius, is explained here using a simple analytical model based on energy conservation. The method used for calculating the afterglow radiation by post-processing the results of the simulations is described in detail. The light curves computed using the results of 1D numerical simulations during the relativistic stage correctly reproduce those calculated assuming the self-similar Blandford-McKee solution for the evolution of the flow. The jet dynamics from our 2D simulations and the resulting afterglow light curves, including the jet break, are in good agreement with those presented in previous works. Finally, we show how the details of the dynamics critically depend on properly resolving the structure of the
NASA Astrophysics Data System (ADS)
De Colle, Fabio; Granot, Jonathan; López-Cámara, Diego; Ramirez-Ruiz, Enrico
2012-02-01
We report on the development of Mezcal-SRHD, a new adaptive mesh refinement, special relativistic hydrodynamics (SRHD) code, developed with the aim of studying the highly relativistic flows in gamma-ray burst sources. The SRHD equations are solved using finite-volume conservative solvers, with second-order interpolation in space and time. The correct implementation of the algorithms is verified by one-dimensional (1D) and multi-dimensional tests. The code is then applied to study the propagation of 1D spherical impulsive blast waves expanding in a stratified medium with ρvpropr -k , bridging between the relativistic and Newtonian phases (which are described by the Blandford-McKee and Sedov-Taylor self-similar solutions, respectively), as well as to a two-dimensional (2D) cylindrically symmetric impulsive jet propagating in a constant density medium. It is shown that the deceleration to nonrelativistic speeds in one dimension occurs on scales significantly larger than the Sedov length. This transition is further delayed with respect to the Sedov length as the degree of stratification of the ambient medium is increased. This result, together with the scaling of position, Lorentz factor, and the shock velocity as a function of time and shock radius, is explained here using a simple analytical model based on energy conservation. The method used for calculating the afterglow radiation by post-processing the results of the simulations is described in detail. The light curves computed using the results of 1D numerical simulations during the relativistic stage correctly reproduce those calculated assuming the self-similar Blandford-McKee solution for the evolution of the flow. The jet dynamics from our 2D simulations and the resulting afterglow light curves, including the jet break, are in good agreement with those presented in previous works. Finally, we show how the details of the dynamics critically depend on properly resolving the structure of the relativistic flow.
NASA Astrophysics Data System (ADS)
Foks, Nathan Leon
The interpretation of geophysical data plays an important role in the analysis of potential field data in resource exploration industries. Two categories of interpretation techniques are discussed in this thesis; boundary detection and geophysical inversion. Fault or boundary detection is a method to interpret the locations of subsurface boundaries from measured data, while inversion is a computationally intensive method that provides 3D information about subsurface structure. My research focuses on these two aspects of interpretation techniques. First, I develop a method to aid in the interpretation of faults and boundaries from magnetic data. These processes are traditionally carried out using raster grid and image processing techniques. Instead, I use unstructured meshes of triangular facets that can extract inferred boundaries using mesh edges. Next, to address the computational issues of geophysical inversion, I develop an approach to reduce the number of data in a data set. The approach selects the data points according to a user specified proxy for its signal content. The approach is performed in the data domain and requires no modification to existing inversion codes. This technique adds to the existing suite of compressive inversion algorithms. Finally, I develop an algorithm to invert gravity data for an interfacing surface using an unstructured mesh of triangular facets. A pertinent property of unstructured meshes is their flexibility at representing oblique, or arbitrarily oriented structures. This flexibility makes unstructured meshes an ideal candidate for geometry based interface inversions. The approaches I have developed provide a suite of algorithms geared towards large-scale interpretation of potential field data, by using an unstructured representation of both the data and model parameters.
A DAFT DL_POLY distributed memory adaptation of the Smoothed Particle Mesh Ewald method
NASA Astrophysics Data System (ADS)
Bush, I. J.; Todorov, I. T.; Smith, W.
2006-09-01
The Smoothed Particle Mesh Ewald method [U. Essmann, L. Perera, M.L. Berkowtz, T. Darden, H. Lee, L.G. Pedersen, J. Chem. Phys. 103 (1995) 8577] for calculating long ranged forces in molecular simulation has been adapted for the parallel molecular dynamics code DL_POLY_3 [I.T. Todorov, W. Smith, Philos. Trans. Roy. Soc. London 362 (2004) 1835], making use of a novel 3D Fast Fourier Transform (DAFT) [I.J. Bush, The Daresbury Advanced Fourier transform, Daresbury Laboratory, 1999] that perfectly matches the Domain Decomposition (DD) parallelisation strategy [W. Smith, Comput. Phys. Comm. 62 (1991) 229; M.R.S. Pinches, D. Tildesley, W. Smith, Mol. Sim. 6 (1991) 51; D. Rapaport, Comput. Phys. Comm. 62 (1991) 217] of the DL_POLY_3 code. In this article we describe software adaptations undertaken to import this functionality and provide a review of its performance.
AMRSim: an object-oriented performance simulator for parallel adaptive mesh refinement
Miller, B; Philip, B; Quinlan, D; Wissink, A
2001-01-08
Adaptive mesh refinement is complicated by both the algorithms and the dynamic nature of the computations. In parallel the complexity of getting good performance is dependent upon the architecture and the application. Most attempts to address the complexity of AMR have lead to the development of library solutions, most have developed object-oriented libraries or frameworks. All attempts to date have made numerous and sometimes conflicting assumptions which make the evaluation of performance of AMR across different applications and architectures difficult or impracticable. The evaluation of different approaches can alternatively be accomplished through simulation of the different AMR processes. In this paper we outline our research work to simulate the processing of adaptive mesh refinement grids using a distributed array class library (P++). This paper presents a combined analytic and empirical approach, since details of the algorithms can be readily predicted (separated into specific phases), while the performance associated with the dynamic behavior must be studied empirically. The result, AMRSim, provides a simple way to develop bounds on the expected performance of AMR calculations subject to constraints given by the algorithms, frameworks, and architecture.
Parallel grid library with adaptive mesh refinement for development of highly scalable simulations
NASA Astrophysics Data System (ADS)
Honkonen, I.; von Alfthan, S.; Sandroos, A.; Janhunen, P.; Palmroth, M.
2012-04-01
As the single CPU core performance is saturating while the number of cores in the fastest supercomputers increases exponentially, the parallel performance of simulations on distributed memory machines is crucial. At the same time, utilizing efficiently the large number of available cores presents a challenge, especially in simulations with run-time adaptive mesh refinement. We have developed a generic grid library (dccrg) aimed at finite volume simulations that is easy to use and scales well up to tens of thousands of cores. The grid has several attractive features: It 1) allows an arbitrary C++ class or structure to be used as cell data; 2) provides a simple interface for adaptive mesh refinement during a simulation; 3) encapsulates the details of MPI communication when updating the data of neighboring cells between processes; and 4) provides a simple interface to run-time load balancing, e.g. domain decomposition, through the Zoltan library. Dccrg is freely available for anyone to use, study and modify under the GNU Lesser General Public License v3. We will present the implementation of dccrg, simple and advanced usage examples and scalability results on various supercomputers and problems.
GAMER: A GRAPHIC PROCESSING UNIT ACCELERATED ADAPTIVE-MESH-REFINEMENT CODE FOR ASTROPHYSICS
Schive, H.-Y.; Tsai, Y.-C.; Chiueh Tzihong
2010-02-01
We present the newly developed code, GPU-accelerated Adaptive-MEsh-Refinement code (GAMER), which adopts a novel approach in improving the performance of adaptive-mesh-refinement (AMR) astrophysical simulations by a large factor with the use of the graphic processing unit (GPU). The AMR implementation is based on a hierarchy of grid patches with an oct-tree data structure. We adopt a three-dimensional relaxing total variation diminishing scheme for the hydrodynamic solver and a multi-level relaxation scheme for the Poisson solver. Both solvers have been implemented in GPU, by which hundreds of patches can be advanced in parallel. The computational overhead associated with the data transfer between the CPU and GPU is carefully reduced by utilizing the capability of asynchronous memory copies in GPU, and the computing time of the ghost-zone values for each patch is diminished by overlapping it with the GPU computations. We demonstrate the accuracy of the code by performing several standard test problems in astrophysics. GAMER is a parallel code that can be run in a multi-GPU cluster system. We measure the performance of the code by performing purely baryonic cosmological simulations in different hardware implementations, in which detailed timing analyses provide comparison between the computations with and without GPU(s) acceleration. Maximum speed-up factors of 12.19 and 10.47 are demonstrated using one GPU with 4096{sup 3} effective resolution and 16 GPUs with 8192{sup 3} effective resolution, respectively.
NASA Astrophysics Data System (ADS)
Yang, Sheng; Kuo, C.-C. Jay
2002-07-01
An Internet-based interactive walkthrough virtual environment is presented in this work to facilitate interactive streaming and browsing of 3D graphic models across the Internet. The models are compressed by the view-dependent progressive mesh compression algorithm to enable the decorrelation of partitions and finer granularity. Following the fundamental framework of mesh representation, an interactive protocol based on the real time streaming protocol (RTSP) is developed to enhance the interaction between the server and the client. Finally, the data of the virtual world is re-organized and transmitted according to the viewer's requests. Experimental results demonstrate that the proposed algorithm reduces the required transmission bandwidth, and provides an acceptable visual quality even at low bit rates.
Multi-dimensional upwind fluctuation splitting scheme with mesh adaption for hypersonic viscous flow
NASA Astrophysics Data System (ADS)
Wood, William Alfred, III
production is shown relative to DMFDSFV. Remarkably the fluctuation splitting scheme shows grid converged skin friction coefficients with only five points in the boundary layer for this case. A viscous Mach 17.6 (perfect gas) cylinder case demonstrates solution monotonicity and heat transfer capability with the fluctuation splitting scheme. While fluctuation splitting is recommended over DMFDSFV, the difference in performance between the schemes is not so great as to obsolete DMFDSFV. The second half of the dissertation develops a local, compact, anisotropic unstructured mesh adaption scheme in conjunction with the multi-dimensional upwind solver, exhibiting a characteristic alignment behavior for scalar problems. This alignment behavior stands in contrast to the curvature clustering nature of the local, anisotropic unstructured adaption strategy based upon a posteriori error estimation that is used for comparison. The characteristic alignment is most pronounced for linear advection, with reduced improvement seen for the more complex non-linear advection and advection-diffusion cases. The adaption strategy is extended to the two-dimensional and axisymmetric Navier-Stokes equations of motion through the concept of fluctuation minimization. The system test case for the adaption strategy is a sting mounted capsule at Mach-10 wind tunnel conditions, considered in both two-dimensional and axisymmetric configurations. For this complex flowfield the adaption results are disappointing since feature alignment does not emerge from the local operations. Aggressive adaption is shown to result in a loss of robustness for the solver, particularly in the bow shock/stagnation point interaction region. Reducing the adaption strength maintains solution robustness but fails to produce significant improvement in the surface heat transfer predictions.
3D model retrieval method based on mesh segmentation
NASA Astrophysics Data System (ADS)
Gan, Yuanchao; Tang, Yan; Zhang, Qingchen
2012-04-01
In the process of feature description and extraction, current 3D model retrieval algorithms focus on the global features of 3D models but ignore the combination of global and local features of the model. For this reason, they show less effective performance to the models with similar global shape and different local shape. This paper proposes a novel algorithm for 3D model retrieval based on mesh segmentation. The key idea is to exact the structure feature and the local shape feature of 3D models, and then to compares the similarities of the two characteristics and the total similarity between the models. A system that realizes this approach was built and tested on a database of 200 objects and achieves expected results. The results show that the proposed algorithm improves the precision and the recall rate effectively.
Scalable and Adaptive Streaming of 3D Mesh to Heterogeneous Devices
NASA Astrophysics Data System (ADS)
Abderrahim, Zeineb; Bouhlel, Mohamed Salim
2016-12-01
This article comprises a presentation of a web platform for the diffusion and visualization of 3D compressed data on the web. Indeed, the major goal of this work resides in the proposal of the transfer adaptation of the three-dimensional data to resources (network bandwidth, the type of visualization terminals, display resolution, user's preferences...). Also, it is an attempt to provide an effective consultation adapted to the user's request (preferences, levels of the requested detail, etc.). Such a platform can adapt the levels of detail to the change in the bandwidth and the rendering time when loading the mesh at the client level. In addition, the levels of detail are adapted to the distance between the object and the camera. These features are able to minimize the latency time and to make the real time interaction possible. The experiences as well as the comparison with the existing solutions show auspicious results in terms of latency, scalability and the quality of the experience offered to the users.
Unstructured mesh algorithms for aerodynamic calculations
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1992-01-01
The use of unstructured mesh techniques for solving complex aerodynamic flows is discussed. The principle advantages of unstructured mesh strategies, as they relate to complex geometries, adaptive meshing capabilities, and parallel processing are emphasized. The various aspects required for the efficient and accurate solution of aerodynamic flows are addressed. These include mesh generation, mesh adaptivity, solution algorithms, convergence acceleration, and turbulence modeling. Computations of viscous turbulent two-dimensional flows and inviscid three-dimensional flows about complex configurations are demonstrated. Remaining obstacles and directions for future research are also outlined.
Simulations of recoiling black holes: adaptive mesh refinement and radiative transfer
NASA Astrophysics Data System (ADS)
Meliani, Zakaria; Mizuno, Yosuke; Olivares, Hector; Porth, Oliver; Rezzolla, Luciano; Younsi, Ziri
2017-01-01
Context. In many astrophysical phenomena, and especially in those that involve the high-energy regimes that always accompany the astronomical phenomenology of black holes and neutron stars, physical conditions that are achieved are extreme in terms of speeds, temperatures, and gravitational fields. In such relativistic regimes, numerical calculations are the only tool to accurately model the dynamics of the flows and the transport of radiation in the accreting matter. Aims: We here continue our effort of modelling the behaviour of matter when it orbits or is accreted onto a generic black hole by developing a new numerical code that employs advanced techniques geared towards solving the equations of general-relativistic hydrodynamics. Methods: More specifically, the new code employs a number of high-resolution shock-capturing Riemann solvers and reconstruction algorithms, exploiting the enhanced accuracy and the reduced computational cost of adaptive mesh-refinement (AMR) techniques. In addition, the code makes use of sophisticated ray-tracing libraries that, coupled with general-relativistic radiation-transfer calculations, allow us to accurately compute the electromagnetic emissions from such accretion flows. Results: We validate the new code by presenting an extensive series of stationary accretion flows either in spherical or axial symmetry that are performed either in two or three spatial dimensions. In addition, we consider the highly nonlinear scenario of a recoiling black hole produced in the merger of a supermassive black-hole binary interacting with the surrounding circumbinary disc. In this way, we can present for the first time ray-traced images of the shocked fluid and the light curve resulting from consistent general-relativistic radiation-transport calculations from this process. Conclusions: The work presented here lays the ground for the development of a generic computational infrastructure employing AMR techniques to accurately and self
Ying, Wenjun; Henriquez, Craig S
2015-01-01
A both space and time adaptive algorithm is presented for simulating electrical wave propagation in the Purkinje system of the heart. The equations governing the distribution of electric potential over the system are solved in time with the method of lines. At each timestep, by an operator splitting technique, the space-dependent but linear diffusion part and the nonlinear but space-independent reactions part in the partial differential equations are integrated separately with implicit schemes, which have better stability and allow larger timesteps than explicit ones. The linear diffusion equation on each edge of the system is spatially discretized with the continuous piecewise linear finite element method. The adaptive algorithm can automatically recognize when and where the electrical wave starts to leave or enter the computational domain due to external current/voltage stimulation, self-excitation, or local change of membrane properties. Numerical examples demonstrating efficiency and accuracy of the adaptive algorithm are presented.
Ying, Wenjun; Henriquez, Craig S.
2015-01-01
A both space and time adaptive algorithm is presented for simulating electrical wave propagation in the Purkinje system of the heart. The equations governing the distribution of electric potential over the system are solved in time with the method of lines. At each timestep, by an operator splitting technique, the space-dependent but linear diffusion part and the nonlinear but space-independent reactions part in the partial differential equations are integrated separately with implicit schemes, which have better stability and allow larger timesteps than explicit ones. The linear diffusion equation on each edge of the system is spatially discretized with the continuous piecewise linear finite element method. The adaptive algorithm can automatically recognize when and where the electrical wave starts to leave or enter the computational domain due to external current/voltage stimulation, self-excitation, or local change of membrane properties. Numerical examples demonstrating efficiency and accuracy of the adaptive algorithm are presented. PMID:26581455
Fakhari, Abbas; Lee, Taehun
2014-03-01
An adaptive-mesh-refinement (AMR) algorithm for the finite-difference lattice Boltzmann method (FDLBM) is presented in this study. The idea behind the proposed AMR is to remove the need for a tree-type data structure. Instead, pointer attributes are used to determine the neighbors of a certain block via appropriate adjustment of its children identifications. As a result, the memory and time required for tree traversal are completely eliminated, leaving us with an efficient algorithm that is easier to implement and use on parallel machines. To allow different mesh sizes at separate parts of the computational domain, the Eulerian formulation of the streaming process is invoked. As a result, there is no need for rescaling the distribution functions or using a temporal interpolation at the fine-coarse grid boundaries. The accuracy and efficiency of the proposed FDLBM AMR are extensively assessed by investigating a variety of vorticity-dominated flow fields, including Taylor-Green vortex flow, lid-driven cavity flow, thin shear layer flow, and the flow past a square cylinder.
NASA Astrophysics Data System (ADS)
Fakhari, Abbas; Lee, Taehun
2014-03-01
An adaptive-mesh-refinement (AMR) algorithm for the finite-difference lattice Boltzmann method (FDLBM) is presented in this study. The idea behind the proposed AMR is to remove the need for a tree-type data structure. Instead, pointer attributes are used to determine the neighbors of a certain block via appropriate adjustment of its children identifications. As a result, the memory and time required for tree traversal are completely eliminated, leaving us with an efficient algorithm that is easier to implement and use on parallel machines. To allow different mesh sizes at separate parts of the computational domain, the Eulerian formulation of the streaming process is invoked. As a result, there is no need for rescaling the distribution functions or using a temporal interpolation at the fine-coarse grid boundaries. The accuracy and efficiency of the proposed FDLBM AMR are extensively assessed by investigating a variety of vorticity-dominated flow fields, including Taylor-Green vortex flow, lid-driven cavity flow, thin shear layer flow, and the flow past a square cylinder.
Adaptive Urban Dispersion Integrated Model
Wissink, A; Chand, K; Kosovic, B; Chan, S; Berger, M; Chow, F K
2005-11-03
Numerical simulations represent a unique predictive tool for understanding the three-dimensional flow fields and associated concentration distributions from contaminant releases in complex urban settings (Britter and Hanna 2003). Utilization of the most accurate urban models, based on fully three-dimensional computational fluid dynamics (CFD) that solve the Navier-Stokes equations with incorporated turbulence models, presents many challenges. We address two in this work; first, a fast but accurate way to incorporate the complex urban terrain, buildings, and other structures to enforce proper boundary conditions in the flow solution; second, ways to achieve a level of computational efficiency that allows the models to be run in an automated fashion such that they may be used for emergency response and event reconstruction applications. We have developed a new integrated urban dispersion modeling capability based on FEM3MP (Gresho and Chan 1998, Chan and Stevens 2000), a CFD model from Lawrence Livermore National Lab. The integrated capability incorporates fast embedded boundary mesh generation for geometrically complex problems and full three-dimensional Cartesian adaptive mesh refinement (AMR). Parallel AMR and embedded boundary gridding support are provided through the SAMRAI library (Wissink et al. 2001, Hornung and Kohn 2002). Embedded boundary mesh generation has been demonstrated to be an automatic, fast, and efficient approach for problem setup. It has been used for a variety of geometrically complex applications, including urban applications (Pullen et al. 2005). The key technology we introduce in this work is the application of AMR, which allows the application of high-resolution modeling to certain important features, such as individual buildings and high-resolution terrain (including important vegetative and land-use features). It also allows the urban scale model to be readily interfaced with coarser resolution meso or regional scale models. This talk
Galaxy Mergers with Adaptive Mesh Refinement: Star Formation and Hot Gas Outflow
Kim, Ji-hoon; Wise, John H.; Abel, Tom; /KIPAC, Menlo Park /Stanford U., Phys. Dept.
2011-06-22
In hierarchical structure formation, merging of galaxies is frequent and known to dramatically affect their properties. To comprehend these interactions high-resolution simulations are indispensable because of the nonlinear coupling between pc and Mpc scales. To this end, we present the first adaptive mesh refinement (AMR) simulation of two merging, low mass, initially gas-rich galaxies (1.8 x 10{sup 10} M{sub {circle_dot}} each), including star formation and feedback. With galaxies resolved by {approx} 2 x 10{sup 7} total computational elements, we achieve unprecedented resolution of the multiphase interstellar medium, finding a widespread starburst in the merging galaxies via shock-induced star formation. The high dynamic range of AMR also allows us to follow the interplay between the galaxies and their embedding medium depicting how galactic outflows and a hot metal-rich halo form. These results demonstrate that AMR provides a powerful tool in understanding interacting galaxies.
NASA Astrophysics Data System (ADS)
Chilton, Sven; Colella, Phillip
2010-11-01
Adaptive mesh refinement (AMR) is an efficient technique for solving systems of partial differential equations numerically. The underlying algorithm determines where and when a base spatial and temporal grid must be resolved further in order to achieve the desired precision and accuracy in the numerical solution. However, propagating wave solutions prove problematic for AMR. In systems with low degrees of dissipation (e.g. the Maxwell-Vlasov system) a wave traveling from a finely resolved region into a coarsely resolved region encounters a numerical impedance mismatch, resulting in spurious reflections off of the coarse-fine grid boundary. These reflected waves then become trapped inside the fine region. Here, we present a scheme for damping these spurious reflections. We demonstrate its application to the scalar wave equation and an implementation for Maxwell's Equations. We also discuss a possible extension to the Maxwell-Vlasov system.
MPI parallelization of full PIC simulation code with Adaptive Mesh Refinement
NASA Astrophysics Data System (ADS)
Matsui, Tatsuki; Nunami, Masanori; Usui, Hideyuki; Moritaka, Toseo
2010-11-01
A new parallelization technique developed for PIC method with adaptive mesh refinement (AMR) is introduced. In AMR technique, the complicated cell arrangements are organized and managed as interconnected pointers with multiple resolution levels, forming a fully threaded tree structure as a whole. In order to retain this tree structure distributed over multiple processes, remote memory access, an extended feature of MPI2 standards, is employed. Another important feature of the present simulation technique is the domain decomposition according to the modified Morton ordering. This algorithm can group up the equal number of particle calculation loops, which allows for the better load balance. Using this advanced simulation code, preliminary results for basic physical problems are exhibited for the validity check, together with the benchmarks to test the performance and the scalability.
THE PLUTO CODE FOR ADAPTIVE MESH COMPUTATIONS IN ASTROPHYSICAL FLUID DYNAMICS
Mignone, A.; Tzeferacos, P.; Zanni, C.; Bodo, G.; Van Straalen, B.; Colella, P.
2012-01-01
We present a description of the adaptive mesh refinement (AMR) implementation of the PLUTO code for solving the equations of classical and special relativistic magnetohydrodynamics (MHD and RMHD). The current release exploits, in addition to the static grid version of the code, the distributed infrastructure of the CHOMBO library for multidimensional parallel computations over block-structured, adaptively refined grids. We employ a conservative finite-volume approach where primary flow quantities are discretized at the cell center in a dimensionally unsplit fashion using the Corner Transport Upwind method. Time stepping relies on a characteristic tracing step where piecewise parabolic method, weighted essentially non-oscillatory, or slope-limited linear interpolation schemes can be handily adopted. A characteristic decomposition-free version of the scheme is also illustrated. The solenoidal condition of the magnetic field is enforced by augmenting the equations with a generalized Lagrange multiplier providing propagation and damping of divergence errors through a mixed hyperbolic/parabolic explicit cleaning step. Among the novel features, we describe an extension of the scheme to include non-ideal dissipative processes, such as viscosity, resistivity, and anisotropic thermal conduction without operator splitting. Finally, we illustrate an efficient treatment of point-local, potentially stiff source terms over hierarchical nested grids by taking advantage of the adaptivity in time. Several multidimensional benchmarks and applications to problems of astrophysical relevance assess the potentiality of the AMR version of PLUTO in resolving flow features separated by large spatial and temporal disparities.
Stabilized Conservative Level Set Method with Adaptive Wavelet-based Mesh Refinement
NASA Astrophysics Data System (ADS)
Shervani-Tabar, Navid; Vasilyev, Oleg V.
2016-11-01
This paper addresses one of the main challenges of the conservative level set method, namely the ill-conditioned behavior of the normal vector away from the interface. An alternative formulation for reconstruction of the interface is proposed. Unlike the commonly used methods which rely on the unit normal vector, Stabilized Conservative Level Set (SCLS) uses a modified renormalization vector with diminishing magnitude away from the interface. With the new formulation, in the vicinity of the interface the reinitialization procedure utilizes compressive flux and diffusive terms only in the normal direction to the interface, thus, preserving the conservative level set properties, while away from the interfaces the directional diffusion mechanism automatically switches to homogeneous diffusion. The proposed formulation is robust and general. It is especially well suited for use with adaptive mesh refinement (AMR) approaches due to need for a finer resolution in the vicinity of the interface in comparison with the rest of the domain. All of the results were obtained using the Adaptive Wavelet Collocation Method, a general AMR-type method, which utilizes wavelet decomposition to adapt on steep gradients in the solution while retaining a predetermined order of accuracy.
Mesh refinement for uncertainty quantification through model reduction
Li, Jing Stinis, Panos
2015-01-01
We present a novel way of deciding when and where to refine a mesh in probability space in order to facilitate uncertainty quantification in the presence of discontinuities in random space. A discontinuity in random space makes the application of generalized polynomial chaos expansion techniques prohibitively expensive. The reason is that for discontinuous problems, the expansion converges very slowly. An alternative to using higher terms in the expansion is to divide the random space in smaller elements where a lower degree polynomial is adequate to describe the randomness. In general, the partition of the random space is a dynamic process since some areas of the random space, particularly around the discontinuity, need more refinement than others as time evolves. In the current work we propose a way to decide when and where to refine the random space mesh based on the use of a reduced model. The idea is that a good reduced model can monitor accurately, within a random space element, the cascade of activity to higher degree terms in the chaos expansion. In turn, this facilitates the efficient allocation of computational sources to the areas of random space where they are more needed. For the Kraichnan–Orszag system, the prototypical system to study discontinuities in random space, we present theoretical results which show why the proposed method is sound and numerical results which corroborate the theory.
MeshGUI: A Mesh Generation and Editing Toolset for the ADCIRC Model
2008-02-08
40 APPENDIX II: Creating Input Data for Meshcreate…………………….. 42 a. Bathymetric Data: Extraction from NRL-DBDB2 Database...The first data file is a boundary file that describes the geographic limits of the mesh to be created . Often the boundary file represents a coastline...constructed to include overland areas as long as supporting topographic information is available. The boundary file can be created using an auxiliary
Consistent properties reconstruction on adaptive Cartesian meshes for complex fluids computations
Xia, Guoping . E-mail: xiag@purdue.edu; Li, Ding; Merkle, Charles L.
2007-07-01
An efficient reconstruction procedure for evaluating the constitutive properties of a complex fluid from general or specialized thermodynamic databases is presented. Properties and their pertinent derivatives are evaluated by means of an adaptive Cartesian mesh in the thermodynamic plane that provides user-specified accuracy over any selected domain. The Cartesian grid produces a binary tree data structure whose search efficiency is competitive with that for an equally spaced table or with simple equations of state such as a perfect gas. Reconstruction is accomplished on a triangular subdivision of the 2D Cartesian mesh that ensures function continuity across cell boundaries in equally and unequally spaced portions of the table to C {sup 0}, C {sup 1} or C {sup 2} levels. The C {sup 0} and C {sup 1} reconstructions fit the equation of state and enthalpy relations separately, while the C {sup 2} reconstruction fits the Helmholtz or Gibbs function enabling EOS/enthalpy consistency also. All three reconstruction levels appear effective for CFD solutions obtained to date. The efficiency of the method is demonstrated through storage and data retrieval examples for air, water and carbon dioxide. The time required for property evaluations is approximately two orders of magnitude faster with the reconstruction procedure than with the complete thermodynamic equations resulting in estimated 3D CFD savings of from 30 to 60. Storage requirements are modest for today's computers, with the C {sup 1} method requiring slightly less storage than those for the C {sup 0} and C {sup 2} reconstructions when the same accuracy is specified. Sample fluid dynamic calculations based upon the procedure show that the C {sup 1} and C {sup 2} methods are approximately a factor of two slower than the C {sup 0} method but that the reconstruction procedure enables arbitrary fluid CFD calculations that are as efficient as those for a perfect gas or an incompressible fluid for all three accuracy
3-D grid refinement using the University of Michigan adaptive mesh library for a pure advective test
NASA Astrophysics Data System (ADS)
Oehmke, R.; Vandenberg, D.; Andronova, N.; Penner, J.; Stout, Q.; Zubov, V.; Jablonowski, C.
2008-05-01
The numerical representation of the partial differential equations (PDE) for high resolution atmospheric dynamical and physical features requires division of the atmospheric volume into a set of 3D grids, each of which has a not quite rectangular form. Each location on the grid contains multiple data that together represent the state of Earth's atmosphere. For successful numerical integration of the PDEs the size of each grid box is used to define the Courant-Friedrichs-Levi criterion in setting the time step. 3D adaptive representations of a sphere are needed to represent the evolution of clouds. In this paper we present the University of Michigan adaptive mesh library - a library that supports the production of parallel codes with use of adaptation on a sphere. The library manages the block-structured data layout, handles ghost cell updates among neighboring blocks and splits blocks as refinements occur. The library has several modules that provide a layer of abstraction for adaptive refinement: blocks, which contain individual cells of user data; shells — the global geometry for the problem, including a sphere, reduced sphere, and now a 3D sphere; a load balancer for placement of blocks onto processors; and a communication support layer which encapsulates all data movement. Users provide data manipulation functions for performing interpolation of user data when refining blocks. We rigorously test the library using refinement of the modeled vertical transport of a tracer with prescribed atmospheric sources and sinks. It is both a 2 and a 3D test, and bridges the performance of the model's dynamics and physics needed for inclusion of cloud formation.
NASA Astrophysics Data System (ADS)
Huang, Rongzong; Wu, Huiying
2016-06-01
A total enthalpy-based lattice Boltzmann (LB) method with adaptive mesh refinement (AMR) is developed in this paper to efficiently simulate solid-liquid phase change problem where variables vary significantly near the phase interface and thus finer grid is required. For the total enthalpy-based LB method, the velocity field is solved by an incompressible LB model with multiple-relaxation-time (MRT) collision scheme, and the temperature field is solved by a total enthalpy-based MRT LB model with the phase interface effects considered and the deviation term eliminated. With a kinetic assumption that the density distribution function for solid phase is at equilibrium state, a volumetric LB scheme is proposed to accurately realize the nonslip velocity condition on the diffusive phase interface and in the solid phase. As compared with the previous schemes, this scheme can avoid nonphysical flow in the solid phase. As for the AMR approach, it is developed based on multiblock grids. An indicator function is introduced to control the adaptive generation of multiblock grids, which can guarantee the existence of overlap area between adjacent blocks for information exchange. Since MRT collision schemes are used, the information exchange is directly carried out in the moment space. Numerical tests are firstly performed to validate the strict satisfaction of the nonslip velocity condition, and then melting problems in a square cavity with different Prandtl numbers and Rayleigh numbers are simulated, which demonstrate that the present method can handle solid-liquid phase change problem with high efficiency and accuracy.
NASA Technical Reports Server (NTRS)
Steger, J. L.; Dougherty, F. C.; Benek, J. A.
1983-01-01
A mesh system composed of multiple overset body-conforming grids is described for adapting finite-difference procedures to complex aircraft configurations. In this so-called 'chimera mesh,' a major grid is generated about a main component of the configuration and overset minor grids are used to resolve all other features. Methods for connecting overset multiple grids and modifications of flow-simulation algorithms are discussed. Computational tests in two dimensions indicate that the use of multiple overset grids can simplify the task of grid generation without an adverse effect on flow-field algorithms and computer code complexity.
NASA Astrophysics Data System (ADS)
Salinas, Pablo; Pavlidis, Dimitrios; Percival, James; Adam, Alexander; Xie, Zhihua; Pain, Christopher; Jackson, Matthew
2015-11-01
We present a new, high-order, control-volume-finite-element (CVFE) method with discontinuous representation for pressure and velocity to simulate multiphase flow in heterogeneous porous media. Time is discretized using an adaptive, fully implicit method. Heterogeneous geologic features are represented as volumes bounded by surfaces. Our approach conserves mass and does not require the use of CVs that span domain boundaries. Computational efficiency is increased by use of dynamic mesh optimization. We demonstrate that the approach, amongst other features, accurately preserves sharp saturation changes associated with high aspect ratio geologic domains, allowing efficient simulation of flow in highly heterogeneous models. Moreover, accurate solutions are obtained at lower cost than an equivalent fine, fixed mesh and conventional CVFE methods. The use of implicit time integration allows the method to efficiently converge using highly anisotropic meshes without having to reduce the time-step. The work is significant for two key reasons. First, it resolves a long-standing problem associated with the use of classical CVFE methods. Second, it reduces computational cost/increases solution accuracy through the use of dynamic mesh optimization and time-stepping with large Courant number. Funding for Dr P. Salinas from ExxonMobil is gratefully acknowledged.
Dynamic implicit 3D adaptive mesh refinement for non-equilibrium radiation diffusion
B. Philip; Z. Wang; M.A. Berrill; M. Birke; M. Pernice
2014-04-01
The time dependent non-equilibrium radiation diffusion equations are important for solving the transport of energy through radiation in optically thick regimes and find applications in several fields including astrophysics and inertial confinement fusion. The associated initial boundary value problems that are encountered often exhibit a wide range of scales in space and time and are extremely challenging to solve. To efficiently and accurately simulate these systems we describe our research on combining techniques that will also find use more broadly for long term time integration of nonlinear multi-physics systems: implicit time integration for efficient long term time integration of stiff multi-physics systems, local control theory based step size control to minimize the required global number of time steps while controlling accuracy, dynamic 3D adaptive mesh refinement (AMR) to minimize memory and computational costs, Jacobian Free Newton–Krylov methods on AMR grids for efficient nonlinear solution, and optimal multilevel preconditioner components that provide level independent solver convergence.
Advances in Rotor Performance and Turbulent Wake Simulation Using DES and Adaptive Mesh Refinement
NASA Technical Reports Server (NTRS)
Chaderjian, Neal M.
2012-01-01
Time-dependent Navier-Stokes simulations have been carried out for a rigid V22 rotor in hover, and a flexible UH-60A rotor in forward flight. Emphasis is placed on understanding and characterizing the effects of high-order spatial differencing, grid resolution, and Spalart-Allmaras (SA) detached eddy simulation (DES) in predicting the rotor figure of merit (FM) and resolving the turbulent rotor wake. The FM was accurately predicted within experimental error using SA-DES. Moreover, a new adaptive mesh refinement (AMR) procedure revealed a complex and more realistic turbulent rotor wake, including the formation of turbulent structures resembling vortical worms. Time-dependent flow visualization played a crucial role in understanding the physical mechanisms involved in these complex viscous flows. The predicted vortex core growth with wake age was in good agreement with experiment. High-resolution wakes for the UH-60A in forward flight exhibited complex turbulent interactions and turbulent worms, similar to the V22. The normal force and pitching moment coefficients were in good agreement with flight-test data.
Dynamic implicit 3D adaptive mesh refinement for non-equilibrium radiation diffusion
NASA Astrophysics Data System (ADS)
Philip, B.; Wang, Z.; Berrill, M. A.; Birke, M.; Pernice, M.
2014-04-01
The time dependent non-equilibrium radiation diffusion equations are important for solving the transport of energy through radiation in optically thick regimes and find applications in several fields including astrophysics and inertial confinement fusion. The associated initial boundary value problems that are encountered often exhibit a wide range of scales in space and time and are extremely challenging to solve. To efficiently and accurately simulate these systems we describe our research on combining techniques that will also find use more broadly for long term time integration of nonlinear multi-physics systems: implicit time integration for efficient long term time integration of stiff multi-physics systems, local control theory based step size control to minimize the required global number of time steps while controlling accuracy, dynamic 3D adaptive mesh refinement (AMR) to minimize memory and computational costs, Jacobian Free Newton-Krylov methods on AMR grids for efficient nonlinear solution, and optimal multilevel preconditioner components that provide level independent solver convergence.
Output-Based Adaptive Meshing Applied to Space Launch System Booster Separation Analysis
NASA Technical Reports Server (NTRS)
Dalle, Derek J.; Rogers, Stuart E.
2015-01-01
This paper presents details of Computational Fluid Dynamic (CFD) simulations of the Space Launch System during solid-rocket booster separation using the Cart3D inviscid code with comparisons to Overflow viscous CFD results and a wind tunnel test performed at NASA Langley Research Center's Unitary PlanWind Tunnel. The Space Launch System (SLS) launch vehicle includes two solid-rocket boosters that burn out before the primary core stage and thus must be discarded during the ascent trajectory. The main challenges for creating an aerodynamic database for this separation event are the large number of basis variables (including orientation of the core, relative position and orientation of the boosters, and rocket thrust levels) and the complex flow caused by the booster separation motors. The solid-rocket boosters are modified from their form when used with the Space Shuttle Launch Vehicle, which has a rich flight history. However, the differences between the SLS core and the Space Shuttle External Tank result in the boosters separating with much narrower clearances, and so reducing aerodynamic uncertainty is necessary to clear the integrated system for flight. This paper discusses an approach that has been developed to analyze about 6000 wind tunnel simulations and 5000 flight vehicle simulations using Cart3D in adaptive-meshing mode. In addition, a discussion is presented of Overflow viscous CFD runs used for uncertainty quantification. Finally, the article presents lessons learned and improvements that will be implemented in future separation databases.
Collins, David C.; Norman, Michael L.; Padoan, Paolo; Xu Hao
2011-04-10
In this work, we present the mass and magnetic distributions found in a recent adaptive mesh refinement magnetohydrodynamic simulation of supersonic, super-Alfvenic, self-gravitating turbulence. Power-law tails are found in both mass density and magnetic field probability density functions, with P({rho}) {proportional_to} {rho}{sup -1.6} and P(B) {proportional_to} B{sup -2.7}. A power-law relationship is also found between magnetic field strength and density, with B {proportional_to} {rho}{sup 0.5}, throughout the collapsing gas. The mass distribution of gravitationally bound cores is shown to be in excellent agreement with recent observation of prestellar cores. The mass-to-flux distribution of cores is also found to be in excellent agreement with recent Zeeman splitting measurements. We also compare the relationship between velocity dispersion and density to the same cores, and find an increasing relationship between the two, with {sigma} {proportional_to} n{sup 0.25}, also in agreement with the observations. We then estimate the potential effects of ambipolar diffusion in our cores and find that due to the weakness of the magnetic field in our simulation, the inclusion of ambipolar diffusion in our simulation will not cause significant alterations of the flow dynamics.
GPU accelerated cell-based adaptive mesh refinement on unstructured quadrilateral grid
NASA Astrophysics Data System (ADS)
Luo, Xisheng; Wang, Luying; Ran, Wei; Qin, Fenghua
2016-10-01
A GPU accelerated inviscid flow solver is developed on an unstructured quadrilateral grid in the present work. For the first time, the cell-based adaptive mesh refinement (AMR) is fully implemented on GPU for the unstructured quadrilateral grid, which greatly reduces the frequency of data exchange between GPU and CPU. Specifically, the AMR is processed with atomic operations to parallelize list operations, and null memory recycling is realized to improve the efficiency of memory utilization. It is found that results obtained by GPUs agree very well with the exact or experimental results in literature. An acceleration ratio of 4 is obtained between the parallel code running on the old GPU GT9800 and the serial code running on E3-1230 V2. With the optimization of configuring a larger L1 cache and adopting Shared Memory based atomic operations on the newer GPU C2050, an acceleration ratio of 20 is achieved. The parallelized cell-based AMR processes have achieved 2x speedup on GT9800 and 18x on Tesla C2050, which demonstrates that parallel running of the cell-based AMR method on GPU is feasible and efficient. Our results also indicate that the new development of GPU architecture benefits the fluid dynamics computing significantly.
Relativistic Flows Using Spatial And Temporal Adaptive Structured Mesh Refinement. I. Hydrodynamics
Wang, Peng; Abel, Tom; Zhang, Weiqun; /KIPAC, Menlo Park
2007-04-02
Astrophysical relativistic flow problems require high resolution three-dimensional numerical simulations. In this paper, we describe a new parallel three-dimensional code for simulations of special relativistic hydrodynamics (SRHD) using both spatially and temporally structured adaptive mesh refinement (AMR). We used method of lines to discrete SRHD equations spatially and used a total variation diminishing (TVD) Runge-Kutta scheme for time integration. For spatial reconstruction, we have implemented piecewise linear method (PLM), piecewise parabolic method (PPM), third order convex essentially non-oscillatory (CENO) and third and fifth order weighted essentially non-oscillatory (WENO) schemes. Flux is computed using either direct flux reconstruction or approximate Riemann solvers including HLL, modified Marquina flux, local Lax-Friedrichs flux formulas and HLLC. The AMR part of the code is built on top of the cosmological Eulerian AMR code enzo, which uses the Berger-Colella AMR algorithm and is parallel with dynamical load balancing using the widely available Message Passing Interface library. We discuss the coupling of the AMR framework with the relativistic solvers and show its performance on eleven test problems.
Adaptive Mesh Refinement for a High-Symmetry Singular Euler Flow
NASA Astrophysics Data System (ADS)
Germaschewski, K.; Bhattacharjee, A.; Grauer, R.
2002-11-01
Starting from a highly symmetric initial condition motivated by the work of Kida [J. Phys. Soc Jpn. 54, 2132 (1995)] and Boratav and Pelz [Phys. Fluids 6, 2757 (1994)], we use the technique of block-structured adaptive mesh refinement (AMR) to numerically investigate the development of a self-similar singular solution to the incompressible Euler equations. The scheme, previously used by Grauer et al [Phys. Rev. Lett. 84, 4850 (1998)], is particularly well suited to follow the development of singular structures as it allows for effective resolutions far beyond those accessible using fixed grid algorithms. A self-similar collapse is observed in the simulation, where the maximum vorticity blows up as 1/(t_crit-t). Ng and Bhattacharjee [Phys Rev E 54, 1530 (1996)] have presented a sufficient condition for a finite-time singularity in this highly symmetric flow involving the fourth-order spatial derivative of the pressure at and near the origin. We test this sufficient condition and investigate the evolution of the spatial range over which this condition holds in our numerical results. We also demonstrate numerically that this singularity is unstable because in a full simulation that does not build in the symmetries of the initial condition, small perturbations introduced by AMR lead to nonsymmetric evolution of the vortices.
3D Boltzmann Simulation of the Io's Plasma Environment with Adaptive Mesh and Particle Refinement
NASA Astrophysics Data System (ADS)
Lipatov, A. S.; Combi, M. R.
2002-12-01
The global dynamics of the ionized and neutral components in the environment of Io plays an important role in the interaction of Jupiter's corotating magnetospheric plasma with Io [Combi et al., 2002; 1998; Kabin et al., 2001]. The stationary simulation of this problem was done in the MHD [Combi et al., 1998; Linker et al, 1998; Kabin et al., 2001] and the electrodynamic [Saur et al., 1999] approaches. In this report, we develop a method of kinetic ion-neutral simulation, which is based on a multiscale adaptive mesh, particle and algorithm refinement. This method employs the fluid description for electrons whereas for ions the drift-kinetic and particle approaches are used. This method takes into account charge-exchange and photoionization processes. The first results of such simulation of the dynamics of ions in the Io's environment are discussed in this report. ~ M R Combi et al., J. Geophys. Res., 103, 9071, 1998. M R Combi, T I Gombosi, K Kabin, Atmospheres in the Solar System: Comparative\\ Aeronomy. Geophys. Monograph Series, 130, 151, 2002. K Kabin et al., Planetary and Space Sci., 49, 337, 2001. J A Linker et al., J. Geophys. Res., 103(E9), 19867, 1998. J Saur et al., J. Geophys. Res., 104, 25105, 1999.
NASA Astrophysics Data System (ADS)
Wang, Cheng; Dong, XinZhuang; Shu, Chi-Wang
2015-10-01
For numerical simulation of detonation, computational cost using uniform meshes is large due to the vast separation in both time and space scales. Adaptive mesh refinement (AMR) is advantageous for problems with vastly different scales. This paper aims to propose an AMR method with high order accuracy for numerical investigation of multi-dimensional detonation. A well-designed AMR method based on finite difference weighted essentially non-oscillatory (WENO) scheme, named as AMR&WENO is proposed. A new cell-based data structure is used to organize the adaptive meshes. The new data structure makes it possible for cells to communicate with each other quickly and easily. In order to develop an AMR method with high order accuracy, high order prolongations in both space and time are utilized in the data prolongation procedure. Based on the message passing interface (MPI) platform, we have developed a workload balancing parallel AMR&WENO code using the Hilbert space-filling curve algorithm. Our numerical experiments with detonation simulations indicate that the AMR&WENO is accurate and has a high resolution. Moreover, we evaluate and compare the performance of the uniform mesh WENO scheme and the parallel AMR&WENO method. The comparison results provide us further insight into the high performance of the parallel AMR&WENO method.
Fibrin sealing versus stapling of hernia meshes in an onlay model in the rat.
Petter-Puchner, Alexander H; Fortelny, R; Mittermayr, R; Ohlinger, W; Redl, H
2005-12-01
Incisional and inguinal hernia repair are among the most common procedures of general surgery. Mesh fixation by means of staples or sutures may lead to severe complications. The use of fibrin sealant (FS) has been suggested as alternative, but data on biocompatibility and adhesive strength of FS in combination with macroporous meshes is limited. Ventral hernia (n = 8 per group) was treated in rats in onlay technique with two types of meshes, fibrin sealed or stapled. TI-Mesh (TMxl) extralight and VYPROII (VPII) were tested 17 days post op. No failure in mechanical tests (tensile and burst strength) occurred in sealed or stapled meshes. Histology revealed equally good tissue integration and neovascularization in all groups. Fibrin sealant yields excellent fixation in experimental hernia repair. This rat model is suitable for testing meshes and fixation techniques.
Tool-assisted mesh generation based on a tissue-growth model.
Smirnov, A V
2003-07-01
An heuristic mesh generation technique is proposed that is based on the model of forced particle motion, an edgewise cell-splitting algorithm and a moving tool concept. The method differs from conventional mesh generators in that it uses outward growth of the mesh, in contrast to the inward growth used in traditional meshing techniques. The method does not require prior meshing and patching of two-dimensional (2D) boundary surfaces. Instead, it uses a pre-defined skeleton of one-dimensional segments, or an arbitrary tool motion in three-dimensional (3D) space. In this respect, the technique can be considered as a 3D extension of a 2D drawing tool and can find applications in virtual reality systems. The method also guarantees the smoothness of the outer boundary of the mesh at each step of mesh generation, which is not the case with traditional propagating-front methods. The approach is based on the model of tissue growth and is suitable for meshing complex networks of bifurcating branches commonly found in biological structures: blood vessels, lungs, neural networks, plants etc. The generated meshes were used in solving unsteady flow and particle transport problems in lungs.
NASA Astrophysics Data System (ADS)
Campa, Alessandro; Esposito, Giuseppe; Belli, Mauro
Cellular response to radiation is often modified by a previous delivery of a small "priming" dose: a smaller amount of damage, defined by the end point being investigated, is observed, and for this reason the effect is called adaptive response. An improved understanding of this effect is essential (as much as for the case of the bystander effect) for a reliable radiation risk assessment when low dose irradiations are involved. Experiments on adaptive response have shown that there are a number of factors that strongly influence the occurrence (and the level) of the adaptation. In particular, priming doses and dose rates have to fall in defined ranges; the same is true for the time interval between the delivery of the small priming dose and the irradiation with the main, larger, dose (called in this case challenging dose). Different hypotheses can be formulated on the main mechanism(s) determining the adaptive response: an increased efficiency of DNA repair, an increased level of antioxidant enzymes, an alteration of cell cycle progression, a chromatin conformation change. An experimental clearcut evidence going definitely in the direction of one of these explanations is not yet available. Modelling can be done at different levels. Simple models, relating the amount of damage, through elementary differential equations, to the dose and dose rate experienced by the cell, are relatively easy to handle, and they can be modified to account for the priming irradiation. However, this can hardly be of decisive help in the explanation of the mechanisms, since each parameter of these models often incorporates in an effective way several cellular processes related to the response to radiation. In this presentation we show our attempts to describe adaptive response with models that explicitly contain, as a dynamical variable, the inducible adaptive agent. At a price of a more difficult treatment, this approach is probably more prone to give support to the experimental studies
A Nonlinear Dynamic Model and Free Vibration Analysis of Deployable Mesh Reflectors
NASA Technical Reports Server (NTRS)
Shi, H.; Yang, B.; Thomson, M.; Fang, H.
2011-01-01
This paper presents a dynamic model of deployable mesh reflectors, in which geometric and material nonlinearities of such a space structure are fully described. Then, by linearization around an equilibrium configuration of the reflector structure, a linearized model is obtained. With this linearized model, the natural frequencies and mode shapes of a reflector can be computed. The nonlinear dynamic model of deployable mesh reflectors is verified by using commercial finite element software in numerical simulation. As shall be seen, the proposed nonlinear model is useful for shape (surface) control of deployable mesh reflectors under thermal loads.
Adaptive Finite Element Methods for Continuum Damage Modeling
NASA Technical Reports Server (NTRS)
Min, J. B.; Tworzydlo, W. W.; Xiques, K. E.
1995-01-01
The paper presents an application of adaptive finite element methods to the modeling of low-cycle continuum damage and life prediction of high-temperature components. The major objective is to provide automated and accurate modeling of damaged zones through adaptive mesh refinement and adaptive time-stepping methods. The damage modeling methodology is implemented in an usual way by embedding damage evolution in the transient nonlinear solution of elasto-viscoplastic deformation problems. This nonlinear boundary-value problem is discretized by adaptive finite element methods. The automated h-adaptive mesh refinements are driven by error indicators, based on selected principal variables in the problem (stresses, non-elastic strains, damage, etc.). In the time domain, adaptive time-stepping is used, combined with a predictor-corrector time marching algorithm. The time selection is controlled by required time accuracy. In order to take into account strong temperature dependency of material parameters, the nonlinear structural solution a coupled with thermal analyses (one-way coupling). Several test examples illustrate the importance and benefits of adaptive mesh refinements in accurate prediction of damage levels and failure time.
NASA Astrophysics Data System (ADS)
Phillips, Carolyn L.
2014-09-01
In a complex self-organizing system, small changes in the interactions between the system's components can result in different emergent macrostructures or macrobehavior. In chemical engineering and material science, such spontaneously self-assembling systems, using polymers, nanoscale or colloidal-scale particles, DNA, or other precursors, are an attractive way to create materials that are precisely engineered at a fine scale. Changes to the interactions can often be described by a set of parameters. Different contiguous regions in this parameter space correspond to different ordered states. Since these ordered states are emergent, often experiment, not analysis, is necessary to create a diagram of ordered states over the parameter space. By issuing queries to points in the parameter space (e.g., performing a computational or physical experiment), ordered states can be discovered and mapped. Queries can be costly in terms of resources or time, however. In general, one would like to learn the most information using the fewest queries. Here we introduce a learning heuristic for issuing queries to map and search a two-dimensional parameter space. Using a method inspired by adaptive mesh refinement, the heuristic iteratively issues batches of queries to be executed in parallel based on past information. By adjusting the search criteria, different types of searches (for example, a uniform search, exploring boundaries, sampling all regions equally) can be flexibly implemented. We show that this method will densely search the space, while preferentially targeting certain features. Using numerical examples, including a study simulating the self-assembly of complex crystals, we show how this heuristic can discover new regions and map boundaries more accurately than a uniformly distributed set of queries.
NASA Astrophysics Data System (ADS)
Henshaw, William D.; Schwendeman, Donald W.
2008-08-01
This paper describes an approach for the numerical solution of time-dependent partial differential equations in complex three-dimensional domains. The domains are represented by overlapping structured grids, and block-structured adaptive mesh refinement (AMR) is employed to locally increase the grid resolution. In addition, the numerical method is implemented on parallel distributed-memory computers using a domain-decomposition approach. The implementation is flexible so that each base grid within the overlapping grid structure and its associated refinement grids can be independently partitioned over a chosen set of processors. A modified bin-packing algorithm is used to specify the partition for each grid so that the computational work is evenly distributed amongst the processors. All components of the AMR algorithm such as error estimation, regridding, and interpolation are performed in parallel. The parallel time-stepping algorithm is illustrated for initial-boundary-value problems involving a linear advection-diffusion equation and the (nonlinear) reactive Euler equations. Numerical results are presented for both equations to demonstrate the accuracy and correctness of the parallel approach. Exact solutions of the advection-diffusion equation are constructed, and these are used to check the corresponding numerical solutions for a variety of tests involving different overlapping grids, different numbers of refinement levels and refinement ratios, and different numbers of processors. The problem of planar shock diffraction by a sphere is considered as an illustration of the numerical approach for the Euler equations, and a problem involving the initiation of a detonation from a hot spot in a T-shaped pipe is considered to demonstrate the numerical approach for the reactive case. For both problems, the accuracy of the numerical solutions is assessed quantitatively through an estimation of the errors from a grid convergence study. The parallel performance of the
Henshaw, W; Schwendeman, D
2007-11-15
This paper describes an approach for the numerical solution of time-dependent partial differential equations in complex three-dimensional domains. The domains are represented by overlapping structured grids, and block-structured adaptive mesh refinement (AMR) is employed to locally increase the grid resolution. In addition, the numerical method is implemented on parallel distributed-memory computers using a domain-decomposition approach. The implementation is flexible so that each base grid within the overlapping grid structure and its associated refinement grids can be independently partitioned over a chosen set of processors. A modified bin-packing algorithm is used to specify the partition for each grid so that the computational work is evenly distributed amongst the processors. All components of the AMR algorithm such as error estimation, regridding, and interpolation are performed in parallel. The parallel time-stepping algorithm is illustrated for initial-boundary-value problems involving a linear advection-diffusion equation and the (nonlinear) reactive Euler equations. Numerical results are presented for both equations to demonstrate the accuracy and correctness of the parallel approach. Exact solutions of the advection-diffusion equation are constructed, and these are used to check the corresponding numerical solutions for a variety of tests involving different overlapping grids, different numbers of refinement levels and refinement ratios, and different numbers of processors. The problem of planar shock diffraction by a sphere is considered as an illustration of the numerical approach for the Euler equations, and a problem involving the initiation of a detonation from a hot spot in a T-shaped pipe is considered to demonstrate the numerical approach for the reactive case. For both problems, the solutions are shown to be well resolved on the finest grid. The parallel performance of the approach is examined in detail for the shock diffraction problem.
Mathias, R T; Eisenberg, R S; Valdiosera, R
1977-01-01
This paper presents the construction, derivation, and test of a mesh model for the electrical properties of the transverse tubular system (T-system) in skeletal muscle. We model the irregular system of tubules as a random network of miniature transmission lines, using differential equations to describe the potential between the nodes and difference equations to describe the potential at the nodes. The solution to the equations can be accurately represented in several approximate forms with simple physical and graphical interpretations. All the parameters of the solution are specified by impedance and morphometric measurements. The effect of wide circumferential spacing between T-system openings is analyzed and the resulting restricted mesh model is shown to be approximated by a mesh with an access resistance. The continuous limit of the mesh model is shown to have the same form as the disk model of the T-system, but with a different expression for the tortuosity factor. The physical meaning of the tortuosity factor is examined, and a short derivation of the disk model is presented that gives results identical to the continuous limit of the mesh model. Both the mesh and restricted mesh models are compared with experimental data on the impedance of muscle fibers of the frog sartorius. The derived value for the resistivity of the lumen of the tubules is not too different from that of the bathing solution, the difference probably arising from the sensitivity of this value to errors in the morphometric measurements. Images FIGURE 4 PMID:831857
NASA Astrophysics Data System (ADS)
Zheng, H. W.; Shu, C.; Chew, Y. T.
2008-07-01
In this paper, an object-oriented and quadrilateral-mesh based solution adaptive algorithm for the simulation of compressible multi-fluid flows is presented. The HLLC scheme (Harten, Lax and van Leer approximate Riemann solver with the Contact wave restored) is extended to adaptively solve the compressible multi-fluid flows under complex geometry on unstructured mesh. It is also extended to the second-order of accuracy by using MUSCL extrapolation. The node, edge and cell are arranged in such an object-oriented manner that each of them inherits from a basic object. A home-made double link list is designed to manage these objects so that the inserting of new objects and removing of the existing objects (nodes, edges and cells) are independent of the number of objects and only of the complexity of O( 1). In addition, the cells with different levels are further stored in different lists. This avoids the recursive calculation of solution of mother (non-leaf) cells. Thus, high efficiency is obtained due to these features. Besides, as compared to other cell-edge adaptive methods, the separation of nodes would reduce the memory requirement of redundant nodes, especially in the cases where the level number is large or the space dimension is three. Five two-dimensional examples are used to examine its performance. These examples include vortex evolution problem, interface only problem under structured mesh and unstructured mesh, bubble explosion under the water, bubble-shock interaction, and shock-interface interaction inside the cylindrical vessel. Numerical results indicate that there is no oscillation of pressure and velocity across the interface and it is feasible to apply it to solve compressible multi-fluid flows with large density ratio (1000) and strong shock wave (the pressure ratio is 10,000) interaction with the interface.
Tuminaro, Raymond S.; Perego, Mauro; Tezaur, Irina Kalashnikova; Salinger, Andrew G.; Price, Stephen
2016-10-06
A multigrid method is proposed that combines ideas from matrix dependent multigrid for structured grids and algebraic multigrid for unstructured grids. It targets problems where a three-dimensional mesh can be viewed as an extrusion of a two-dimensional, unstructured mesh in a third dimension. Our motivation comes from the modeling of thin structures via finite elements and, more specifically, the modeling of ice sheets. Extruded meshes are relatively common for thin structures and often give rise to anisotropic problems when the thin direction mesh spacing is much smaller than the broad direction mesh spacing. Within our approach, the first few multigrid hierarchy levels are obtained by applying matrix dependent multigrid to semicoarsen in a structured thin direction fashion. After sufficient structured coarsening, the resulting mesh contains only a single layer corresponding to a two-dimensional, unstructured mesh. Algebraic multigrid can then be employed in a standard manner to create further coarse levels, as the anisotropic phenomena is no longer present in the single layer problem. The overall approach remains fully algebraic, with the minor exception that some additional information is needed to determine the extruded direction. Furthermore, this facilitates integration of the solver with a variety of different extruded mesh applications.
Tuminaro, Raymond S.; Perego, Mauro; Tezaur, Irina Kalashnikova; ...
2016-10-06
A multigrid method is proposed that combines ideas from matrix dependent multigrid for structured grids and algebraic multigrid for unstructured grids. It targets problems where a three-dimensional mesh can be viewed as an extrusion of a two-dimensional, unstructured mesh in a third dimension. Our motivation comes from the modeling of thin structures via finite elements and, more specifically, the modeling of ice sheets. Extruded meshes are relatively common for thin structures and often give rise to anisotropic problems when the thin direction mesh spacing is much smaller than the broad direction mesh spacing. Within our approach, the first few multigridmore » hierarchy levels are obtained by applying matrix dependent multigrid to semicoarsen in a structured thin direction fashion. After sufficient structured coarsening, the resulting mesh contains only a single layer corresponding to a two-dimensional, unstructured mesh. Algebraic multigrid can then be employed in a standard manner to create further coarse levels, as the anisotropic phenomena is no longer present in the single layer problem. The overall approach remains fully algebraic, with the minor exception that some additional information is needed to determine the extruded direction. Furthermore, this facilitates integration of the solver with a variety of different extruded mesh applications.« less
Development of improved analytical models for mesh reflector surfaces
NASA Technical Reports Server (NTRS)
Brand, J. C.; Kauffman, J. F.
1983-01-01
Several methods for computing the reflection coefficients from mesh surfaces are discussed. Some methods mentioned have severe limitations, and the spectral approach appears to be the most attractive alternative. In spite of some inherent problems, the solutions obtained with this method will offer not only the reflection coefficients but also the currents carried on the mesh. This would allow separation of the power lost to resistive terms from that due to transmission loss. Overall, good results and rapid convergence should be obtained from this method when proper care is applied.
NASA Astrophysics Data System (ADS)
Moura, R. C.; Silva, A. F. C.; Bigarella, E. D. V.; Fazenda, A. L.; Ortega, M. A.
2016-08-01
This paper proposes two important improvements to shock-capturing strategies using a discontinuous Galerkin scheme, namely, accurate shock identification via finite-time Lyapunov exponent (FTLE) operators and efficient shock treatment through a point-implicit discretization of a PDE-based artificial viscosity technique. The advocated approach is based on the FTLE operator, originally developed in the context of dynamical systems theory to identify certain types of coherent structures in a flow. We propose the application of FTLEs in the detection of shock waves and demonstrate the operator's ability to identify strong and weak shocks equally well. The detection algorithm is coupled with a mesh refinement procedure and applied to transonic and supersonic flows. While the proposed strategy can be used potentially with any numerical method, a high-order discontinuous Galerkin solver is used in this study. In this context, two artificial viscosity approaches are employed to regularize the solution near shocks: an element-wise constant viscosity technique and a PDE-based smooth viscosity model. As the latter approach is more sophisticated and preferable for complex problems, a point-implicit discretization in time is proposed to reduce the extra stiffness introduced by the PDE-based technique, making it more competitive in terms of computational cost.
Spherical geodesic mesh generation
Fung, Jimmy; Kenamond, Mark Andrew; Burton, Donald E.; Shashkov, Mikhail Jurievich
2015-02-27
In ALE simulations with moving meshes, mesh topology has a direct influence on feature representation and code robustness. In three-dimensional simulations, modeling spherical volumes and features is particularly challenging for a hydrodynamics code. Calculations on traditional spherical meshes (such as spin meshes) often lead to errors and symmetry breaking. Although the underlying differencing scheme may be modified to rectify this, the differencing scheme may not be accessible. This work documents the use of spherical geodesic meshes to mitigate solution-mesh coupling. These meshes are generated notionally by connecting geodesic surface meshes to produce triangular-prismatic volume meshes. This mesh topology is fundamentally different from traditional mesh topologies and displays superior qualities such as topological symmetry. This work describes the geodesic mesh topology as well as motivating demonstrations with the FLAG hydrocode.
Karell, Mara A; Langstaff, Helen K; Halazonetis, Demetrios J; Minghetti, Caterina; Frelat, Mélanie; Kranioti, Elena F
2016-09-01
The commingling of human remains often hinders forensic/physical anthropologists during the identification process, as there are limited methods to accurately sort these remains. This study investigates a new method for pair-matching, a common individualization technique, which uses digital three-dimensional models of bone: mesh-to-mesh value comparison (MVC). The MVC method digitally compares the entire three-dimensional geometry of two bones at once to produce a single value to indicate their similarity. Two different versions of this method, one manual and the other automated, were created and then tested for how well they accurately pair-matched humeri. Each version was assessed using sensitivity and specificity. The manual mesh-to-mesh value comparison method was 100 % sensitive and 100 % specific. The automated mesh-to-mesh value comparison method was 95 % sensitive and 60 % specific. Our results indicate that the mesh-to-mesh value comparison method overall is a powerful new tool for accurately pair-matching commingled skeletal elements, although the automated version still needs improvement.
NASA Technical Reports Server (NTRS)
Steinthorsson, E.; Modiano, David; Colella, Phillip
1994-01-01
A methodology for accurate and efficient simulation of unsteady, compressible flows is presented. The cornerstones of the methodology are a special discretization of the Navier-Stokes equations on structured body-fitted grid systems and an efficient solution-adaptive mesh refinement technique for structured grids. The discretization employs an explicit multidimensional upwind scheme for the inviscid fluxes and an implicit treatment of the viscous terms. The mesh refinement technique is based on the AMR algorithm of Berger and Colella. In this approach, cells on each level of refinement are organized into a small number of topologically rectangular blocks, each containing several thousand cells. The small number of blocks leads to small overhead in managing data, while their size and regular topology means that a high degree of optimization can be achieved on computers with vector processors.
Greene, Patrick T.; Schofield, Samuel P.; Nourgaliev, Robert
2017-01-27
A new mesh smoothing method designed to cluster cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered fields, such as a volume fractionmore » or index function, is provided. Results show that the low-order level set works equally well as the actual level set for mesh smoothing. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Lastly, dynamic cases with moving interfaces show the new method is capable of maintaining a desired resolution near the interface with an acceptable number of relaxation iterations per time step, which demonstrates the method's potential to be used as a mesh relaxer for arbitrary Lagrangian Eulerian (ALE) methods.« less
NASA Astrophysics Data System (ADS)
Greene, Patrick T.; Schofield, Samuel P.; Nourgaliev, Robert
2017-04-01
A new mesh smoothing method designed to cluster cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered fields, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well as the actual level set for mesh smoothing. Meshes generated for a number of interface geometries are presented, including cases with multiple level sets. Dynamic cases with moving interfaces show the new method is capable of maintaining a desired resolution near the interface with an acceptable number of relaxation iterations per time step, which demonstrates the method's potential to be used as a mesh relaxer for arbitrary Lagrangian Eulerian (ALE) methods.
NASA Technical Reports Server (NTRS)
Turon, Albert; Davila, Carlos G.; Camanho, Pedro P.; Costa, Josep
2005-01-01
This paper presents a methodology to determine the parameters used in the simulation of delamination in composite materials using decohesion finite elements. A closed-form expression is developed to define the stiffness of the cohesive layer. A novel procedure that allows the use of coarser meshes of decohesion elements in large-scale computations is proposed. The procedure ensures that the energy dissipated by the fracture process is correctly computed. It is shown that coarse-meshed models defined using the approach proposed here yield the same results as the models with finer meshes normally used in the simulation of fracture processes.
Coarse mesh transport theory model for heterogeneous systems
NASA Astrophysics Data System (ADS)
Ilas, Danut
To improve fuel utilization, recent reactor cores have become substantially more heterogeneous. In these cores, use of variable fuel enrichments and strong absorbers lead to high neutron flux gradients, which may limit the accuracy (validity) of diffusion theory based methods. In fact, the diffusion equation itself may become a poor approximation of the Boltzmann equation, the exact equation that describes the neutron flux. Therefore, numerical methods to solve the transport equation efficiently over a large heterogeneous region (such as a reactor core) are very desirable in case where the diffusion approximation breaks down. Presently, the only methods capable of computing the power (flux) distributions very accurately throughout a large system such as a nuclear reactor core are the Monte-Carlo or the fine-mesh transport theory methods. Both these methods suffer from the long computational time which makes them useless for routine core calculations. Starting from a variational principle that admits trial functions that can be discontinuous at coarse mesh (assembly) interfaces, we propose a method to solve the transport equation on a spatial grid made up of meshes as large as the size of a fuel assembly. The variational principle is derived for the most general case, but further methods are developed for one-dimensional geometry with the angular variable treated by discrete ordinates. The method uses the finite element approach for the space variable with basis functions precomputed for each element to obtain an algebraic linear system of equations. The eigenvalue of this system is the multiplication constant and the eigenvector represents the incoming angular fluxes for each coarse mesh. The latter allows the reconstruction of the fine mesh solution (angular flux) throughout the domain of interest when used with the basis functions (surface Green's function) for each coarse mesh. The method requires no homogenization procedure that can be a serious source of
Lipnikov, Konstantin; Agouzal, Abdellatif; Vassilevski, Yuri
2009-01-01
We present a new technology for generating meshes minimizing the interpolation and discretization errors or their gradients. The key element of this methodology is construction of a space metric from edge-based error estimates. For a mesh with N{sub h} triangles, the error is proportional to N{sub h}{sup -1} and the gradient of error is proportional to N{sub h}{sup -1/2} which are optimal asymptotics. The methodology is verified with numerical experiments.
FEMHD: An adaptive finite element method for MHD and edge modelling
Strauss, H.R.
1995-07-01
This paper describes the code FEMHD, an adaptive finite element MHD code, which is applied in a number of different manners to model MHD behavior and edge plasma phenomena on a diverted tokamak. The code uses an unstructured triangular mesh in 2D and wedge shaped mesh elements in 3D. The code has been adapted to look at neutral and charged particle dynamics in the plasma scrape off region, and into a full MHD-particle code.
A solution-adaptive mesh algorithm for dynamic/static refinement of two and three dimensional grids
NASA Technical Reports Server (NTRS)
Benson, Rusty A.; Mcrae, D. S.
1991-01-01
An adaptive grid algorithm has been developed in two and three dimensions that can be used dynamically with a solver or as part of a grid refinement process. The algorithm employs a transformation from the Cartesian coordinate system to a general coordinate space, which is defined as a parallelepiped in three dimensions. A weighting function, independent for each coordinate direction, is developed that will provide the desired refinement criteria in regions of high solution gradient. The adaptation is performed in the general coordinate space and the new grid locations are returned to the Cartesian space via a simple, one-step inverse mapping. The algorithm for relocation of the mesh points in the parametric space is based on the center of mass for distributed weights. Dynamic solution-adaptive results are presented for laminar flows in two and three dimensions.
Animating a Human Body Mesh with Maya for Doppler Signature Computer Modeling
2009-06-01
February 2009. 3. Remcom Web page. http://www.remcom.com (accessed November 2006). 4. Autodesk Web page. http://www.autodesk.com/ maya (accessed June...Animating a Human Body Mesh with Maya for Doppler Signature Computer Modeling by Getachew Kirose ARL-TN-0351 June 2009...Animating a Human Body Mesh with Maya for Doppler Signature Computer Modeling Getachew Kirose Sensors and Electron Devices Directorate
NASA Astrophysics Data System (ADS)
Grauer, R.; Germaschewski, K.
The goal of this presentation is threefold. First, the role of singular structures like shocks, vortex tubes and current sheets for understanding intermittency in small scale turbulence is demonstrated. Secondly, in order to investigate the time evolution of singular structures, effective numerical techniques have to be applied, like block structured adaptive mesh refinement combined with recent advances in treating hyperbolic equations. And thirdly, the developed numerical techniques can perfectly be applied to the question of fast reconnection demonstrated by the example of compressible Hall-MHD including electron and ion inertia. 1 Why is it worth studying singular structures? The motivation for studying singular structures has several sources. In turbulent fluid and plasma flows the formation of nearly singular structures like shocks, vortex tubes or current sheets provide an effective mechanism to transport energy from large to small scales. In the last years it has become clear that the nature of the singular structures is a key feature of small scale intermittency. In a phenomenological way this is established in She-Leveque like models (She and Leveque, 1994; Grauer, Krug and Marliani, 1994; Politano and Pouquet, 1995; M¨uller and Biskamp, 2000), which are able to describe some of the scaling properties of high order structure functions. An additional source which highlights the importance of singular structures originates from studies of a toy model of turbulence, the so-called Burgers turbulence. The very left tail of the probability distribution of velocity increments can be calculated using the instanton approach (Balkovsky, Falkovich, Kolokolov and Lebedev, 1997). Here it is interesting to note that the main contribution in the relevant path integral stems from the the singular structures which are shocks in the burgers turbulence. From a mathematical point of view the question whether
Adaptive optimal quantization for 3D mesh representation in the spherical coordinate system
NASA Astrophysics Data System (ADS)
Ahn, Jeong-Hwan; Ho, Yo-Sung
1998-12-01
In recent days, applications using 3D models are increasing. Since the 3D model contains a huge amount of information, compression of the 3D model data is necessary for efficient storage or transmission. In this paper, we propose an adaptive encoding scheme to compress the geometry information of the 3D model. Using the Levinson-Durbin algorithm, the encoder first predicts vertex positions along a vertex spanning tree. After each prediction error is normalized, the prediction error vector of each vertex point is represented in the spherical coordinate system (r,(theta) ,(phi) ). Each r is then quantizes by an optimal uniform quantizer. A pair of each ((theta) ,(phi) ) is also successively encoded by partitioning the surface of the sphere according to the quantized value of r. The proposed scheme demonstrates improved coding efficiency by exploiting the statistical properties of r and ((theta) ,(phi) ).
NASA Astrophysics Data System (ADS)
Greene, Patrick; Schofield, Sam; Nourgaliev, Robert
2016-11-01
A new mesh smoothing method designed to cluster cells near a dynamically evolving interface is presented. The method is based on weighted condition number mesh relaxation with the weight function being computed from a level set representation of the interface. The weight function is expressed as a Taylor series based discontinuous Galerkin (DG) projection, which makes the computation of the derivatives of the weight function needed during the condition number optimization process a trivial matter. For cases when a level set is not available, a fast method for generating a low-order level set from discrete cell-centered fields, such as a volume fraction or index function, is provided. Results show that the low-order level set works equally well for the weight function as the actual level set. The method retains the excellent smoothing capabilities of condition number relaxation, while providing a method for clustering mesh cells near regions of interest. Dynamic cases for moving interfaces are presented to demonstrate the method's potential usefulness as a mesh relaxer for arbitrary Lagrangian Eulerian (ALE) methods. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
NASA Astrophysics Data System (ADS)
Sanzana, P.; Jankowfsky, S.; Branger, F.; Braud, I.; Vargas, X.; Hitschfeld, N.; Gironás, J.
2013-08-01
Distributed hydrological models rely on a spatial discretization composed of homogeneous units representing different areas within the catchment. Hydrological Response Units (HRUs) typically form the basis of such a discretization. HRUs are generally obtained by intersecting raster or vector layers of land uses, soil types, geology and sub-catchments. Polylines maps representing ditches and river drainage networks can also be used. However this overlapping may result in a mesh with numerical and topological problems not highly representative of the terrain. Thus, a pre-processing is needed to improve the mesh in order to avoid negative effects on the performance of the hydrological model. This paper proposes computer-assisted mesh generation tools to obtain a more regular and physically meaningful mesh of HRUs suitable for hydrologic modeling. We combined existing tools with newly developed scripts implemented in GRASS GIS. The developed scripts address the following problems: (1) high heterogeneity in Digital Elevation Model derived properties within the HRUs, (2) correction of concave polygons or polygons with holes inside, (3) segmentation of very large polygons, and (4) bad estimations of units' perimeter and distances among them. The improvement process was applied and tested using two small catchments in France. The improvement of the spatial discretization was further assessed by comparing the representation and arrangement of overland flow paths in the original and improved meshes. Overall, a more realistic physical representation was obtained with the improved meshes, which should enhance the computation of surface and sub-surface flows in a hydrologic model.
Klein, R.I. |; Bell, J.; Pember, R.; Kelleher, T.
1993-04-01
The authors present results for high resolution hydrodynamic calculations of the growth and development of instabilities in shock driven imploding spherical geometries in both 2D and 3D. They solve the Eulerian equations of hydrodynamics with a high order Godunov approach using local adaptive mesh refinement to study the temporal and spatial development of the turbulent mixing layer resulting from both Richtmyer Meshkov and Rayleigh Taylor instabilities. The use of a high resolution Eulerian discretization with adaptive mesh refinement permits them to study the detailed three-dimensional growth of multi-mode perturbations far into the non-linear regime for converging geometries. They discuss convergence properties of the simulations by calculating global properties of the flow. They discuss the time evolution of the turbulent mixing layer and compare its development to a simple theory for a turbulent mix model in spherical geometry based on Plesset`s equation. Their 3D calculations show that the constant found in the planar incompressible experiments of Read and Young`s may not be universal for converging compressible flow. They show the 3D time trace of transitional onset to a mixing state using the temporal evolution of volume rendered imaging. Their preliminary results suggest that the turbulent mixing layer loses memory of its initial perturbations for classical Richtmyer Meshkov and Rayleigh Taylor instabilities in spherically imploding shells. They discuss the time evolution of mixed volume fraction and the role of vorticity in converging 3D flows in enhancing the growth of a turbulent mixing layer.
NASA Astrophysics Data System (ADS)
Fang, F.; Zhang, T.; Pavlidis, D.; Pain, C. C.; Buchan, A. G.; Navon, I. M.
2014-10-01
A novel reduced order model (ROM) based on proper orthogonal decomposition (POD) has been developed for a finite-element (FE) adaptive mesh air pollution model. A quadratic expansion of the non-linear terms is employed to ensure the method remained efficient. This is the first time such an approach has been applied to air pollution LES turbulent simulation through three dimensional landscapes. The novelty of this work also includes POD's application within a FE-LES turbulence model that uses adaptive resolution. The accuracy of the reduced order model is assessed and validated for a range of 2D and 3D urban street canyon flow problems. By comparing the POD solutions against the fine detail solutions obtained from the full FE model it is shown that the accuracy is maintained, where fine details of the air flows are captured, whilst the computational requirements are reduced. In the examples presented below the size of the reduced order models is reduced by factors up to 2400 in comparison to the full FE model while the CPU time is reduced by up to 98% of that required by the full model.
Optimal sinusoidal modelling of gear mesh vibration signals for gear diagnosis and prognosis
NASA Astrophysics Data System (ADS)
Man, Zhihong; Wang, Wenyi; Khoo, Suiyang; Yin, Juliang
2012-11-01
In this paper, the synchronous signal average of gear mesh vibration signals is modelled with the multiple modulated sinusoidal representations. The signal model parameters are optimised against the measured signal averages by using the batch learning of the least squares technique. With the optimal signal model, all components of a gear mesh vibration signal, including the amplitude modulations, the phase modulations and the impulse vibration component induced by gear tooth cracking, are identified and analysed with insight of the gear tooth crack development and propagation. In particular, the energy distribution of the impulse vibration signal, extracted from the optimal signal model, provides sufficient information for monitoring and diagnosing the evolution of the tooth cracking process, leading to the prognosis of gear tooth cracking. The new methodologies for gear mesh signal modelling and the diagnosis of the gear tooth fault development and propagation are validated with a set of rig test data, which has shown excellent performance.
Cortical imaging on a head template: a simulation study using a resistor mesh model (RMM).
Chauveau, Nicolas; Franceries, Xavier; Aubry, Florent; Celsis, Pierre; Rigaud, Bernard
2008-09-01
The T1 head template model used in Statistical Parametric Mapping Version 2000 (SPM2), was segmented into five layers (scalp, skull, CSF, grey and white matter) and implemented in 2 mm voxels. We designed a resistor mesh model (RMM), based on the finite volume method (FVM) to simulate the electrical properties of this head model along the three axes for each voxel. Then, we introduced four dipoles of high eccentricity (about 0.8) in this RMM, separately and simultaneously, to compute the potentials for two sets of conductivities. We used the direct cortical imaging technique (CIT) to recover the simulated dipoles, using 60 or 107 electrodes and with or without addition of Gaussian white noise (GWN). The use of realistic conductivities gave better CIT results than standard conductivities, lowering the blurring effect on scalp potentials and displaying more accurate position areas when CIT was applied to single dipoles. Simultaneous dipoles were less accurately localized, but good qualitative and stable quantitative results were obtained up to 5% noise level for 107 electrodes and up to 10% noise level for 60 electrodes, showing that a compromise must be found to optimize both the number of electrodes and the noise level. With the RMM defined in 2 mm voxels, the standard 128-electrode cap and 5% noise appears to be the upper limit providing reliable source positions when direct CIT is used. The admittance matrix defining the RMM is easy to modify so as to adapt to different conductivities. The next step will be the adaptation of individual real head T2 images to the RMM template and the introduction of anisotropy using diffusion imaging (DI).
A nonhydrostatic unstructured-mesh soundproof model for simulation of internal gravity waves
NASA Astrophysics Data System (ADS)
Smolarkiewicz, Piotr; Szmelter, Joanna
2011-12-01
A semi-implicit edge-based unstructured-mesh model is developed that integrates nonhydrostatic soundproof equations, inclusive of anelastic and pseudo-incompressible systems of partial differential equations. The model builds on nonoscillatory forward-in-time MPDATA approach using finite-volume discretization and unstructured meshes with arbitrarily shaped cells. Implicit treatment of gravity waves benefits both accuracy and stability of the model. The unstructured-mesh solutions are compared to equivalent structured-grid results for intricate, multiscale internal-wave phenomenon of a non-Boussinesq amplification and breaking of deep stratospheric gravity waves. The departures of the anelastic and pseudoincompressible results are quantified in reference to a recent asymptotic theory [Achatz et al. 2010, J. Fluid Mech., 663, 120-147)].
Three-dimensional gravity modeling and focusing inversion using rectangular meshes.
Commer, M.
2011-03-01
Rectangular grid cells are commonly used for the geophysical modeling of gravity anomalies, owing to their flexibility in constructing complex models. The straightforward handling of cubic cells in gravity inversion algorithms allows for a flexible imposition of model regularization constraints, which are generally essential in the inversion of static potential field data. The first part of this paper provides a review of commonly used expressions for calculating the gravity of a right polygonal prism, both for gravity and gradiometry, where the formulas of Plouff and Forsberg are adapted. The formulas can be cast into general forms practical for implementation. In the second part, a weighting scheme for resolution enhancement at depth is presented. Modelling the earth using highly digitized meshes, depth weighting schemes are typically applied to the model objective functional, subject to minimizing the data misfit. The scheme proposed here involves a non-linear conjugate gradient inversion scheme with a weighting function applied to the non-linear conjugate gradient scheme's gradient vector of the objective functional. The low depth resolution due to the quick decay of the gravity kernel functions is counteracted by suppressing the search directions in the parameter space that would lead to near-surface concentrations of gravity anomalies. Further, a density parameter transformation function enabling the imposition of lower and upper bounding constraints is employed. Using synthetic data from models of varying complexity and a field data set, it is demonstrated that, given an adequate depth weighting function, the gravity inversion in the transform space can recover geologically meaningful models requiring a minimum of prior information and user interaction.
NASA Astrophysics Data System (ADS)
Angelidis, Dionysios; Sotiropoulos, Fotis
2015-11-01
The geometrical details of wind turbines determine the structure of the turbulence in the near and far wake and should be taken in account when performing high fidelity calculations. Multi-resolution simulations coupled with an immersed boundary method constitutes a powerful framework for high-fidelity calculations past wind farms located over complex terrains. We develop a 3D Immersed-Boundary Adaptive Mesh Refinement flow solver (IB-AMR) which enables turbine-resolving LES of wind turbines. The idea of using a hybrid staggered/non-staggered grid layout adopted in the Curvilinear Immersed Boundary Method (CURVIB) has been successfully incorporated on unstructured meshes and the fractional step method has been employed. The overall performance and robustness of the second order accurate, parallel, unstructured solver is evaluated by comparing the numerical simulations against conforming grid calculations and experimental measurements of laminar and turbulent flows over complex geometries. We also present turbine-resolving multi-scale LES considering all the details affecting the induced flow field; including the geometry of the tower, the nacelle and especially the rotor blades of a wind tunnel scale turbine. This material is based upon work supported by the Department of Energy under Award Number DE-EE0005482 and the Sandia National Laboratories.
NASA Astrophysics Data System (ADS)
Fakhari, Abbas; Bolster, Diogo
2016-11-01
A three-dimensional (3D) adaptive mesh refinement (AMR) algorithm on structured Cartesian grids is developed, and supplemented by a mesoscopic multiphase-flow solver based on state-of-the-art lattice Boltzmann methods (LBM). Using this in-house AMR-LBM routine, we present fully 3D simulations of partial coalescence of a liquid drop with an initially flat interface at small Ohnesorge and Bond numbers. Qualitatively, our numerical simulations are in excellent agreement with experimental observations. Partial coalescence cascades are successfully observed at very small Ohnesorge numbers (Oh 10-4). The fact that the partial coalescence is absent in similar 2D simulations suggests that the Rayleigh-Plateau instability may be the principle driving mechanism responsible for this phenomenon.
Lopez-Camara, D.; Lazzati, Davide; Morsony, Brian J.; Begelman, Mitchell C.
2013-04-10
We present the results of special relativistic, adaptive mesh refinement, 3D simulations of gamma-ray burst jets expanding inside a realistic stellar progenitor. Our simulations confirm that relativistic jets can propagate and break out of the progenitor star while remaining relativistic. This result is independent of the resolution, even though the amount of turbulence and variability observed in the simulations is greater at higher resolutions. We find that the propagation of the jet head inside the progenitor star is slightly faster in 3D simulations compared to 2D ones at the same resolution. This behavior seems to be due to the fact that the jet head in 3D simulations can wobble around the jet axis, finding the spot of least resistance to proceed. Most of the average jet properties, such as density, pressure, and Lorentz factor, are only marginally affected by the dimensionality of the simulations and therefore results from 2D simulations can be considered reliable.
NASA Astrophysics Data System (ADS)
Zhang, A.; Guo, Z.; Xiong, S.-M.
2017-03-01
Eutectic pattern transition under an externally imposed temperature gradient was studied using the phase field method coupled with a novel parallel adaptive-mesh-refinement (Para-AMR) algorithm. Numerical tests revealed that the Para-AMR algorithm could improve the computational efficiency by two orders of magnitude and thus made it possible to perform large-scale simulations without any compromising accuracy. Results showed that the direction of the temperature gradient played a crucial role in determining the eutectic patterns during solidification, which agreed well with experimental observations. In particular, the presence of the transverse temperature gradient could tilt the eutectic patterns, and in 3D simulations, the eutectic microstructure would alter from lamellar to rod-like and/or from rod-like to dumbbell-shaped. Furthermore, under a radial temperature gradient, the eutectic would evolve from a dumbbell-shaped or clover-shaped pattern to an isolated rod-like pattern.
N-body simulations for f(R) gravity using a self-adaptive particle-mesh code
Zhao Gongbo; Koyama, Kazuya; Li Baojiu
2011-02-15
We perform high-resolution N-body simulations for f(R) gravity based on a self-adaptive particle-mesh code MLAPM. The chameleon mechanism that recovers general relativity on small scales is fully taken into account by self-consistently solving the nonlinear equation for the scalar field. We independently confirm the previous simulation results, including the matter power spectrum, halo mass function, and density profiles, obtained by Oyaizu et al.[Phys. Rev. D 78, 123524 (2008)] and Schmidt et al.[Phys. Rev. D 79, 083518 (2009)], and extend the resolution up to k{approx}20 h/Mpc for the measurement of the matter power spectrum. Based on our simulation results, we discuss how the chameleon mechanism affects the clustering of dark matter and halos on full nonlinear scales.
Masterlark, Timothy; Lu, Zhong; Rykhus, Russell P.
2006-01-01
nterferometric synthetic aperture radar (InSAR) imagery documents the consistent subsidence, during the interval 1992–1999, of a pyroclastic flow deposit (PFD) emplaced during the 1986 eruption of Augustine Volcano, Alaska. We construct finite element models (FEMs) that simulate thermoelastic contraction of the PFD to account for the observed subsidence. Three-dimensional problem domains of the FEMs include a thermoelastic PFD embedded in an elastic substrate. The thickness of the PFD is initially determined from the difference between post- and pre-eruption digital elevation models (DEMs). The initial excess temperature of the PFD at the time of deposition, 640°C, is estimated from FEM predictions and an InSAR image via standard least-squares inverse methods. Although the FEM predicts the major features of the observed transient deformation, systematic prediction errors (RMSE = 2.2 cm) are most likely associated with errors in the a priori PFD thickness distribution estimated from the DEM differences. We combine an InSAR image, FEMs, and an adaptive mesh algorithm to iteratively optimize the geometry of the PFD with respect to a minimized misfit between the predicted thermoelastic deformation and observed deformation. Prediction errors from an FEM, which includes an optimized PFD geometry and the initial excess PFD temperature estimated from the least-squares analysis, are sub-millimeter (RMSE = 0.3 mm). The average thickness (9.3 m), maximum thickness (126 m), and volume (2.1×107m3) of the PFD, estimated using the adaptive mesh algorithm, are about twice as large as the respective estimations for the a priori PFD geometry. Sensitivity analyses suggest unrealistic PFD thickness distributions are required for initial excess PFD temperatures outside of the range 500–800°C.
Acquire High Quality Meshes of Scale Models for AN Automatic Modelling Process
NASA Astrophysics Data System (ADS)
Giraud, F.; Jacquot, K.; Chevrier, C.; Halin, G.
2013-07-01
Urban scale models depicting whole towns such as the hundred-scale model collection known as plans-reliefs are a valuable source of information of cities and their surroundings. These physical representations of French strongholds from the 17th through the 19th century suffer from many problems that are, among other things, wear and tear or the lack of visibility and accessibility. A virtual collection would allow remote accessibility for visitors as well as history researchers. Moreover, it may also be linked to other digital collections and therefore, promote the collection to make people come to the museums to see the physical scale models. We also work on other physical town scale models like Epinal for which the scale is a bit higher. In a first part, we define a protocol for acquiring 3D meshes of town scale models from both photogrammetric and scanning methods. Then we compare the results of both methods The photogrammetric protocol has been elaborated by choosing the most accurate software, 123DCatch, which asks for about 60 pictures, and defining the settings needed to obtain exploitable photographs. In the same way, we defined the devices and settings needed for the laser scan acquisition method. In a second part, we segment the 3D meshes in planes by using Geomagic, which has been chosen between several programs, for its accurate resulting geometry.
Lee, W H; Kim, T S; Kim, Andrew T; Lee, S Y
2008-01-01
Realistic finite element (FE) head models have been successfully applied to bioelectromagnetic problems due to a realistic representation of arbitrary head geometry with inclusion of anisotropic material properties. In this paper, we propose a new automatic FE mesh generation scheme to generate a diffusion tensor MRI (DT-MRI) white matter anisotropy content-adaptive FE head model. We term this kind of mesh as wMesh. With this meshing technique, the anisotropic electrical conductivities derived from DT-MRIs can be best incorporated into the model. The influence of the white matter anisotropy on the EEG forward solutions has been studied via our wMesh head models. The scalp potentials computed from the anisotropic wMesh models against those of the isotropic models have been compared. The results describe that there are substantial changes in the scalp electrical potentials between the isotropic and anisotropic models, indicating that the inclusion of the white matter anisotropy is critical for accurate computation of E/MEG forward and inverse solutions. This fully automatic anisotropy-adaptive wMesh meshing scheme could be useful for modeling of individual-specific FE head models with better incorporation of the white matter anisotropic property towards bioelectromagnetic imaging.
Manual for automatic generation of finite element models of spiral bevel gears in mesh
NASA Technical Reports Server (NTRS)
Bibel, G. D.; Reddy, S.; Kumar, A.
1994-01-01
The goal of this research is to develop computer programs that generate finite element models suitable for doing 3D contact analysis of faced milled spiral bevel gears in mesh. A pinion tooth and a gear tooth are created and put in mesh. There are two programs: Points.f and Pat.f to perform the analysis. Points.f is based on the equation of meshing for spiral bevel gears. It uses machine tool settings to solve for an N x M mesh of points on the four surfaces, pinion concave and convex, and gear concave and convex. Points.f creates the file POINTS.OUT, an ASCI file containing N x M points for each surface. (N is the number of node points along the length of the tooth, and M is nodes along the height.) Pat.f reads POINTS.OUT and creates the file tl.out. Tl.out is a series of PATRAN input commands. In addition to the mesh density on the tooth face, additional user specified variables are the number of finite elements through the thickness, and the number of finite elements along the tooth full fillet. A full fillet is assumed to exist for both the pinion and gear.
Selection of finite-element mesh parameters in modeling the growth of hydraulic fracturing cracks
NASA Astrophysics Data System (ADS)
Kurguzov, V. D.
2016-12-01
The effect of the mesh geometry on the accuracy of solutions obtained by the finite-element method for problems of linear fracture mechanics is investigated. The guidelines have been formulated for constructing an optimum mesh for several routine problems involving elements with linear and quadratic approximation of displacements. The accuracy of finite-element solutions is estimated based on the degree of the difference between the calculated stress-intensity factor (SIF) and its value obtained analytically. In problems of hydrofracturing of oil-bearing formation, the pump-in pressure of injected water produces a distributed load on crack flanks as opposed to standard fracture mechanics problems that have analytical solutions, where a load is applied to the external boundaries of the computational region and the cracks themselves are kept free from stresses. Some model pressure profiles, as well as pressure profiles taken from real hydrodynamic computations, have been considered. Computer models of cracks with allowance for the pre-stressed state, fracture toughness, and elastic properties of materials are developed in the MSC.Marc 2012 finite-element analysis software. The Irwin force criterion is used as a criterion of brittle fracture and the SIFs are computed using the Cherepanov-Rice invariant J-integral. The process of crack propagation in a linearly elastic isotropic body is described in terms of the elastic energy release rate G and modeled using the VCCT (Virtual Crack Closure Technique) approach. It has been found that the solution accuracy is sensitive to the mesh configuration. Several parameters that are decisive in constructing effective finite-element meshes, namely, the minimum element size, the distance between mesh nodes in the vicinity of a crack tip, and the ratio of the height of an element to its length, have been established. It has been shown that a mesh that consists of only small elements does not improve the accuracy of the solution.
3D active shape models of human brain structures: application to patient-specific mesh generation
NASA Astrophysics Data System (ADS)
Ravikumar, Nishant; Castro-Mateos, Isaac; Pozo, Jose M.; Frangi, Alejandro F.; Taylor, Zeike A.
2015-03-01
The use of biomechanics-based numerical simulations has attracted growing interest in recent years for computer-aided diagnosis and treatment planning. With this in mind, a method for automatic mesh generation of brain structures of interest, using statistical models of shape (SSM) and appearance (SAM), for personalised computational modelling is presented. SSMs are constructed as point distribution models (PDMs) while SAMs are trained using intensity profiles sampled from a training set of T1-weighted magnetic resonance images. The brain structures of interest are, the cortical surface (cerebrum, cerebellum & brainstem), lateral ventricles and falx-cerebri membrane. Two methods for establishing correspondences across the training set of shapes are investigated and compared (based on SSM quality): the Coherent Point Drift (CPD) point-set registration method and B-spline mesh-to-mesh registration method. The MNI-305 (Montreal Neurological Institute) average brain atlas is used to generate the template mesh, which is deformed and registered to each training case, to establish correspondence over the training set of shapes. 18 healthy patients' T1-weightedMRimages form the training set used to generate the SSM and SAM. Both model-training and model-fitting are performed over multiple brain structures simultaneously. Compactness and generalisation errors of the BSpline-SSM and CPD-SSM are evaluated and used to quantitatively compare the SSMs. Leave-one-out cross validation is used to evaluate SSM quality in terms of these measures. The mesh-based SSM is found to generalise better and is more compact, relative to the CPD-based SSM. Quality of the best-fit model instance from the trained SSMs, to test cases are evaluated using the Hausdorff distance (HD) and mean absolute surface distance (MASD) metrics.
Adaptive Numerical Algorithms in Space Weather Modeling
NASA Technical Reports Server (NTRS)
Toth, Gabor; vanderHolst, Bart; Sokolov, Igor V.; DeZeeuw, Darren; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Nakib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav
2010-01-01
Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different physics in different domains. A multi-physics system can be modeled by a software framework comprising of several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solar wind Roe Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamics (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit numerical
Adaptive numerical algorithms in space weather modeling
NASA Astrophysics Data System (ADS)
Tóth, Gábor; van der Holst, Bart; Sokolov, Igor V.; De Zeeuw, Darren L.; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Najib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav
2012-02-01
Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different relevant physics in different domains. A multi-physics system can be modeled by a software framework comprising several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solarwind Roe-type Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamic (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit
Medical case-based retrieval: integrating query MeSH terms for query-adaptive multi-modal fusion
NASA Astrophysics Data System (ADS)
Seco de Herrera, Alba G.; Foncubierta-Rodríguez, Antonio; Müller, Henning
2015-03-01
Advances in medical knowledge give clinicians more objective information for a diagnosis. Therefore, there is an increasing need for bibliographic search engines that can provide services helping to facilitate faster information search. The ImageCLEFmed benchmark proposes a medical case-based retrieval task. This task aims at retrieving articles from the biomedical literature that are relevant for differential diagnosis of query cases including a textual description and several images. In the context of this campaign many approaches have been investigated showing that the fusion of visual and text information can improve the precision of the retrieval. However, fusion does not always lead to better results. In this paper, a new query-adaptive fusion criterion to decide when to use multi-modal (text and visual) or only text approaches is presented. The proposed method integrates text information contained in MeSH (Medical Subject Headings) terms extracted and visual features of the images to find synonym relations between them. Given a text query, the query-adaptive fusion criterion decides when it is suitable to also use visual information for the retrieval. Results show that this approach can decide if a text or multi{modal approach should be used with 77.15% of accuracy.
Modeling, Modal Properties, and Mesh Stiffness Variation Instabilities of Planetary Gears
NASA Technical Reports Server (NTRS)
Parker, Robert G.; Lin, Jian; Krantz, Timothy L. (Technical Monitor)
2001-01-01
Planetary gear noise and vibration are primary concerns in their applications in helicopters, automobiles, aircraft engines, heavy machinery and marine vehicles. Dynamic analysis is essential to the noise and vibration reduction. This work analytically investigates some critical issues and advances the understanding of planetary gear dynamics. A lumped-parameter model is built for the dynamic analysis of general planetary gears. The unique properties of the natural frequency spectra and vibration modes are rigorously characterized. These special structures apply for general planetary gears with cyclic symmetry and, in practically important case, systems with diametrically opposed planets. The special vibration properties are useful for subsequent research. Taking advantage of the derived modal properties, the natural frequency and vibration mode sensitivities to design parameters are investigated. The key parameters include mesh stiffnesses, support/bearing stiffnesses, component masses, moments of inertia, and operating speed. The eigen-sensitivities are expressed in simple, closed-form formulae associated with modal strain and kinetic energies. As disorders (e.g., mesh stiffness variation. manufacturing and assembling errors) disturb the cyclic symmetry of planetary gears, their effects on the free vibration properties are quantitatively examined. Well-defined veering rules are derived to identify dramatic changes of natural frequencies and vibration modes under parameter variations. The knowledge of free vibration properties, eigen-sensitivities, and veering rules provide important information to effectively tune the natural frequencies and optimize structural design to minimize noise and vibration. Parametric instabilities excited by mesh stiffness variations are analytically studied for multi-mesh gear systems. The discrepancies of previous studies on parametric instability of two-stage gear chains are clarified using perturbation and numerical methods. The
Refining 3D Earth models by unifying geological and geophysical information on unstructured meshes
NASA Astrophysics Data System (ADS)
Lelièvre, P. G.; Carter-McAuslan, A.; Tycholiz, C.; Farquharson, C. G.; Hurich, C. A.
2012-04-01
Earth models used for mineral exploration or other subsurface investigations should be consistent with all available geological and geophysical information. Geophysical inversion provides the means to integrate geological information, geophysical survey data, and physical property measurements taken on rock samples. Incorporation of geological information into inversions is always an iterative process. One begins with the geologists' best guess about the Earth (i.e. the geological model) and the models recovered from geophysical inversion may indicate that the geological model should be changed slightly prior to the next iteration of the procedure. In this way, geological and geophysical data can be combined through inversion and we can move towards the creation of a common Earth model consistent with all the available data. As more information is incorporated, the inherent non-uniqueness of the inverse problem is reduced, yielding a higher potential to resolve deeper features that are less well-constrained by the geophysical data alone. Geological ore deposit models are commonly created during delineation drilling. The accuracy of these models is crucial when used to determine if a deposit is economic. 3D geological Earth models typically comprise wireframe surfaces that represent the geological contacts between different rock units. The contacts may be known at points from down-hole intersections and surface mapping, and can be interpolated between boreholes and extrapolated outwards. Contacts may also be interpreted from seismic traces. Wireframe surfaces, comprising tessellated triangular facets, are sufficiently flexible to allow the representation of arbitrarily complicated geological structures. These surfaces can be honoured exactly within fully unstructured 3D volumetric tetrahedral meshes. In contrast, geophysical forward modelling and inversion algorithms typically work with rectilinear meshes when parameterizing the subsurface because this simplifies
A parallel second-order adaptive mesh algorithm for incompressible flow in porous media.
Pau, George S H; Almgren, Ann S; Bell, John B; Lijewski, Michael J
2009-11-28
In this paper, we present a second-order accurate adaptive algorithm for solving multi-phase, incompressible flow in porous media. We assume a multi-phase form of Darcy's law with relative permeabilities given as a function of the phase saturation. The remaining equations express conservation of mass for the fluid constituents. In this setting, the total velocity, defined to be the sum of the phase velocities, is divergence free. The basic integration method is based on a total-velocity splitting approach in which we solve a second-order elliptic pressure equation to obtain a total velocity. This total velocity is then used to recast component conservation equations as nonlinear hyperbolic equations. Our approach to adaptive refinement uses a nested hierarchy of logically rectangular grids with simultaneous refinement of the grids in both space and time. The integration algorithm on the grid hierarchy is a recursive procedure in which coarse grids are advanced in time, fine grids are advanced multiple steps to reach the same time as the coarse grids and the data at different levels are then synchronized. The single-grid algorithm is described briefly, but the emphasis here is on the time-stepping procedure for the adaptive hierarchy. Numerical examples are presented to demonstrate the algorithm's accuracy and convergence properties and to illustrate the behaviour of the method.
A Parallel Second-Order Adaptive Mesh Algorithm for Incompressible Flow in Porous Media
Pau, George Shu Heng; Almgren, Ann S.; Bell, John B.; Lijewski, Michael J.
2008-04-01
In this paper we present a second-order accurate adaptive algorithm for solving multiphase, incompressible flows in porous media. We assume a multiphase form of Darcy's law with relative permeabilities given as a function of the phase saturation. The remaining equations express conservation of mass for the fluid constituents. In this setting the total velocity, defined to be the sum of the phase velocities, is divergence-free. The basic integration method is based on a total-velocity splitting approach in which we solve a second-order elliptic pressure equation to obtain a total velocity. This total velocity is then used to recast component conservation equations as nonlinear hyperbolic equations. Our approach to adaptive refinement uses a nested hierarchy of logically rectangular grids with simultaneous refinement of the grids in both space and time. The integration algorithm on the grid hierarchy is a recursive procedure in which coarse grids are advanced in time, fine grids areadvanced multiple steps to reach the same time as the coarse grids and the data atdifferent levels are then synchronized. The single grid algorithm is described briefly,but the emphasis here is on the time-stepping procedure for the adaptive hierarchy. Numerical examples are presented to demonstrate the algorithm's accuracy and convergence properties and to illustrate the behavior of the method.
Valiant load-balanced robust routing under hose model for WDM mesh networks
NASA Astrophysics Data System (ADS)
Zhang, Xiaoning; Li, Lemin; Wang, Sheng
2006-09-01
In this paper, we propose Valiant Load-Balanced robust routing scheme for WDM mesh networks under the model of polyhedral uncertainty (i.e., hose model), and the proposed routing scheme is implemented with traffic grooming approach. Our Objective is to maximize the hose model throughput. A mathematic formulation of Valiant Load-Balanced robust routing is presented and three fast heuristic algorithms are also proposed. When implementing Valiant Load-Balanced robust routing scheme to WDM mesh networks, a novel traffic-grooming algorithm called MHF (minimizing hop first) is proposed. We compare the three heuristic algorithms with the VPN tree under the hose model. Finally we demonstrate in the simulation results that MHF with Valiant Load-Balanced robust routing scheme outperforms the traditional traffic-grooming algorithm in terms of the throughput for the uniform/non-uniform traffic matrix under the hose model.
Rate sensitive continuum damage models and mesh dependence in finite element analyses.
Ljustina, Goran; Fagerström, Martin; Larsson, Ragnar
2014-01-01
The experiences from orthogonal machining simulations show that the Johnson-Cook (JC) dynamic failure model exhibits significant element size dependence. Such mesh dependence is a direct consequence of the utilization of local damage models. The current contribution is an investigation of the extent of the possible pathological mesh dependence. A comparison of the resulting JC model behavior combined with two types of damage evolution is considered. The first damage model is the JC dynamic failure model, where the development of the "damage" does not affect the response until the critical state is reached. The second one is a continuum damage model, where the damage variable is affecting the material response continuously during the deformation. Both the plasticity and the damage models are rate dependent, and the damage evolutions for both models are defined as a postprocessing of the effective stress response. The investigation is conducted for a series of 2D shear tests utilizing different FE representations of the plane strain plate with pearlite material properties. The results show for both damage models, using realistic pearlite material parameters, that similar extent of the mesh dependence is obtained and that the possible viscous regularization effects are absent in the current investigation.
Surface mesh to voxel data registration for patient-specific anatomical modeling
NASA Astrophysics Data System (ADS)
de Oliveira, Júlia E. E.; Giessler, Paul; Keszei, András.; Herrler, Andreas; Deserno, Thomas M.
2016-03-01
Virtual Physiological Human (VPH) models are frequently used for training, planning, and performing medical procedures. The Regional Anaesthesia Simulator and Assistant (RASimAs) project has the goal of increasing the application and effectiveness of regional anesthesia (RA) by combining a simulator of ultrasound-guided and electrical nerve-stimulated RA procedures and a subject-specific assistance system through an integration of image processing, physiological models, subject-specific data, and virtual reality. Individualized models enrich the virtual training tools for learning and improving regional anaesthesia (RA) skills. Therefore, we suggest patient-specific VPH models that are composed by registering the general mesh-based models with patient voxel data-based recordings. Specifically, the pelvis region has been focused for the support of the femoral nerve block. The processing pipeline is composed of different freely available toolboxes such as MatLab, the open Simulation framework (SOFA), and MeshLab. The approach of Gilles is applied for mesh-to-voxel registration. Personalized VPH models include anatomical as well as mechanical properties of the tissues. Two commercial VPH models (Zygote and Anatomium) were used together with 34 MRI data sets. Results are presented for the skin surface and pelvic bones. Future work will extend the registration procedure to cope with all model tissue (i.e., skin, muscle, bone, vessel, nerve, fascia) in a one-step procedure and extrapolating the personalized models to body regions actually being out of the captured field of view.
Finite element meshing of ANSYS (trademark) solid models
NASA Technical Reports Server (NTRS)
Kelley, F. S.
1987-01-01
A large scale, general purpose finite element computer program, ANSYS, developed and marketed by Swanson Analysis Systems, Inc. is discussed. ANSYS was perhaps the first commercially available program to offer truly interactive finite element model generation. ANSYS's purpose is for solid modeling. This application is briefly discussed and illustrated.
The Dynamic Modelling of a Spur Gear in Mesh Including Friction and a Crack
NASA Astrophysics Data System (ADS)
Howard, Ian; Jia, Shengxiang; Wang, Jiande
2001-09-01
To improve the current generation of diagnostic techniques, many researchers are actively developing advanced dynamic models of gear case vibration to ascertain the effect of different types of gear train damage. This paper details a simplified gear dynamic model aimed at exploring the effect of friction on the resultant gear case vibration. The model incorporates the effect of variations in gear tooth torsional mesh stiffness, developed using finite element analysis, as the gears mesh together. The method of introducing the frictional force between teeth into the dynamic equations is given. The comparison between the results with friction and without friction was investigated using Matlab and Simulink models developed from the differential equations. The effects the single tooth crack has on the frequency spectrum and on the common diagnostic functions of the resulting gearbox component vibrations are also shown.
Adaptive h -refinement for reduced-order models: ADAPTIVE h -refinement for reduced-order models
Carlberg, Kevin T.
2014-11-05
Our work presents a method to adaptively refine reduced-order models a posteriori without requiring additional full-order-model solves. The technique is analogous to mesh-adaptive h-refinement: it enriches the reduced-basis space online by ‘splitting’ a given basis vector into several vectors with disjoint support. The splitting scheme is defined by a tree structure constructed offline via recursive k-means clustering of the state variables using snapshot data. This method identifies the vectors to split online using a dual-weighted-residual approach that aims to reduce error in an output quantity of interest. The resulting method generates a hierarchy of subspaces online without requiring large-scale operationsmore » or full-order-model solves. Furthermore, it enables the reduced-order model to satisfy any prescribed error tolerance regardless of its original fidelity, as a completely refined reduced-order model is mathematically equivalent to the original full-order model. Experiments on a parameterized inviscid Burgers equation highlight the ability of the method to capture phenomena (e.g., moving shocks) not contained in the span of the original reduced basis.« less
Mesh-free particle continuum-based modeling of complex geophysical flows
NASA Astrophysics Data System (ADS)
Shakibaeinia, A.
2015-12-01
Complex geophysical flows such as tsunamis, water floods, and avalanche/debris flows, wave atmosphere interaction and violent river flow are the multiphysics flows characterized by large interfacial deformation and fragmentations and frequent topology change in various spatial and temporal scales. Dealing with such characteristics is still beyond the capabilities of many conventional mesh-based Eulerian numerical methods. Mesh-free particle (Lagrangian) numerical methods, such as Moving Particle Semi-implicit (MPS) and Smoothed particle hydrodynamics (SPH), which have been developed for continuum mechanics, offer a unique opportunity to deal with such complexities. This research presents a mesh-free particle method for complex multi-physics geophysical flows, based on a quasi incompressible MPS-based formulation for turbulent liquid, solid and gas multiphase systems. In combination with the various constitutive relations it is capable of dealing with the non-Newtonian behaviors in geophysical flows (e.g. for case of granular flows). The capabilities of proposed numerical method are evaluated and demonstrated for a series of computationally demanding examples including submarine landslide (and resulting tsunami wave), river flow/ice dynamics, and wave-air interaction. The capabilities of the model in accurate prediction of flow characteristics such as interfacial deformation, velocity/pressure fields, turbulence and mixing processes are demonstrated. Results of this study not only evaluates the capabilities of mesh-free particle methods for complex geophysical flows, but only provides a reliable tool for in deep analysis of mechanisms and processes involved in such flow systems.
NASA Astrophysics Data System (ADS)
Pelties, C.; Käser, M.
2010-12-01
We will present recent developments concerning the extensions of the ADER-DG method to solve three dimensional dynamic rupture problems on unstructured tetrahedral meshes. The simulation of earthquake rupture dynamics and seismic wave propagation using a discontinuous Galerkin (DG) method in 2D was recently presented by J. de la Puente et al. (2009). A considerable feature of this study regarding spontaneous rupture problems was the combination of the DG scheme and a time integration method using Arbitrarily high-order DERivatives (ADER) to provide high accuracy in space and time with the discretization on unstructured meshes. In the resulting discrete velocity-stress formulation of the elastic wave equations variables are naturally discontinuous at the interfaces between elements. The so-called Riemann problem can then be solved to obtain well defined values of the variables at the discontinuity itself. This is in particular valid for the fault at which a certain friction law has to be evaluated. Hence, the fault’s geometry is honored by the computational mesh. This way, complex fault planes can be modeled adequately with small elements while fast mesh coarsening is possible with increasing distance from the fault. Due to the strict locality of the scheme using only direct neighbor communication, excellent parallel behavior can be observed. A further advantage of the scheme is that it avoids spurious high-frequency contributions in the slip rate spectra and therefore does not require artificial Kelvin-Voigt damping or filtering of synthetic seismograms. In order to test the accuracy of the ADER-DG method the Southern California Earthquake Center (SCEC) benchmark for spontaneous rupture simulations was employed. Reference: J. de la Puente, J.-P. Ampuero, and M. Käser (2009), Dynamic rupture modeling on unstructured meshes using a discontinuous Galerkin method, JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 114, B10302, doi:10.1029/2008JB006271
NASA Technical Reports Server (NTRS)
Mikhaylov, Rebecca; Dawson, Douglas; Kwack, Eug
2014-01-01
NASA's Earth observing Soil Moisture Active & Passive (SMAP) Mission is scheduled to launch in November 2014 into a 685 km near-polar, sun synchronous orbit. SMAP will provide comprehensive global mapping measurements of soil moisture and freeze/thaw state in order to enhance understanding of the processes that link the water, energy, and carbon cycles. The primary objectives of SMAP are to improve worldwide weather and flood forecasting, enhance climate prediction, and refine drought and agriculture monitoring during its 3 year mission. The SMAP instrument architecture incorporates an L-band radar and an L-band radiometer which share a common feed horn and parabolic mesh reflector. The instrument rotates about the nadir axis at approximately 15 rpm, thereby providing a conically scanning wide swath antenna beam that is capable of achieving global coverage within 3 days. In order to make the necessary precise surface emission measurements from space, a temperature knowledge of 60 deg C for the mesh reflector is required. In order to show compliance, a thermal vacuum test was conducted using a portable solar simulator to illuminate a non flight, but flight-like test article through the quartz window of the vacuum chamber. The molybdenum wire of the antenna mesh is too fine to accommodate thermal sensors for direct temperature measurements. Instead, the mesh temperature was inferred from resistance measurements made during the test. The test article was rotated to five separate angles between 10 deg and 90 deg via chamber breaks to simulate the maximum expected on-orbit solar loading during the mission. The resistance measurements were converted to temperature via a resistance versus temperature calibration plot that was constructed from data collected in a separate calibration test. A simple thermal model of two different representations of the mesh (plate and torus) was created to correlate the mesh temperature predictions to within 60 deg C. The on-orbit mesh
Adaptivity via mesh movement with three-dimensional block-structured grids
Catherall, D.
1996-12-31
The method described here is one in which grid nodes are redistributed so that they are attracted towards regions of high solution activity. The major difficulty in attempting this arises from the degree of grid smoothness and orthogonality required by the flow solver. These requirements are met by suitable choice of grid equations, to be satisfied by the adapted grid, and by the inclusion of certain source terms, for added control in regions where grid movement is limited by the local geometry. The method has been coded for multiblock grids, so that complex configurations may be treated. It is demonstrated here for inviscid supercritical flow with two test cases: an ONERA M6 wing with a rounded tip, and a forward-swept wing/fuselage configuration (M151).
Model of aircraft noise adaptation
NASA Technical Reports Server (NTRS)
Dempsey, T. K.; Coates, G. D.; Cawthorn, J. M.
1977-01-01
Development of an aircraft noise adaptation model, which would account for much of the variability in the responses of subjects participating in human response to noise experiments, was studied. A description of the model development is presented. The principal concept of the model, was the determination of an aircraft adaptation level which represents an annoyance calibration for each individual. Results showed a direct correlation between noise level of the stimuli and annoyance reactions. Attitude-personality variables were found to account for varying annoyance judgements.
Hybrid Adaptive Flight Control with Model Inversion Adaptation
NASA Technical Reports Server (NTRS)
Nguyen, Nhan
2011-01-01
This study investigates a hybrid adaptive flight control method as a design possibility for a flight control system that can enable an effective adaptation strategy to deal with off-nominal flight conditions. The hybrid adaptive control blends both direct and indirect adaptive control in a model inversion flight control architecture. The blending of both direct and indirect adaptive control provides a much more flexible and effective adaptive flight control architecture than that with either direct or indirect adaptive control alone. The indirect adaptive control is used to update the model inversion controller by an on-line parameter estimation of uncertain plant dynamics based on two methods. The first parameter estimation method is an indirect adaptive law based on the Lyapunov theory, and the second method is a recursive least-squares indirect adaptive law. The model inversion controller is therefore made to adapt to changes in the plant dynamics due to uncertainty. As a result, the modeling error is reduced that directly leads to a decrease in the tracking error. In conjunction with the indirect adaptive control that updates the model inversion controller, a direct adaptive control is implemented as an augmented command to further reduce any residual tracking error that is not entirely eliminated by the indirect adaptive control.
A mesh-free particle model for simulation of mobile-bed dam break
NASA Astrophysics Data System (ADS)
Shakibaeinia, Ahmad; Jin, Yee-Chung
2011-06-01
Mesh-free particle (Lagrangian) methods such as Moving Particle Semi-Implicit (MPS) and Smoothed Particle Hydrodynamics (SPH) are the latest generation of methods in the field of computational fluid dynamics that attracts lots of attention in modeling applications where large interfacial deformations and fragmentations exist. Due to their mesh-free nature, these methods are capable of simulating any kind of boundary/interface deformation and fragmentations. This study aims to develop a new mesh-free particle model based on the weakly compressible MPS (WC-MPS) formulation for modeling of a dam break over a mobile bed, which is a highly erosive and transient flow problem. A multiphase model, capable of handling the density and viscosity discontinuity and in which the solid (sediment) phase is treated as a non-Newtonian fluid, is introduced. The resulting model is first validated using a two-phase dam break problem and is then applied to the mobile-bed dam break problem with different conditions, comparing the results to those obtained from some experimental works.
ADAPTIVE EYE MODEL - Poster Paper
NASA Astrophysics Data System (ADS)
Galetskiy, Sergey O.; Kudryashov, Alexey V.
2008-01-01
We propose experimental adaptive eye model based on flexible 18-electrode bimorph mirror reproducing human eye aberrations up to 4th radial order of Zernike polynomials at frequency of 10Hz. The accuracy of aberrations reproduction in most cases is better than λ/10 RMS. The model is introduced to aberrometer for human eye aberrations compensation to improve visual acuity test.
Rasia, Elena; Lau, Erwin T.; Nagai, Daisuke; Avestruz, Camille; Borgani, Stefano; Dolag, Klaus; Granato, Gian Luigi; Murante, Giuseppe; Ragone-Figueroa, Cinthia; Mazzotta, Pasquale; Nelson, Kaylea
2014-08-20
Analyses of cosmological hydrodynamic simulations of galaxy clusters suggest that X-ray masses can be underestimated by 10%-30%. The largest bias originates from both violation of hydrostatic equilibrium (HE) and an additional temperature bias caused by inhomogeneities in the X-ray-emitting intracluster medium (ICM). To elucidate this large dispersion among theoretical predictions, we evaluate the degree of temperature structures in cluster sets simulated either with smoothed-particle hydrodynamics (SPH) or adaptive-mesh refinement (AMR) codes. We find that the SPH simulations produce larger temperature variations connected to the persistence of both substructures and their stripped cold gas. This difference is more evident in nonradiative simulations, whereas it is reduced in the presence of radiative cooling. We also find that the temperature variation in radiative cluster simulations is generally in agreement with that observed in the central regions of clusters. Around R {sub 500} the temperature inhomogeneities of the SPH simulations can generate twice the typical HE mass bias of the AMR sample. We emphasize that a detailed understanding of the physical processes responsible for the complex thermal structure in ICM requires improved resolution and high-sensitivity observations in order to extend the analysis to higher temperature systems and larger cluster-centric radii.
Selvan, S Easter; Borckmans, Pierre B; Chattopadhyay, A; Absil, P-A
2013-09-01
It is seemingly paradoxical to the classical definition of the independent component analysis (ICA), that in reality, the true sources are often not strictly uncorrelated. With this in mind, this letter concerns a framework to extract quasi-uncorrelated sources with finite supports by optimizing a range-based contrast function under unit-norm constraints (to handle the inherent scaling indeterminacy of ICA) but without orthogonality constraints. Albeit the appealing contrast properties of the range-based function (e.g., the absence of mixing local optima), the function is not differentiable everywhere. Unfortunately, there is a dearth of literature on derivative-free optimizers that effectively handle such a nonsmooth yet promising contrast function. This is the compelling reason for the design of a nonsmooth optimization algorithm on a manifold of matrices having unit-norm columns with the following objectives: to ascertain convergence to a Clarke stationary point of the contrast function and adhere to the necessary unit-norm constraints more naturally. The proposed nonsmooth optimization algorithm crucially relies on the design and analysis of an extension of the mesh adaptive direct search (MADS) method to handle locally Lipschitz objective functions defined on the sphere. The applicability of the algorithm in the ICA domain is demonstrated with simulations involving natural, face, aerial, and texture images.
NASA Astrophysics Data System (ADS)
Fakhari, Abbas; Lee, Taehun
2013-11-01
A novel adaptive mesh refinement (AMR) algorithm for the numerical solution of fluid flow problems is presented in this study. The proposed AMR algorithm can be used to solve partial differential equations including, but not limited to, the Navier-Stokes equations using an AMR technique. Here, the lattice Boltzmann method (LBM) is employed as a substitute of the nearly incompressible Navier-Stokes equations. Besides its simplicity, the proposed AMR algorithm is straightforward and yet efficient. The idea is to remove the need for a tree-type data structure by using the pointer attributes in a unique way, along with an appropriate adjustment of the child block's IDs, to determine the neighbors of a certain block. Thanks to the unique way of invoking pointers, there is no need to construct a quad-tree (in 2D) or oct-tree (in 3D) data structure for maintaining the connectivity data between different blocks. As a result, the memory and time required for tree traversal are completely eliminated, leaving us with a clean and efficient algorithm that is easier to implement and use on parallel machines. Several benchmark studies are carried out to assess the accuracy and efficiency of the proposed AMR-LBM, including lid-driven cavity flow, vortex shedding past a square cylinder, and Kelvin-Helmholtz instability for single-phase and multiphase fluids.
MAST-2D diffusive model for flood prediction on domains with triangular Delaunay unstructured meshes
NASA Astrophysics Data System (ADS)
Aricò, C.; Sinagra, M.; Begnudelli, L.; Tucciarelli, T.
2011-11-01
A new methodology for the solution of the 2D diffusive shallow water equations over Delaunay unstructured triangular meshes is presented. Before developing the new algorithm, the following question is addressed: it is worth developing and using a simplified shallow water model, when well established algorithms for the solution of the complete one do exist? The governing Partial Differential Equations are discretized using a procedure similar to the linear conforming Finite Element Galerkin scheme, with a different flux formulation and a special flux treatment that requires Delaunay triangulation but entire solution monotonicity. A simple mesh adjustment is suggested, that attains the Delaunay condition for all the triangle sides without changing the original nodes location and also maintains the internal boundaries. The original governing system is solved applying a fractional time step procedure, that solves consecutively a convective prediction system and a diffusive correction system. The non linear components of the problem are concentrated in the prediction step, while the correction step leads to the solution of a linear system of the order of the number of computational cells. A semi-analytical procedure is applied for the solution of the prediction step. The discretized formulation of the governing equations allows to handle also wetting and drying processes without any additional specific treatment. Local energy dissipations, mainly the effect of vertical walls and hydraulic jumps, can be easily included in the model. Several numerical experiments have been carried out in order to test (1) the stability of the proposed model with regard to the size of the Courant number and to the mesh irregularity, (2) its computational performance, (3) the convergence order by means of mesh refinement. The model results are also compared with the results obtained by a fully dynamic model. Finally, the application to a real field case with a Venturi channel is presented.
A mechanical model for deformable and mesh pattern wheel of lunar roving vehicle
NASA Astrophysics Data System (ADS)
Liang, Zhongchao; Wang, Yongfu; Chen, Gang (Sheng); Gao, Haibo
2015-12-01
As an indispensable tool for astronauts on lunar surface, the lunar roving vehicle (LRV) is of great significance for manned lunar exploration. An LRV moves on loose and soft lunar soil, so the mechanical property of its wheels directly affects the mobility performance. The wheels used for LRV have deformable and mesh pattern, therefore, the existing mechanical theory of vehicle wheel cannot be used directly for analyzing the property of LRV wheels. In this paper, a new mechanical model for LRV wheel is proposed. At first, a mechanical model for a rigid normal wheel is presented, which involves in multiple conventional parameters such as vertical load, tangential traction force, lateral force, and slip ratio. Secondly, six equivalent coefficients are introduced to amend the rigid normal wheel model to fit for the wheels with deformable and mesh-pattern in LRV application. Thirdly, the values of the six equivalent coefficients are identified by using experimental data obtained in an LRV's single wheel testing. Finally, the identified mechanical model for LRV's wheel with deformable and mesh pattern are further verified and validated by using additional experimental results.
On Modeling and Analysis of MIMO Wireless Mesh Networks with Triangular Overlay Topology
Cao, Zhanmao; Wu, Chase Q.; Zhang, Yuanping; ...
2015-01-01
Multiple input multiple output (MIMO) wireless mesh networks (WMNs) aim to provide the last-mile broadband wireless access to the Internet. Along with the algorithmic development for WMNs, some fundamental mathematical problems also emerge in various aspects such as routing, scheduling, and channel assignment, all of which require an effective mathematical model and rigorous analysis of network properties. In this paper, we propose to employ Cartesian product of graphs (CPG) as a multichannel modeling approach and explore a set of unique properties of triangular WMNs. In each layer of CPG with a single channel, we design a node coordinate scheme thatmore » retains the symmetric property of triangular meshes and develop a function for the assignment of node identity numbers based on their coordinates. We also derive a necessary-sufficient condition for interference-free links and combinatorial formulas to determine the number of the shortest paths for channel realization in triangular WMNs.« less
2015-01-07
and Anisotropic All-Hexahedral Mesh Generation Using a Hybrid Octree and Bubble Packing. 4. Octree-based Dual Contouring Method for Triangular and...Parameterization Using Eigenfunction-based Cross Field. Graphical Models, 76(6):691-705, 2014 8. X. Liang, Y. Zhang. An Octree-based Dual Contouring Method for...and Evaluation, the 27th Conference on Computer Animation and Social Agents (CASA 2014). Houston, TX. May 26-28, 2014. 13. K. Hu, J. Qian, Y. Zhang
The 2D and 3D hypersonic flows with unstructured meshes
NASA Technical Reports Server (NTRS)
Thareja, Rajiv
1993-01-01
Viewgraphs on 2D and 3D hypersonic flows with unstructured meshes are presented. Topics covered include: mesh generation, mesh refinement, shock-shock interaction, velocity contours, mesh movement, vehicle bottom surface, and adapted meshes.
Merging for Particle-Mesh Complex Particle Kinetic Modeling of the Multiple Plasma Beams
NASA Technical Reports Server (NTRS)
Lipatov, Alexander S.
2011-01-01
We suggest a merging procedure for the Particle-Mesh Complex Particle Kinetic (PMCPK) method in case of inter-penetrating flow (multiple plasma beams). We examine the standard particle-in-cell (PIC) and the PMCPK methods in the case of particle acceleration by shock surfing for a wide range of the control numerical parameters. The plasma dynamics is described by a hybrid (particle-ion-fluid-electron) model. Note that one may need a mesh if modeling with the computation of an electromagnetic field. Our calculations use specified, time-independent electromagnetic fields for the shock, rather than self-consistently generated fields. While a particle-mesh method is a well-verified approach, the CPK method seems to be a good approach for multiscale modeling that includes multiple regions with various particle/fluid plasma behavior. However, the CPK method is still in need of a verification for studying the basic plasma phenomena: particle heating and acceleration by collisionless shocks, magnetic field reconnection, beam dynamics, etc.
NASA Astrophysics Data System (ADS)
Tonello, N.; Eude, Y.; de Laage de Meux, B.; Ferrand, M.
2017-01-01
The steady-state operation of the Francis-99, Tokke turbine [1-3] has been simulated numerically at different loads using the open source, CAD and CFD software, SALOME [4] Code_Saturne [5]. The full 3D mesh of the Tokke turbine provided for the Second Francis-99 Workshop has been adapted and modified to work with the solver. Results are compared for the frozen-rotor and the unsteady, conservative sliding mesh approach over three operating points, showing that good agreement with the experimental data is obtained with both models without having to tune the CFD models for each operating point. Approaches to the simulation of transient operation are also presented with results of work in progress.
Chedid, R.; Najjar, N.
1996-09-01
One of the inconveniences associated with the existing finite-element packages is the need for an educated user to develop a correct mesh at the preprocessing level. Procedures which start with a coarse mesh and attempt serious refinements, as is the case in most adaptive finite-element packages, are time consuming and costly. Hence, it is very important to develop a tool that can provide a mesh that either leads immediately to an acceptable solution, or would require fewer correcting steps to achieve better results. In this paper, the authors present a technique for automatic mesh generation based on artificial neural networks (ANN). The essence of this technique is to predict the mesh density distribution of a given model, and then supply this information to a Kohonen neural network which provides the final mesh. Prediction of mesh density is accomplished by a simple feedforward neural network which has the ability to learn the relationship between mesh density and model geometric features. It will be shown that ANN are able to recognize delicate areas where a sharp variation of the magnetic field is expected. Examples of 2-D models are provided to illustrate the usefulness of the proposed technique.
Evolution of the mandibular mesh implant.
Salyer, K E; Johns, D F; Holmes, R E; Layton, J G
1977-07-01
Between 1960 and 1972, the Dallas Veterans Administration Hospital Maxillofacial Research Laboratory developed and made over 150 cast-mesh implants. Successive designs were ovoid, circular, and double-lumened in cross section to improve implant strength, surface area for bioattachment, and adjustability. Sleeves, collars, and bows were employed in the assembly of these implants, with an acrylic condylar head attached when indicated. In 1972, our laboratory developed a mandibular mesh tray, cast in one piece on a single sprue, with preservation of the vertically adjustable ramus. Stainless steel replaced Vitallium because of its greater malleability. Essentially, a lost-wax technique is used to cast the mesh tray. The model of a mandibular segment is duplicated as a refractory model. Mesh wax, made in our own custom-made die, is adapted to the refractory model. The unit is then sprued and invested. The wax is fired our of the mold in a gas furnace. Casting is done by the transferral of molten stainless steel from the crucible to the mold by centrifugal force in an electro-induction casting machine. Other mesh implants that have been developed are made from wire mesh, Dacron mesh, cast Ticonium, and hydroformed titanium.
Conversion of Component-Based Point Definition to VSP Model and Higher Order Meshing
NASA Technical Reports Server (NTRS)
Ordaz, Irian
2011-01-01
Vehicle Sketch Pad (VSP) has become a powerful conceptual and parametric geometry tool with numerous export capabilities for third-party analysis codes as well as robust surface meshing capabilities for computational fluid dynamics (CFD) analysis. However, a capability gap currently exists for reconstructing a fully parametric VSP model of a geometry generated by third-party software. A computer code called GEO2VSP has been developed to close this gap and to allow the integration of VSP into a closed-loop geometry design process with other third-party design tools. Furthermore, the automated CFD surface meshing capability of VSP are demonstrated for component-based point definition geometries in a conceptual analysis and design framework.
OMEGA: The operational multiscale environment model with grid adaptivity
Bacon, D.P.
1995-07-01
This review talk describes the OMEGA code, used for weather simulation and the modeling of aerosol transport through the atmosphere. Omega employs a 3D mesh of wedge shaped elements (triangles when viewed from above) that adapt with time. Because wedges are laid out in layers of triangular elements, the scheme can utilize structured storage and differencing techniques along the elevation coordinate, and is thus a hybrid of structured and unstructured methods. The utility of adaptive gridding in this moded, near geographic features such as coastlines, where material properties change discontinuously, is illustrated. Temporal adaptivity was used additionally to track moving internal fronts, such as clouds of aerosol contaminants. The author also discusses limitations specific to this problem, including manipulation of huge data bases and fixed turn-around times. In practice, the latter requires a carefully tuned optimization between accuracy and computation speed.
NASA Astrophysics Data System (ADS)
Skutnik, Steven E.; Davis, David R.
2016-05-01
The use of passive gamma and neutron signatures from fission indicators is a common means of estimating used fuel burnup, enrichment, and cooling time. However, while characteristic fission product signatures such as 134Cs, 137Cs, 154Eu, and others are generally reliable estimators for used fuel burnup within the context where the assembly initial enrichment and the discharge time are known, in the absence of initial enrichment and/or cooling time information (such as when applying NDA measurements in a safeguards/verification context), these fission product indicators no longer yield a unique solution for assembly enrichment, burnup, and cooling time after discharge. Through the use of a new Mesh-Adaptive Direct Search (MADS) algorithm, it is possible to directly probe the shape of this "degeneracy space" characteristic of individual nuclides (and combinations thereof), both as a function of constrained parameters (such as the assembly irradiation history) and unconstrained parameters (e.g., the cooling time before measurement and the measurement precision for particular indicator nuclides). In doing so, this affords the identification of potential means of narrowing the uncertainty space of potential assembly enrichment, burnup, and cooling time combinations, thereby bounding estimates of assembly plutonium content. In particular, combinations of gamma-emitting nuclides with distinct half-lives (e.g., 134Cs with 137Cs and 154Eu) in conjunction with gross neutron counting (via 244Cm) are able to reasonably constrain the degeneracy space of possible solutions to a space small enough to perform useful discrimination and verification of fuel assemblies based on their irradiation history.
Moving Overlapping Grids with Adaptive Mesh Refinement for High-Speed Reactive and Non-reactive Flow
Henshaw, W D; Schwendeman, D W
2005-08-30
We consider the solution of the reactive and non-reactive Euler equations on two-dimensional domains that evolve in time. The domains are discretized using moving overlapping grids. In a typical grid construction, boundary-fitted grids are used to represent moving boundaries, and these grids overlap with stationary background Cartesian grids. Block-structured adaptive mesh refinement (AMR) is used to resolve fine-scale features in the flow such as shocks and detonations. Refinement grids are added to base-level grids according to an estimate of the error, and these refinement grids move with their corresponding base-level grids. The numerical approximation of the governing equations takes place in the parameter space of each component grid which is defined by a mapping from (fixed) parameter space to (moving) physical space. The mapped equations are solved numerically using a second-order extension of Godunov's method. The stiff source term in the reactive case is handled using a Runge-Kutta error-control scheme. We consider cases when the boundaries move according to a prescribed function of time and when the boundaries of embedded bodies move according to the surface stress exerted by the fluid. In the latter case, the Newton-Euler equations describe the motion of the center of mass of the each body and the rotation about it, and these equations are integrated numerically using a second-order predictor-corrector scheme. Numerical boundary conditions at slip walls are described, and numerical results are presented for both reactive and non-reactive flows in order to demonstrate the use and accuracy of the numerical approach.
NASA Astrophysics Data System (ADS)
Komitopoulos, Nikolaos; Vakouftsis, Christos
2014-10-01
A Two-Degree Of Freedom analytical model of meshing in a single-stage spur gear set was developed and used for time-domain dynamic simulation. Apart from the time-varying tooth stiffness, the individual tooth mass, reduced to the meshing point, was also taken into consideration and modeled. The simulations that were performed by means of MatLab software using numerical methods highlight the effect of the individual tooth mass in the dynamic response of the gear stage.
Discretization of the Joule heating term for plasma discharge fluid models in unstructured meshes
Deconinck, T.; Mahadevan, S.; Raja, L.L.
2009-07-01
The fluid (continuum) approach is commonly used for simulation of plasma phenomena in electrical discharges at moderate to high pressures (>10's mTorr). The description comprises governing equations for charged and neutral species transport and energy equations for electrons and the heavy species, coupled to equations for the electromagnetic fields. The coupling of energy from the electrostatic field to the plasma species is modeled by the Joule heating term which appears in the electron and heavy species (ion) energy equations. Proper numerical discretization of this term is necessary for accurate description of discharge energetics; however, discretization of this term poses a special problem in the case of unstructured meshes owing to the arbitrary orientation of the faces enclosing each cell. We propose a method for the numerical discretization of the Joule heating term using a cell-centered finite volume approach on unstructured meshes with closed convex cells. The Joule heating term is computed by evaluating both the electric field and the species flux at the cell center. The dot product of these two vector quantities is computed to obtain the Joule heating source term. We compare two methods to evaluate the species flux at the cell center. One is based on reconstructing the fluxes at the cell centers from the fluxes at the face centers. The other recomputes the flux at the cell center using the common drift-diffusion approximation. The reconstructed flux scheme is the most stable method and yields reasonably accurate results on coarse meshes.
NASA Astrophysics Data System (ADS)
Schubert, Jochen E.; Sanders, Brett F.; Smith, Martin J.; Wright, Nigel G.
2008-12-01
Urban flood inundation modeling with a hydrodynamic flow solver is addressed in this paper, focusing on strategies to effectively integrate geospatial data for unstructured mesh generation, building representation and flow resistance parameterization. Data considered include Light Detection and Ranging (LiDAR) terrain height surveys, aerial imagery and vector datasets such as building footprint polygons. First, a unstructured mesh-generation technique we term the building-hole method (BH) is developed whereby building footprint data define interior domain boundaries or mesh holes. A wall boundary condition depicts the impact of buildings on flood hydrodynamics. BH provides an alternative to the more commonly used method of raising terrain heights where buildings coincide with the mesh. We term this the building-block method (BB). Application of BH and BB to a flooding site in Glasgow, Scotland identifies a number of tradeoffs to consider at resolutions ranging from 1 to 5 m. At fine resolution, BH is shown to be similarly accurate but execute faster than BB. And at coarse resolution, BH is shown to preserve the geometry of buildings and maintain better accuracy than BB, but requires a longer run time. Meshes that ignore buildings completely ( no-building method or NB) also support surprisingly good flood inundation predictions at coarse resolution compared to BH and BB. NB also supports faster execution times than BH at coarse resolution because the latter uses localized refinements that mandate a greater number of computational cells. However, with mesh refinement, NB converges to a different (and presumably less-accurate) solution compared to BH and BB. Using the same test conditions, Hunter et al. [Hunter NM, Bates PD, Neelz S, Pender G, Villanueva I, Wright NG, Liang D, et al. Benchmarking 2D hydraulic models for urban flood simulations. ICE J Water Manage 2008;161(1):13-30] compared the performance of dynamic-wave and diffusive-wave models and reported that
NASA Astrophysics Data System (ADS)
Skamarock, W. C.
2015-12-01
One of the major problems in atmospheric model applications is the representation of deep convection within the models; explicit simulation of deep convection on fine meshes performs much better than sub-grid parameterized deep convection on coarse meshes. Unfortunately, the high cost of explicit convective simulation has meant it has only been used to down-scale global simulations in weather prediction and regional climate applications, typically using traditional one-way interactive nesting technology. We have been performing real-time weather forecast tests using a global non-hydrostatic atmospheric model (the Model for Prediction Across Scales, MPAS) that employs a variable-resolution unstructured Voronoi horizontal mesh (nominally hexagons) to span hydrostatic to nonhydrostatic scales. The smoothly varying Voronoi mesh eliminates many downscaling problems encountered using traditional one- or two-way grid nesting. Our test weather forecasts cover two periods - the 2015 Spring Forecast Experiment conducted at the NOAA Storm Prediction Center during the month of May in which we used a 50-3 km mesh, and the PECAN field program examining nocturnal convection over the US during the months of June and July in which we used a 15-3 km mesh. An important aspect of this modeling system is that the model physics be scale-aware, particularly the deep convection parameterization. These MPAS simulations employ the Grell-Freitas scale-aware convection scheme. Our test forecasts show that the scheme produces a gradual transition in the deep convection, from the deep unstable convection being handled entirely by the convection scheme on the coarse mesh regions (dx > 15 km), to the deep convection being almost entirely explicit on the 3 km NA region of the meshes. We will present results illustrating the performance of critical aspects of the MPAS model in these tests.
NASA Astrophysics Data System (ADS)
Schroeder, Philipp W.; Lube, Gert
2017-04-01
This paper presents heavily grad-div and pressure jump stabilised, equal- and mixed-order discontinuous Galerkin finite element methods for non-isothermal incompressible flows based on the Oberbeck-Boussinesq approximation. In this framework, the enthalpy-porosity model for multiphase flow in melting and solidification problems can be employed. By considering the differentially heated cavity and the melting of pure gallium in a rectangular enclosure, it is shown that both boundary layers and sharp moving interior layers can be handled naturally by the proposed class of non-conforming methods. Due to the stabilising effect of the grad-div term and the robustness of discontinuous Galerkin methods, it is possible to solve the underlying problems accurately on coarse, non-adapted meshes. The interaction of heavy grad-div stabilisation and discontinuous Galerkin methods significantly improves the mass conservation properties and the overall accuracy of the numerical scheme which is observed for the first time. Hence, it is inferred that stabilised discontinuous Galerkin methods are highly robust as well as computationally efficient numerical methods to deal with natural convection problems arising in incompressible computational thermo-fluid dynamics.
Fabrication of compliant hybrid grafts supported with elastomeric meshes.
Kobashi, T; Matsuda, T
1999-01-01
We devised tubular hybrid medial tissues with mechanical properties similar to those of native arteries, which were composed of bovine smooth muscle cells (SMCs) and type I collagen with minimal reinforcement with knitted fabric meshes made of synthetic elastomers. Three hybrid medial tissue models that incorporated segmented polyester (mesh A) or polyurethane-nylon (mesh B) meshes were designed: the inner, sandwich, and wrapping models. Hybrid medial tissues were prepared by pouring a cold mixed solution of SMCs and collagen into a tubular glass mold consisting of an inner mandrel and an outer sheath and subsequent thermal gelation, followed by further culture for 7 days. For the inner model, the mandrel was wrapped with a mesh. For the sandwich model, a cylindrically shaped mesh was incorporated into a space between the mandrel and the sheath. The wrapping model was prepared by wrapping a 7-day-incubated nonmesh gel with a mesh. The inner diameter was 3 mm, irrespective of the model, and the length was 2.5-4.0 cm, depending on the model. The intraluminal pressure-external diameter relationship showed that nonmesh and inner models had a very low burst strength below 50 mmHg, while the sandwich model ruptured at around 110-120 mmHg; no rupturing below 240 mmHg was observed for the wrapping model, regardless of the type of mesh used. Compliance values of wrapping and sandwich models were close to those of native arteries. Pressure-dependent distensibility characteristics similar to native arteries were observed for a mesh A wrapping model, whereas a mesh B wrapping model expanded almost linearly as intraluminal pressure increased, which appeared to be due to elasticity of the incorporated mesh. Thus, design criteria for hybrid vascular grafts with appropriate biomechanical matching with host arteries were established. Such hybrid grafts may be mechanically adapted in an arterial system.
Dynamic modelling of flexibly supported gears using iterative convergence of tooth mesh stiffness
NASA Astrophysics Data System (ADS)
Xue, Song; Howard, Ian
2016-12-01
This paper presents a new gear dynamic model for flexibly supported gear sets aiming to improve the accuracy of gear fault diagnostic methods. In the model, the operating gear centre distance, which can affect the gear design parameters, like the gear mesh stiffness, has been selected as the iteration criteria because it will significantly deviate from its nominal value for a flexible supported gearset when it is operating. The FEA method was developed for calculation of the gear mesh stiffnesses with varying gear centre distance, which can then be incorporated by iteration into the gear dynamic model. The dynamic simulation results from previous models that neglect the operating gear centre distance change and those from the new model that incorporate the operating gear centre distance change were obtained by numerical integration of the differential equations of motion using the Newmark method. Some common diagnostic tools were utilized to investigate the difference and comparison of the fault diagnostic results between the two models. The results of this paper indicate that the major difference between the two diagnostic results for the cracked tooth exists in the extended duration of the crack event and in changes to the phase modulation of the coherent time synchronous averaged signal even though other notable differences from other diagnostic results can also be observed.
Pavarino, E.; Neves, L. A.; Machado, J. M.; de Godoy, M. F.; Shiyou, Y.; Momente, J. C.; Zafalon, G. F. D.; Pinto, A. R.; Valêncio, C. R.
2013-01-01
The Finite Element Method is a well-known technique, being extensively applied in different areas. Studies using the Finite Element Method (FEM) are targeted to improve cardiac ablation procedures. For such simulations, the finite element meshes should consider the size and histological features of the target structures. However, it is possible to verify that some methods or tools used to generate meshes of human body structures are still limited, due to nondetailed models, nontrivial preprocessing, or mainly limitation in the use condition. In this paper, alternatives are demonstrated to solid modeling and automatic generation of highly refined tetrahedral meshes, with quality compatible with other studies focused on mesh generation. The innovations presented here are strategies to integrate Open Source Software (OSS). The chosen techniques and strategies are presented and discussed, considering cardiac structures as a first application context. PMID:23762031
NASA Technical Reports Server (NTRS)
Wood, William A., III
2002-01-01
A multi-dimensional upwind fluctuation splitting scheme is developed and implemented for two-dimensional and axisymmetric formulations of the Navier-Stokes equations on unstructured meshes. Key features of the scheme are the compact stencil, full upwinding, and non-linear discretization which allow for second-order accuracy with enforced positivity. Throughout, the fluctuation splitting scheme is compared to a current state-of-the-art finite volume approach, a second-order, dual mesh upwind flux difference splitting scheme (DMFDSFV), and is shown to produce more accurate results using fewer computer resources for a wide range of test cases. A Blasius flat plate viscous validation case reveals a more accurate upsilon-velocity profile for fluctuation splitting, and the reduced artificial dissipation production is shown relative to DMFDSFV. Remarkably, the fluctuation splitting scheme shows grid converged skin friction coefficients with only five points in the boundary layer for this case. The second half of the report develops a local, compact, anisotropic unstructured mesh adaptation scheme in conjunction with the multi-dimensional upwind solver, exhibiting a characteristic alignment behavior for scalar problems. The adaptation strategy is extended to the two-dimensional and axisymmetric Navier-Stokes equations of motion through the concept of fluctuation minimization.
NASA Astrophysics Data System (ADS)
Gill, Stuart P. D.; Knebe, Alexander; Gibson, Brad K.; Flynn, Chris; Ibata, Rodrigo A.; Lewis, Geraint F.
2003-04-01
An adaptive multi grid approach to simulating the formation of structure from collisionless dark matter is described. MLAPM (Multi-Level Adaptive Particle Mesh) is one of the most efficient serial codes available on the cosmological "market" today. As part of Swinburne University's role in the development of the Square Kilometer Array, we are implementing hydrodynamics, feedback, and radiative transfer within the MLAPM adaptive mesh, in order to simulate baryonic processes relevant to the interstellar and intergalactic media at high redshift. We will outline our progress to date in applying the existing MLAPM to a study of the decay of satellite galaxies within massive host potentials.
NASA Astrophysics Data System (ADS)
Lenkiewicz, Przemyslaw; Pereira, Manuela; Freire, Mário M.; Fernandes, José
2013-12-01
In this article, we propose a novel image segmentation method called the whole mesh deformation (WMD) model, which aims at addressing the problems of modern medical imaging. Such problems have raised from the combination of several factors: (1) significant growth of medical image volumes sizes due to increasing capabilities of medical acquisition devices; (2) the will to increase the complexity of image processing algorithms in order to explore new functionality; (3) change in processor development and turn towards multi processing units instead of growing bus speeds and the number of operations per second of a single processing unit. Our solution is based on the concept of deformable models and is characterized by a very effective and precise segmentation capability. The proposed WMD model uses a volumetric mesh instead of a contour or a surface to represent the segmented shapes of interest, which allows exploiting more information in the image and obtaining results in shorter times, independently of image contents. The model also offers a good ability for topology changes and allows effective parallelization of workflow, which makes it a very good choice for large datasets. We present a precise model description, followed by experiments on artificial images and real medical data.
An Interpreted Language and System for the Visualization of Unstructured Meshes
NASA Technical Reports Server (NTRS)
Moran, Patrick J.; Gerald-Yamasaki, Michael (Technical Monitor)
1998-01-01
We present an interpreted language and system supporting the visualization of unstructured meshes and the manipulation of shapes defined in terms of mesh subsets. The language features primitives inspired by geometric modeling, mathematical morphology and algebraic topology. The adaptation of the topology ideas to an interpreted environment, along with support for programming constructs such, as user function definition, provide a flexible system for analyzing a mesh and for calculating with shapes defined in terms of the mesh. We present results demonstrating some of the capabilities of the language, based on an implementation called the Shape Calculator, for tetrahedral meshes in R^3.
TRANSL8GDECIM8. Data Translation and Filtering for Large 3D Triangle Mesh Models
Janucik, F.X.; Ross, D.M.
1993-09-01
The TRANSL8GDECIM8 system consists of two programs: TRANSL8G and DECIM8. The TRANSL8G program facilitates the interchange, topology generation, error checking, and enhancement of large 3D triangle meshes. Such data is frequently used to represent conceptual designs, scientific visualization volume modeling, or discrete sample data. Interchange is provided between several popular commercial and defacto standard geometry formats. Error checking is included to identify duplicate and zero area triangles. Model enhancement features include common vertex joining, consistent triangle vertex ordering, vertex normal vector averaging, and triangle strip generation. Many of the traditional O(n squared) algorithms required to provide the above features have been recast and are O(n) which support large mesh sizes. The DECIM8 program is based on a data filter algorithm that significantly reduces the number of triangles required to represent three dimensional (3D) models of geometry, scientific visualization results, and discretely sampled data. The algorithm uses a combined incremental and iterative strategy. It eliminates local patches of triangles whose geometries are not appreciably different and replaces them with fewer larger triangles. The algorithm has been used to reduce triangles in large conceptual design models to facilitate virtual walk throughs and to enable interactive viewing of large 3D iso-surface volume visualizations.
Generation of the 30 M-Mesh Global Digital Surface Model by Alos Prism
NASA Astrophysics Data System (ADS)
Tadono, T.; Nagai, H.; Ishida, H.; Oda, F.; Naito, S.; Minakawa, K.; Iwamoto, H.
2016-06-01
Topographical information is fundamental to many geo-spatial related information and applications on Earth. Remote sensing satellites have the advantage in such fields because they are capable of global observation and repeatedly. Several satellite-based digital elevation datasets were provided to examine global terrains with medium resolutions e.g. the Shuttle Radar Topography Mission (SRTM), the global digital elevation model by the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER GDEM). A new global digital surface model (DSM) dataset using the archived data of the Panchromatic Remote-sensing Instrument for Stereo Mapping (PRISM) onboard the Advanced Land Observing Satellite (ALOS, nicknamed "Daichi") has been completed on March 2016 by Japan Aerospace Exploration Agency (JAXA) collaborating with NTT DATA Corp. and Remote Sensing Technology Center, Japan. This project is called "ALOS World 3D" (AW3D), and its dataset consists of the global DSM dataset with 0.15 arcsec. pixel spacing (approx. 5 m mesh) and ortho-rectified PRISM image with 2.5 m resolution. JAXA is also processing the global DSM with 1 arcsec. spacing (approx. 30 m mesh) based on the AW3D DSM dataset, and partially releasing it free of charge, which calls "ALOS World 3D 30 m mesh" (AW3D30). The global AW3D30 dataset will be released on May 2016. This paper describes the processing status, a preliminary validation result of the AW3D30 DSM dataset, and its public release status. As a summary of the preliminary validation of AW3D30 DSM, 4.40 m (RMSE) of the height accuracy of the dataset was confirmed using 5,121 independent check points distributed in the world.
NASA Astrophysics Data System (ADS)
Paparella, F.; Oliveri, F.
2009-04-01
The interplay of advection, reaction and diffusion terms in ADR equations is a rather difficult one to be modeled numerically. The kind of spurious oscillations that is usually harmless for non-reacting scalars is often amplified without bounds by reaction terms. Furthermore, in most biogeochimical applications, such as mesoscale or global-scale plankton modeling, the diffusive fluxes may be smaller than the numerical ones. Inspired by the particle-mesh methods used by cosmologists, we propose to discretize on a grid only the diffusive term of the equation, and solve the advection-reaction terms as ordinary differential equations along the characteristic lines. Diffusion happens by letting the concentration field carried by each particle to relax towards the diffusive field known on the grid, without redistributing the particles. This method, in the limit of vanishing diffusivity and for a fixed mesh size, recovers the advection-reaction solution with no numerical diffusion. We show some example numerical solutions of the ADR equations stemming from a simple predator-prey model.
Documentation for MeshKit - Reactor Geometry (&mesh) Generator
Jain, Rajeev; Mahadevan, Vijay
2015-09-30
This report gives documentation for using MeshKit’s Reactor Geometry (and mesh) Generator (RGG) GUI and also briefly documents other algorithms and tools available in MeshKit. RGG is a program designed to aid in modeling and meshing of complex/large hexagonal and rectilinear reactor cores. RGG uses Argonne’s SIGMA interfaces, Qt and VTK to produce an intuitive user interface. By integrating a 3D view of the reactor with the meshing tools and combining them into one user interface, RGG streamlines the task of preparing a simulation mesh and enables real-time feedback that reduces accidental scripting mistakes that could waste hours of meshing. RGG interfaces with MeshKit tools to consolidate the meshing process, meaning that going from model to mesh is as easy as a button click. This report is designed to explain RGG v 2.0 interface and provide users with the knowledge and skills to pilot RGG successfully. Brief documentation of MeshKit source code, tools and other algorithms available are also presented for developers to extend and add new algorithms to MeshKit. RGG tools work in serial and parallel and have been used to model complex reactor core models consisting of conical pins, load pads, several thousands of axially varying material properties of instrumentation pins and other interstices meshes.
Adaptive moving finite volume scheme for flood inundation modeling under dry and complex topography
NASA Astrophysics Data System (ADS)
Zhou, F.; Chen, G.
2012-04-01
To assess and alleviate the risk of flood inundation on local scale, the use of numerical models with high accuracy, spatial resolution, and efficiency is crucial for the reliability of the solutions to provide the forecasts and early-warnings of flood inundation at large or meso-scales. Different with traditional numerical models on fixed meshes, an adaptive moving finite volume scheme on moving meshes is proposed for flood inundation modeling under dry and complex topography, this scheme aims to improve the predictive accuracy, spatial resolution, and computational efficiency as well as the satisfaction of well-balanced positivity preserving properties. The crucial feature of our scheme is to move fixed number of unstructured triangular meshes adaptively for approximating the time-variant patterns of flow variables and then to update flow variables through PDEs discretization on new meshes. At each time step of simulation, this scheme consists of three parts, giving in time n for instance: (1) adaptive mesh movement equation for adapting vertex from xij(n, v) to xij(n,v+1) where v is the iteration step, this equation can be transferred as Euler-Lagrange ones⛛· (ω⛛x) = 0, in which the monitor functionω is determined by the solution and the gradient of solution; (2) geometrical conservative interpolation for remapping flow variables from Ui(n, v) to Ui(n,v+1), when ||xij(n,v+1)-xij(n, v)||≤10-6 or v=5, then set xij(n, +∞):= xij(n,v+1) and Uj(n, +∞):= Uj(n,v+1), and (3) HLL-based PDEs discretization for updating flow variables from Ui(n,+∞) to Ui(n+1,0), the treatments of bed slope source terms and wet-dry interface are based on second-order reconstruction of Audusse et al., (2004) and Audusse and Bristeau (2005). Two analytical and two experimental test cases were performed to verify the advantages of the proposed scheme over non-adaptive methods. The results revealed two attractive features: (i) this scheme could achieve high-accuracy and high
NASA Astrophysics Data System (ADS)
Schartmann, M.; Ballone, A.; Burkert, A.; Gillessen, S.; Genzel, R.; Pfuhl, O.; Eisenhauer, F.; Plewa, P. M.; Ott, T.; George, E. M.; Habibi, M.
2015-10-01
The dusty, ionized gas cloud G2 is currently passing the massive black hole in the Galactic Center at a distance of roughly 2400 Schwarzschild radii. We explore the possibility of a starting point of the cloud within the disks of young stars. We make use of the large amount of new observations in order to put constraints on G2's origin. Interpreting the observations as a diffuse cloud of gas, we employ three-dimensional hydrodynamical adaptive mesh refinement (AMR) simulations with the PLUTO code and do a detailed comparison with observational data. The simulations presented in this work update our previously obtained results in multiple ways: (1) high resolution three-dimensional hydrodynamical AMR simulations are used, (2) the cloud follows the updated orbit based on the Brackett-γ data, (3) a detailed comparison to the observed high-quality position-velocity (PV) diagrams and the evolution of the total Brackett-γ luminosity is done. We concentrate on two unsolved problems of the diffuse cloud scenario: the unphysical formation epoch only shortly before the first detection and the too steep Brackett-γ light curve obtained in simulations, whereas the observations indicate a constant Brackett-γ luminosity between 2004 and 2013. For a given atmosphere and cloud mass, we find a consistent model that can explain both, the observed Brackett-γ light curve and the PV diagrams of all epochs. Assuming initial pressure equilibrium with the atmosphere, this can be reached for a starting date earlier than roughly 1900, which is close to apo-center and well within the disks of young stars.
Schartmann, M.; Ballone, A.; Burkert, A.; Gillessen, S.; Genzel, R.; Pfuhl, O.; Eisenhauer, F.; Plewa, P. M.; Ott, T.; George, E. M.; Habibi, M.
2015-10-01
The dusty, ionized gas cloud G2 is currently passing the massive black hole in the Galactic Center at a distance of roughly 2400 Schwarzschild radii. We explore the possibility of a starting point of the cloud within the disks of young stars. We make use of the large amount of new observations in order to put constraints on G2's origin. Interpreting the observations as a diffuse cloud of gas, we employ three-dimensional hydrodynamical adaptive mesh refinement (AMR) simulations with the PLUTO code and do a detailed comparison with observational data. The simulations presented in this work update our previously obtained results in multiple ways: (1) high resolution three-dimensional hydrodynamical AMR simulations are used, (2) the cloud follows the updated orbit based on the Brackett-γ data, (3) a detailed comparison to the observed high-quality position–velocity (PV) diagrams and the evolution of the total Brackett-γ luminosity is done. We concentrate on two unsolved problems of the diffuse cloud scenario: the unphysical formation epoch only shortly before the first detection and the too steep Brackett-γ light curve obtained in simulations, whereas the observations indicate a constant Brackett-γ luminosity between 2004 and 2013. For a given atmosphere and cloud mass, we find a consistent model that can explain both, the observed Brackett-γ light curve and the PV diagrams of all epochs. Assuming initial pressure equilibrium with the atmosphere, this can be reached for a starting date earlier than roughly 1900, which is close to apo-center and well within the disks of young stars.
Toward An Unstructured Mesh Database
NASA Astrophysics Data System (ADS)
Rezaei Mahdiraji, Alireza; Baumann, Peter Peter
2014-05-01
Unstructured meshes are used in several application domains such as earth sciences (e.g., seismology), medicine, oceanography, cli- mate modeling, GIS as approximate representations of physical objects. Meshes subdivide a domain into smaller geometric elements (called cells) which are glued together by incidence relationships. The subdivision of a domain allows computational manipulation of complicated physical structures. For instance, seismologists model earthquakes using elastic wave propagation solvers on hexahedral meshes. The hexahedral con- tains several hundred millions of grid points and millions of hexahedral cells. Each vertex node in the hexahedrals stores a multitude of data fields. To run simulation on such meshes, one needs to iterate over all the cells, iterate over incident cells to a given cell, retrieve coordinates of cells, assign data values to cells, etc. Although meshes are used in many application domains, to the best of our knowledge there is no database vendor that support unstructured mesh features. Currently, the main tool for querying and manipulating unstructured meshes are mesh libraries, e.g., CGAL and GRAL. Mesh li- braries are dedicated libraries which includes mesh algorithms and can be run on mesh representations. The libraries do not scale with dataset size, do not have declarative query language, and need deep C++ knowledge for query implementations. Furthermore, due to high coupling between the implementations and input file structure, the implementations are less reusable and costly to maintain. A dedicated mesh database offers the following advantages: 1) declarative querying, 2) ease of maintenance, 3) hiding mesh storage structure from applications, and 4) transparent query optimization. To design a mesh database, the first challenge is to define a suitable generic data model for unstructured meshes. We proposed ImG-Complexes data model as a generic topological mesh data model which extends incidence graph model to multi
An adaptive multigrid model for hurricane track prediction
NASA Technical Reports Server (NTRS)
Fulton, Scott R.
1993-01-01
This paper describes a simple numerical model for hurricane track prediction which uses a multigrid method to adapt the model resolution as the vortex moves. The model is based on the modified barotropic vorticity equation, discretized in space by conservative finite differences and in time by a Runge-Kutta scheme. A multigrid method is used to solve an elliptic problem for the streamfunction at each time step. Nonuniform resolution is obtained by superimposing uniform grids of different spatial extent; these grids move with the vortex as it moves. Preliminary numerical results indicate that the local mesh refinement allows accurate prediction of the hurricane track with substantially less computer time than required on a single uniform grid.
NASA Technical Reports Server (NTRS)
Turon, A.; Davila, C. G.; Camanho, P. P.; Costa, J.
2007-01-01
This paper presents a methodology to determine the parameters to be used in the constitutive equations of Cohesive Zone Models employed in the simulation of delamination in composite materials by means of decohesion finite elements. A closed-form expression is developed to define the stiffness of the cohesive layer. A novel procedure that allows the use of coarser meshes of decohesion elements in large-scale computations is also proposed. The procedure ensures that the energy dissipated by the fracture process is computed correctly. It is shown that coarse-meshed models defined using the approach proposed here yield the same results as the models with finer meshes normally used for the simulation of fracture processes.
Attwood, S E; Caldwell, M T; Marks, P; McDermott, M; Stephens, R B
1994-07-01
Prosthetic mesh for laparoscopic inguinal hernia repair has become popular but the method of its placement is controversial. Mesh placed within the peritoneum may cause adhesion formation and further complications. The aim of this study was to examine the laparoscopic placement of a mesh, comparing intraperitoneal vs extraperitoneal insertion. In a porcine model (n = 15) a polypropylene mesh was placed laparoscopically over the anterior abdominal wall. On the left side the mesh was stapled on the parietal peritoneum. On the right side the peritoneum was incised, an extraperitoneal space was dissected, the mesh was inserted, and the peritoneum was closed over it. The animals were maintained for 2 weeks. At postmortem there were adhesions in two of those placed extraperitoneally and five of those placed intraperitoneally (P = 0.19, Fisher's exact test). The adhesions comprised fibrous peritoneal bands to loops of small intestine. Both methods of laparoscopic mesh placement were associated with a small but significant incidence of adhesion formation.
NASA Astrophysics Data System (ADS)
Gansen, A.; Hachemi, M. El; Belouettar, S.; Hassan, O.; Morgan, K.
2016-09-01
The standard Yee algorithm is widely used in computational electromagnetics because of its simplicity and divergence free nature. A generalization of the classical Yee scheme to 3D unstructured meshes is adopted, based on the use of a Delaunay primal mesh and its high quality Voronoi dual. This allows the problem of accuracy losses, which are normally associated with the use of the standard Yee scheme and a staircased representation of curved material interfaces, to be circumvented. The 3D dual mesh leapfrog-scheme which is presented has the ability to model both electric and magnetic anisotropic lossy materials. This approach enables the modelling of problems, of current practical interest, involving structured composites and metamaterials.
Towards Spherical Mesh Gravity and Magnetic Modelling in an HPC Environment
NASA Astrophysics Data System (ADS)
Lane, R. J.; Brodie, R. C.; de Hoog, M.; Navin, J.; Chen, C.; Du, J.; Liang, Q.; Wang, H.; Li, Y.
2013-12-01
Staff at Geoscience Australia (GA), Australia's Commonwealth Government geoscientific agency, have routinely performed 3D gravity and magnetic modelling as part of geoscience investigations. For this work, we have used software programs that have been based on a Cartesian mesh spatial framework. These programs have come as executable files that were compiled to operate in a Windows environment on single core personal computers (PCs). To cope with models with higher resolution and larger extents, we developed an approach whereby a large problem could be broken down into a number of overlapping smaller models (';tiles') that could be modelled separately, with the results combined back into a single output model. To speed up the processing, we established a Condor distributed network from existing desktop PCs. A number of factors have caused us to consider a new approach to this modelling work. The drivers for change include; 1) models with very large lateral extents where the effects of Earth curvature are a consideration, 2) a desire to ensure that the modelling of separate regions is carried out in a consistent and managed fashion, 3) migration of scientific computing to off-site High Performance Computing (HPC) facilities, and 4) development of virtual globe environments for integration and visualization of 3D spatial objects. Some of the more surprising realizations to emerge have been that; 1) there aren't any readily available commercial software packages for modelling gravity and magnetic data in a spherical mesh spatial framework, 2) there are many different types of HPC environments, 3) no two HPC environments are the same, and 4) the most common virtual globe environment (i.e., Google Earth) doesn't allow spatial objects to be displayed below the topographic/bathymetric surface. Our response has been to do the following; 1) form a collaborative partnership with researchers at the Colorado School of Mines (CSM) and the China University of Geosciences (CUG
Multigrid solution of compressible turbulent flow on unstructured meshes using a two-equation model
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.; Matinelli, L.
1994-01-01
The steady state solution of the system of equations consisting of the full Navier-Stokes equations and two turbulence equations has been obtained using a multigrid strategy of unstructured meshes. The flow equations and turbulence equations are solved in a loosely coupled manner. The flow equations are advanced in time using a multistage Runge-Kutta time-stepping scheme with a stability-bound local time step, while turbulence equations are advanced in a point-implicit scheme with a time step which guarantees stability and positivity. Low-Reynolds-number modifications to the original two-equation model are incorporated in a manner which results in well-behaved equations for arbitrarily small wall distances. A variety of aerodynamic flows are solved, initializing all quantities with uniform freestream values. Rapid and uniform convergence rates for the flow and turbulence equations are observed.
Goal-oriented model adaptivity for viscous incompressible flows
NASA Astrophysics Data System (ADS)
van Opstal, T. M.; Bauman, P. T.; Prudhomme, S.; van Brummelen, E. H.
2015-06-01
In van Opstal et al. (Comput Mech 50:779-788, 2012) airbag inflation simulations were performed where the flow was approximated by Stokes flow. Inside the intricately folded initial geometry the Stokes assumption is argued to hold. This linearity assumption leads to a boundary-integral representation, the key to bypassing mesh generation and remeshing. It therefore enables very large displacements with near-contact. However, such a coarse assumption cannot hold throughout the domain, where it breaks down one needs to revert to the original model. The present work formalizes this idea. A model adaptive approach is proposed, in which the coarse model (a Stokes boundary-integral equation) is locally replaced by the original high-fidelity model (Navier-Stokes) based on a-posteriori estimates of the error in a quantity of interest. This adaptive modeling framework aims at taking away the burden and heuristics of manually partitioning the domain while providing new insight into the physics. We elucidate how challenges pertaining to model disparity can be addressed. Essentially, the solution in the interior of the coarse model domain is reconstructed as a post-processing step. We furthermore present a two-dimensional numerical experiments to show that the error estimator is reliable.
Verification Test of the SURF and SURFplus Models in xRage: Part III Affect of Mesh Alignment
Menikoff, Ralph
2016-08-15
The previous studies used an underdriven detonation wave in 1-dimension (steady ZND reaction zone profile followed by a scale-invariant rarefaction wave) for PBX 9502 as a verification test of the implementation of the SURF and SURFplus models in the xRage code. Since the SURF rate is a function of the lead shock pressure, the question arises as to the effect on accuracy of variations in the detected shock pressure due to the alignment of the shock front with the mesh. To study the effect of mesh alignment we simulate a cylindrically diverging detonation wave using a planar 2-D mesh. The leading issue is the magnitude of azimuthal asymmetries in the numerical solution. The 2-D test case does not have an exact analytic solution. To quantify the accuracy, the 2-D solution along rays through the origin are compared to a highly resolved 1-D simulation in cylindrical geometry.
Biomimetic collagen/elastin meshes for ventral hernia repair in a rat model.
Minardi, Silvia; Taraballi, Francesca; Wang, Xin; Cabrera, Fernando J; Van Eps, Jeffrey L; Robbins, Andrew B; Sandri, Monica; Moreno, Michael R; Weiner, Bradley K; Tasciotti, Ennio
2017-03-01
Ventral hernia repair remains a major clinical need. Herein, we formulated a type I collagen/elastin crosslinked blend (CollE) for the fabrication of biomimetic meshes for ventral hernia repair. To evaluate the effect of architecture on the performance of the implants, CollE was formulated both as flat sheets (CollE Sheets) and porous scaffolds (CollE Scaffolds). The morphology, hydrophylicity and in vitro degradation were assessed by SEM, water contact angle and differential scanning calorimetry, respectively. The stiffness of the meshes was determined using a constant stretch rate uniaxial tensile test, and compared to that of native tissue. CollE Sheets and Scaffolds were tested in vitro with human bone marrow-derived mesenchymal stem cells (h-BM-MSC), and finally implanted in a rat ventral hernia model. Neovascularization and tissue regeneration within the implants was evaluated at 6weeks, by histology, immunofluorescence, and q-PCR. It was found that CollE Sheets and Scaffolds were not only biomechanically sturdy enough to provide immediate repair of the hernia defect, but also promoted tissue restoration in only 6weeks. In fact, the presence of elastin enhanced the neovascularization in both sheets and scaffolds. Overall, CollE Scaffolds displayed mechanical properties more closely resembling those of native tissue, and induced higher gene expression of the entire marker genes tested, associated with de novo matrix deposition, angiogenesis, adipogenesis and skeletal muscles, compared to CollE Sheets. Altogether, this data suggests that the improved mechanical properties and bioactivity of CollE Sheets and Scaffolds make them valuable candidates for applications of ventral hernia repair.
Unstructured-mesh modeling of the Congo river-to-sea continuum
NASA Astrophysics Data System (ADS)
Bars, Yoann Le; Vallaeys, Valentin; Deleersnijder, Éric; Hanert, Emmanuel; Carrere, Loren; Channelière, Claire
2016-04-01
With the second largest outflow in the world and one of the widest hydrological basins, the Congo River is of a major importance both locally and globally. However, relatively few studies have been conducted on its hydrology, as compared to other great rivers such as the Amazon, Nile, Yangtze, or Mississippi. The goal of this study is therefore to help fill this gap and provide the first high-resolution simulation of the Congo river-estuary-coastal sea continuum. To this end, we are using a discontinuous-Galerkin finite element marine model that solves the two-dimensional depth-averaged shallow water equations on an unstructured mesh. To ensure a smooth transition from river to coastal sea, we have considered a model that encompasses both hydrological and coastal ocean processes. An important difficulty in setting up this model was to find data to parameterize and validate it, as it is a rather remote and understudied area. Therefore, an important effort in this study has been to establish a methodology to take advantage of all the data sources available including nautical charts that had to be digitalized. The model surface elevation has then been validated with respect to an altimetric database. Model results suggest the existence of gyres in the vicinity of the river mouth that have never been documented before. The effect of those gyres on the Congo River dynamics has been further investigated by simulating the transport of Lagrangian particles and computing the water age.
Multigrid techniques for unstructured meshes
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1995-01-01
An overview of current multigrid techniques for unstructured meshes is given. The basic principles of the multigrid approach are first outlined. Application of these principles to unstructured mesh problems is then described, illustrating various different approaches, and giving examples of practical applications. Advanced multigrid topics, such as the use of algebraic multigrid methods, and the combination of multigrid techniques with adaptive meshing strategies are dealt with in subsequent sections. These represent current areas of research, and the unresolved issues are discussed. The presentation is organized in an educational manner, for readers familiar with computational fluid dynamics, wishing to learn more about current unstructured mesh techniques.
Improving Tropical Cyclone Track and Intensity in a Global Model with Local Mesh Refinement
NASA Astrophysics Data System (ADS)
Zarzycki, C. M.; Jablonowski, C.
2014-12-01
Even with recent improvements in general circulation model (GCM) resolution, tropical cyclones (TCs) are typically underresolved, resulting in fewer or weaker storms than observed. In an effort to alleviate these issues, the use of limited area models (LAMs) allowing for higher resolutions has become popular. However, LAMs require lateral boundary conditions and typically lack two-way communication with the exterior domain. Variable-resolution GCMs can serve as the bridge between traditional global models and high-resolution LAMs. These models can reach 10 km or finer resolution in low-latitude ocean basins where TCs are prevalent. They do so while maintaining global continuity, therefore eliminating the need for externally-forced and possibly numerically and physically inconsistent boundary conditions required by LAMs. Recent developments allow the Community Atmosphere Model's (CAM) Spectral Element (SE) dynamical core to be run on unstructured, statically-nested, variable-resolution grids. We present deterministic CAM-SE model simulations of TCs during recent summers and compare the model's prediction of storm track and intensity to other global and regional models as well as observations. The simulations are run on a 55 km global cubed-sphere grid with additional refinement to 13 km over the Atlantic and Eastern Pacific Oceans. Forecasts are integrated for eight days and the period of analysis spans three months (August, September, and October) during 2012 and 2013. We compare these simulations to identically initialized model runs without mesh refinement to demonstrate the impact of high resolution on TC behavior in CAM. We also investigate cyclone genesis and whether locally high resolution in a global model leads to improved forecast skill at longer lead times. In addition, the impact of the localized refined patch on the remainder of the coarser global solution during the simulation period is discussed.
NASA Technical Reports Server (NTRS)
Malek, Miroslaw; Ozden, Banu
1990-01-01
Efficient testing techniques for two-dimensional mesh interconnection networks are presented. The tests cover faults in the arbitration logic of the switches; this includes an examination of fault detection in the data paths, routing, and control circuitry, including the conflict resolution capabilities of mesh interconnection networks using topological test methods. The proposed methods are not implementation specific and can be applied to any design with a mesh topology. The topology and behavior of the network are described and definitions are presented. The fault model is defined and parallel testing methods for the entire network are given.
An Adaptive Critic Approach to Reference Model Adaptation
NASA Technical Reports Server (NTRS)
Krishnakumar, K.; Limes, G.; Gundy-Burlet, K.; Bryant, D.
2003-01-01
Neural networks have been successfully used for implementing control architectures for different applications. In this work, we examine a neural network augmented adaptive critic as a Level 2 intelligent controller for a C- 17 aircraft. This intelligent control architecture utilizes an adaptive critic to tune the parameters of a reference model, which is then used to define the angular rate command for a Level 1 intelligent controller. The present architecture is implemented on a high-fidelity non-linear model of a C-17 aircraft. The goal of this research is to improve the performance of the C-17 under degraded conditions such as control failures and battle damage. Pilot ratings using a motion based simulation facility are included in this paper. The benefits of using an adaptive critic are documented using time response comparisons for severe damage situations.
Mesh-based Monte Carlo code for fluorescence modeling in complex tissues with irregular boundaries
NASA Astrophysics Data System (ADS)
Wilson, Robert H.; Chen, Leng-Chun; Lloyd, William; Kuo, Shiuhyang; Marcelo, Cynthia; Feinberg, Stephen E.; Mycek, Mary-Ann
2011-07-01
There is a growing need for the development of computational models that can account for complex tissue morphology in simulations of photon propagation. We describe the development and validation of a user-friendly, MATLAB-based Monte Carlo code that uses analytically-defined surface meshes to model heterogeneous tissue geometry. The code can use information from non-linear optical microscopy images to discriminate the fluorescence photons (from endogenous or exogenous fluorophores) detected from different layers of complex turbid media. We present a specific application of modeling a layered human tissue-engineered construct (Ex Vivo Produced Oral Mucosa Equivalent, EVPOME) designed for use in repair of oral tissue following surgery. Second-harmonic generation microscopic imaging of an EVPOME construct (oral keratinocytes atop a scaffold coated with human type IV collagen) was employed to determine an approximate analytical expression for the complex shape of the interface between the two layers. This expression can then be inserted into the code to correct the simulated fluorescence for the effect of the irregular tissue geometry.
Predictor-Based Model Reference Adaptive Control
NASA Technical Reports Server (NTRS)
Lavretsky, Eugene; Gadient, Ross; Gregory, Irene M.
2009-01-01
This paper is devoted to robust, Predictor-based Model Reference Adaptive Control (PMRAC) design. The proposed adaptive system is compared with the now-classical Model Reference Adaptive Control (MRAC) architecture. Simulation examples are presented. Numerical evidence indicates that the proposed PMRAC tracking architecture has better than MRAC transient characteristics. In this paper, we presented a state-predictor based direct adaptive tracking design methodology for multi-input dynamical systems, with partially known dynamics. Efficiency of the design was demonstrated using short period dynamics of an aircraft. Formal proof of the reported PMRAC benefits constitute future research and will be reported elsewhere.
A diversified portfolio model of adaptability.
Chandra, Siddharth; Leong, Frederick T L
2016-12-01
A new model of adaptability, the diversified portfolio model (DPM) of adaptability, is introduced. In the 1950s, Markowitz developed the financial portfolio model by demonstrating that investors could optimize the ratio of risk and return on their portfolios through risk diversification. The DPM integrates attractive features of a variety of models of adaptability, including Linville's self-complexity model, the risk and resilience model, and Bandura's social cognitive theory. The DPM draws on the concept of portfolio diversification, positing that diversified investment in multiple life experiences, life roles, and relationships promotes positive adaptation to life's challenges. The DPM provides a new integrative model of adaptability across the biopsychosocial levels of functioning. More importantly, the DPM addresses a gap in the literature by illuminating the antecedents of adaptive processes studied in a broad array of psychological models. The DPM is described in relation to the biopsychosocial model and propositions are offered regarding its utility in increasing adaptiveness. Recommendations for future research are also offered. (PsycINFO Database Record
Winny, Markus; Maegel, Lavinia; Grethe, Leonie; Lippmann, Torsten; Jonigk, Danny; Schrem, Harald; Kaltenborn, Alexander; Klempnauer, Juergen; Poehnert, Daniel
2016-01-01
Background: Adhesions to intraperitoneally implanted meshes (IPOM) are a common problem following hernia surgery and may cause severe complications. Recently, we showed that missing peritoneal coverage of the intestine is a decisive factor for adhesion formation and 4DryField® PH (4DF) gel significantly prevents intestine-to-mesh adhesions even with use of uncoated Ultrapro® polypropylene mesh (UPM). The present study investigates adhesion prevention capability of coated Parietex® mesh (PTM) and Proceed® mesh (PCM) in comparison to 4DF treated UPM. Methods: 20 rats were randomized into two groups. A 1.5 x 2 cm patch of PTM or PCM was attached to the abdominal wall and the cecum was depleted from peritoneum by abrasion. After seven days incidence of intestine-to-mesh adhesions was evaluated using Lauder and Hoffmann adhesion scores. Histological specimens were evaluated; statistics were performed using student's t-test. The data were compared with recently published data of 4DF treated uncoated UPM. Results: Use of PTM or PCM did not significantly diminish development of intestine-to-mesh adhesions (adhesion reduction rate PTM: 29%, p = 0.069 and PCM: 25%, p = 0.078). Histological results confirmed macroscopic finding of agglutination of intestine and abdominal wall with the mesh in between. Compared to these data, the use of UPM combined with 4DF gel reveals significantly better adhesion prevention capability (p < 0.0001) as shown in earlier studies. However, in clinical situation interindividual differences in adhesion induction mechanisms cannot be excluded by this experimental approach as healing responses towards the different materials might vary. Conclusion: This study shows that in case of impaired intestinal peritoneum coated PTM and PCM do not provide significant adhesion prevention. In contrast, use of UPM combined with 4DF gel achieved a significant reduction of adhesions. Hence, in case of injury of the visceral peritoneum, application of a
Winny, Markus; Maegel, Lavinia; Grethe, Leonie; Lippmann, Torsten; Jonigk, Danny; Schrem, Harald; Kaltenborn, Alexander; Klempnauer, Juergen; Poehnert, Daniel
2016-01-01
Background: Adhesions to intraperitoneally implanted meshes (IPOM) are a common problem following hernia surgery and may cause severe complications. Recently, we showed that missing peritoneal coverage of the intestine is a decisive factor for adhesion formation and 4DryField(®) PH (4DF) gel significantly prevents intestine-to-mesh adhesions even with use of uncoated Ultrapro(®) polypropylene mesh (UPM). The present study investigates adhesion prevention capability of coated Parietex(®) mesh (PTM) and Proceed(®) mesh (PCM) in comparison to 4DF treated UPM. Methods: 20 rats were randomized into two groups. A 1.5 x 2 cm patch of PTM or PCM was attached to the abdominal wall and the cecum was depleted from peritoneum by abrasion. After seven days incidence of intestine-to-mesh adhesions was evaluated using Lauder and Hoffmann adhesion scores. Histological specimens were evaluated; statistics were performed using student's t-test. The data were compared with recently published data of 4DF treated uncoated UPM. Results: Use of PTM or PCM did not significantly diminish development of intestine-to-mesh adhesions (adhesion reduction rate PTM: 29%, p = 0.069 and PCM: 25%, p = 0.078). Histological results confirmed macroscopic finding of agglutination of intestine and abdominal wall with the mesh in between. Compared to these data, the use of UPM combined with 4DF gel reveals significantly better adhesion prevention capability (p < 0.0001) as shown in earlier studies. However, in clinical situation interindividual differences in adhesion induction mechanisms cannot be excluded by this experimental approach as healing responses towards the different materials might vary. Conclusion: This study shows that in case of impaired intestinal peritoneum coated PTM and PCM do not provide significant adhesion prevention. In contrast, use of UPM combined with 4DF gel achieved a significant reduction of adhesions. Hence, in case of injury of the visceral peritoneum, application of a
NASA Astrophysics Data System (ADS)
Wang, Hairen; Lou, Zheng; Qian, Yuan; Zheng, Xianzhong; Zuo, Yingxi
2016-03-01
The optimization of a primary mirror support system is one of the most critical problems in the design of large telescopes. Here, we propose a hybrid optimization methodology of variable densities mesh model (HOMVDMM) for the axial supporting design, which has three key steps: (1) creating a variable densities mesh model, which will partition the mirror into several sparse mesh areas and several dense mesh areas; (2) global optimization based on the zero-order optimization method for the support of primary mirror with a large tolerance; (3) based on the optimization results of the second step, further optimization with first-order optimization method in dense mesh areas by a small tolerance. HOMVDMM exploits the complementary merits of both the zero- and first-order optimizations, with the former in global scale and the latter in small scale. As an application, the axial support of the primary mirror of the 2.5-m wide-field survey telescope (WFST) is optimized by HOMVDMM. These three designs are obtained via a comparative study of different supporting points including 27 supporting points, 39 supporting points, and 54 supporting points. Their residual half-path length errors are 28.78, 9.32, and 5.29 nm. The latter two designs both meet the specification of WFST. In each of the three designs, a global optimization value with high accuracy will be obtained in an hour on an ordinary PC. As the results suggest, the overall performance of HOMVDMM is superior to the first-order optimization method as well as the zero-order optimization method.
An adaptive finite element approach to modelling sediment laden density currents
NASA Astrophysics Data System (ADS)
Parkinson, S.; Hill, J.; Allison, P. A.; Piggott, M. D.
2012-04-01
Modelling sediment-laden density currents at real-world scales is a challenging task. Here we present Fluidity, which uses dynamic adaptive re-meshing to reduce computational costs whilst maintaining sufficient resolution where and when it is required. This allows small-scale processes to be captured in large scale simulations. Density currents, also known as gravity or buoyancy currents, occur wherever two fluids with different densities meet. They can occur at scales of up to hundred kilometres in the ocean when continental shelves collapse. This process releases large quantities of sediment into the ocean which increase the bulk density of the fluid to form a density current. These currents can carry sediment hundreds of kilometres, at speeds of up to a hundred kilometres per hour, over the sea bed. They can be tsunamigenic and they have the potential to cause significant damage to submarine infrastructure, such as submarine telecommunications cables or oil and gas infrastructure. They are also a key process for movement of organic material into the depths of the ocean. Due to this, they play an important role in the global carbon cycle on the Earth, forming a significant component of the stratigraphic record, and their deposits can form useful sources of important hydrocarbons. Modelling large scale sediment laden density currents is a very challenging problem. Particles within the current are suspended by turbulence that occurs at length scales that are several orders of magnitude smaller than the size of the current. Models that resolve the vertical structure of the flow require a very large, highly resolved mesh, and substantial computing power to solve. Here, we verify our adaptive model by comparison with a set of laboratory experiments by Gladstone et al. [1998] on the propagation and sediment deposition of bidisperse gravity currents. Comparisons are also made with fixed mesh solutions, and it is shown that accuracy can be maintained with fewer elements
On Fractional Model Reference Adaptive Control
Shi, Bao; Dong, Chao
2014-01-01
This paper extends the conventional Model Reference Adaptive Control systems to fractional ones based on the theory of fractional calculus. A control law and an incommensurate fractional adaptation law are designed for the fractional plant and the fractional reference model. The stability and tracking convergence are analyzed using the frequency distributed fractional integrator model and Lyapunov theory. Moreover, numerical simulations of both linear and nonlinear systems are performed to exhibit the viability and effectiveness of the proposed methodology. PMID:24574897
Guzik, S; McCorquodale, P; Colella, P
2011-12-16
A fourth-order accurate finite-volume method is presented for solving time-dependent hyperbolic systems of conservation laws on mapped grids that are adaptively refined in space and time. Novel considerations for formulating the semi-discrete system of equations in computational space combined with detailed mechanisms for accommodating the adapting grids ensure that conservation is maintained and that the divergence of a constant vector field is always zero (freestream-preservation property). Advancement in time is achieved with a fourth-order Runge-Kutta method.
Campbell, J Q; Coombs, D J; Rao, M; Rullkoetter, P J; Petrella, A J
2016-09-06
The purpose of this study was to seek broad verification and validation of human lumbar spine finite element models created using a previously published automated algorithm. The automated algorithm takes segmented CT scans of lumbar vertebrae, automatically identifies important landmarks and contact surfaces, and creates a finite element model. Mesh convergence was evaluated by examining changes in key output variables in response to mesh density. Semi-direct validation was performed by comparing experimental results for a single specimen to the automated finite element model results for that specimen with calibrated material properties from a prior study. Indirect validation was based on a comparison of results from automated finite element models of 18 individual specimens, all using one set of generalized material properties, to a range of data from the literature. A total of 216 simulations were run and compared to 186 experimental data ranges in all six primary bending modes up to 7.8Nm with follower loads up to 1000N. Mesh convergence results showed less than a 5% difference in key variables when the original mesh density was doubled. The semi-direct validation results showed that the automated method produced results comparable to manual finite element modeling methods. The indirect validation results showed a wide range of outcomes due to variations in the geometry alone. The studies showed that the automated models can be used to reliably evaluate lumbar spine biomechanics, specifically within our intended context of use: in pure bending modes, under relatively low non-injurious simulated in vivo loads, to predict torque rotation response, disc pressures, and facet forces.
Multiresolution mesh segmentation based on surface roughness and wavelet analysis
NASA Astrophysics Data System (ADS)
Roudet, Céline; Dupont, Florent; Baskurt, Atilla
2007-01-01
During the last decades, the three-dimensional objects have begun to compete with traditional multimedia (images, sounds and videos) and have been used by more and more applications. The common model used to represent them is a surfacic mesh due to its intrinsic simplicity and efficacity. In this paper, we present a new algorithm for the segmentation of semi-regular triangle meshes, via multiresolution analysis. Our method uses several measures which reflect the roughness of the surface for all meshes resulting from the decomposition of the initial model into different fine-to-coarse multiresolution meshes. The geometric data decomposition is based on the lifting scheme. Using that formulation, we have compared various interpolant prediction operators, associated or not with an update step. For each resolution level, the resulting approximation mesh is then partitioned into classes having almost constant roughness thanks to a clustering algorithm. Resulting classes gather regions having the same visual appearance in term of roughness. The last step consists in decomposing the mesh into connex groups of triangles using region growing ang merging algorithms. These connex surface patches are of particular interest for adaptive mesh compression, visualisation, indexation or watermarking.
NASA Astrophysics Data System (ADS)
Jackman, Kevin Richard
Traditional approaches in gamma-ray spectroscopy for determining the absolute full-energy peak efficiencies of germanium detectors are primarily either too time consuming or not economically viable. In addition, these approaches are difficult to use for arbitrary source shapes and counting geometries. An open source software package, KMESS (Kevin's Mesh Efficiency Simulator Software), was developed to address these problems. KMESS uses a new semi-empirical mesh-grid method to predict the absolute full-energy peak efficiencies of n- and p-type germanium detectors in both coaxial and closed-ended configurations. The model assumes that any gamma-ray source shape can be treated as a collection of point sources. The code was written in a modular form, making it easy to adapt for other detector configurations and materials. A suite of web-based graphical front-end tools was also developed to make the execution of KMESS user-friendly. KMESS can predict most full-energy peak efficiencies to within 10% accuracy for the energy range 100--1800 keV in less than 10 minutes.
NASA Astrophysics Data System (ADS)
Iarve, E. V.; Mollenhauer, D.; Kim, R.
2004-09-01
A three-dimensional ply-level modeling of multiple matrix cracking near an open hole in a quasi-isotropic composite laminate was performed. A mesh-independent displacement discontinuity modeling method based on higher-order shape functions was constructed for this purpose. The mesh configuration is dictated by the boundaries of a specimen, such as the presence of a hole, whereas the matrix cracking surfaces are aligned with the fiber direction in a given ply. The surface of the displacement jump associated with matrix cracking was defined in terms of the domain Heaviside function approximated by using higher-order polynomial B-splines. Several matrix cracks in each ply of a [0/45/90/-45] s composite were modeled, and their effect on the fiber-direction stress magnitude in the 0° ply was examined. Up to 35% relaxation of the fiber-direction strain amplitude due to matrix cracking (splitting) in the 0° ply was predicted. The moir? interferometry was used to experimentally determine the strain and displacement fields in the surface layer of the same composite, previously prestressed beyond the damage initiation load. A good correlation between the experimental data and the stress redistribution predicted by the mesh-independent damage modeling technique was observed.
Toward Interoperable Mesh, Geometry and Field Components for PDE Simulation Development
Chand, K K; Diachin, L F; Li, X; Ollivier-Gooch, C; Seol, E S; Shephard, M; Tautges, T; Trease, H
2005-07-11
Mesh-based PDE simulation codes are becoming increasingly sophisticated and rely on advanced meshing and discretization tools. Unfortunately, it is still difficult to interchange or interoperate tools developed by different communities to experiment with various technologies or to develop new capabilities. To address these difficulties, we have developed component interfaces designed to support the information flow of mesh-based PDE simulations. We describe this information flow and discuss typical roles and services provided by the geometry, mesh, and field components of the simulation. Based on this delineation for the roles of each component, we give a high-level description of the abstract data model and set of interfaces developed by the Department of Energy's Interoperable Tools for Advanced Petascale Simulation (ITAPS) center. These common interfaces are critical to our interoperability goal, and we give examples of several services based upon these interfaces including mesh adaptation and mesh improvement.
NASA Astrophysics Data System (ADS)
Dönmez, Orhan
2004-09-01
In this paper, the general procedure to solve the general relativistic hydrodynamical (GRH) equations with adaptive-mesh refinement (AMR) is presented. In order to achieve, the GRH equations are written in the conservation form to exploit their hyperbolic character. The numerical solutions of GRH equations are obtained by high resolution shock Capturing schemes (HRSC), specifically designed to solve nonlinear hyperbolic systems of conservation laws. These schemes depend on the characteristic information of the system. The Marquina fluxes with MUSCL left and right states are used to solve GRH equations. First, different test problems with uniform and AMR grids on the special relativistic hydrodynamics equations are carried out to verify the second-order convergence of the code in one, two and three dimensions. Results from uniform and AMR grid are compared. It is found that adaptive grid does a better job when the number of resolution is increased. Second, the GRH equations are tested using two different test problems which are Geodesic flow and Circular motion of particle In order to do this, the flux part of GRH equations is coupled with source part using Strang splitting. The coupling of the GRH equations is carried out in a treatment which gives second order accurate solutions in space and time.
Graphical Models and Computerized Adaptive Testing.
ERIC Educational Resources Information Center
Almond, Russell G.; Mislevy, Robert J.
1999-01-01
Considers computerized adaptive testing from the perspective of graphical modeling (GM). GM provides methods for making inferences about multifaceted skills and knowledge and for extracting data from complex performances. Provides examples from language-proficiency assessment. (SLD)
Tangle-Free Mesh Motion for Ablation Simulations
NASA Technical Reports Server (NTRS)
Droba, Justin
2016-01-01
Problems involving mesh motion-which should not be mistakenly associated with moving mesh methods, a class of adaptive mesh redistribution techniques-are of critical importance in numerical simulations of the thermal response of melting and ablative materials. Ablation is the process by which material vaporizes or otherwise erodes due to strong heating. Accurate modeling of such materials is of the utmost importance in design of passive thermal protection systems ("heatshields") for spacecraft, the layer of the vehicle that ensures survival of crew and craft during re-entry. In an explicit mesh motion approach, a complete thermal solve is first performed. Afterwards, the thermal response is used to determine surface recession rates. These values are then used to generate boundary conditions for an a posteriori correction designed to update the location of the mesh nodes. Most often, linear elastic or biharmonic equations are used to model this material response, traditionally in a finite element framework so that complex geometries can be simulated. A simple scheme for moving the boundary nodes involves receding along the surface normals. However, for all but the simplest problem geometries, evolution in time following such a scheme will eventually bring the mesh to intersect and "tangle" with itself, inducing failure. This presentation demonstrates a comprehensive and sophisticated scheme that analyzes the local geometry of each node with help from user-provided clues to eliminate the tangle and enable simulations on a wide-class of difficult problem geometries. The method developed is demonstrated for linear elastic equations but is general enough that it may be adapted to other modeling equations. The presentation will explicate the inner workings of the tangle-free mesh motion algorithm for both two and three-dimensional meshes. It will show abstract examples of the method's success, including a verification problem that demonstrates its accuracy and
Ziegelwanger, Harald; Majdak, Piotr; Kreuzer, Wolfgang
2015-07-01
Head-related transfer functions (HRTFs) can be numerically calculated by applying the boundary element method on the geometry of a listener's head and pinnae. The calculation results are defined by geometrical, numerical, and acoustical parameters like the microphone used in acoustic measurements. The scope of this study was to estimate requirements on the size and position of the microphone model and on the discretization of the boundary geometry as triangular polygon mesh for accurate sound localization. The evaluation involved the analysis of localization errors predicted by a sagittal-plane localization model, the comparison of equivalent head radii estimated by a time-of-arrival model, and the analysis of actual localization errors obtained in a sound-localization experiment. While the average edge length (AEL) of the mesh had a negligible effect on localization performance in the lateral dimension, the localization performance in sagittal planes, however, degraded for larger AELs with the geometrical error as dominant factor. A microphone position at an arbitrary position at the entrance of the ear canal, a microphone size of 1 mm radius, and a mesh with 1 mm AEL yielded a localization performance similar to or better than observed with acoustically measured HRTFs.
User Modeling in Adaptive Hypermedia Educational Systems
ERIC Educational Resources Information Center
Martins, Antonio Constantino; Faria, Luiz; Vaz de Carvalho, Carlos; Carrapatoso, Eurico
2008-01-01
This document is a survey in the research area of User Modeling (UM) for the specific field of Adaptive Learning. The aims of this document are: To define what it is a User Model; To present existing and well known User Models; To analyze the existent standards related with UM; To compare existing systems. In the scientific area of User Modeling…
Designing Interference-Robust Wireless Mesh Networks Using a Defender-Attacker-Defender Model
2015-02-01
design. vi THIS PAGE INTENTIONALLY LEFT BLANK vii TABLE OF CONTENTS I. INTRODUCTION ...THIS PAGE INTENTIONALLY LEFT BLANK 1 I. INTRODUCTION Wireless mesh networks (WMNs) are interconnected systems of wireless access points (APs...operations in austere environments, such as combat and humanitarian assistance disaster relief (HA/DR) operations. See Nicholas (2009) for an introduction
Tsunami modelling with adaptively refined finite volume methods
LeVeque, R.J.; George, D.L.; Berger, M.J.
2011-01-01
Numerical modelling of transoceanic tsunami propagation, together with the detailed modelling of inundation of small-scale coastal regions, poses a number of algorithmic challenges. The depth-averaged shallow water equations can be used to reduce this to a time-dependent problem in two space dimensions, but even so it is crucial to use adaptive mesh refinement in order to efficiently handle the vast differences in spatial scales. This must be done in a 'wellbalanced' manner that accurately captures very small perturbations to the steady state of the ocean at rest. Inundation can be modelled by allowing cells to dynamically change from dry to wet, but this must also be done carefully near refinement boundaries. We discuss these issues in the context of Riemann-solver-based finite volume methods for tsunami modelling. Several examples are presented using the GeoClaw software, and sample codes are available to accompany the paper. The techniques discussed also apply to a variety of other geophysical flows. ?? 2011 Cambridge University Press.
NASA Astrophysics Data System (ADS)
Kim, Kyeong Ok; Choi, Byung Ho; Jung, Kyung Tae
2016-04-01
The performance of an integrally coupled wave-tide-surge model using the unstructured mesh system has been tested for the typhoon Bolaven which is regarded as the most powerful storm to strike the Korean Peninsula in nearly a decade with wind gusts measured up to 50 m/s, causing serious damages with 19 victims. Use of the unstructured mesh in coastal sea regions of marginal scale allows all energy from deep to shallow waters to be seamlessly followed; the physics of wave-circulation interactions can be then correctly resolved. The model covers the whole Yellow and East China Seas with locally refined meshes near the regions of Gageo Island (offshore southwestern corner of the Korean Peninsula) and south of Jeju Island (Gangjeong and Seogwipo ports). The wind and pressure fields during the passage of typhoon Bolaven are generated by the blending method. Generally the numerical atmospheric model cannot satisfactorily reproduce the strength of typhoons due to dynamic and resolution restrictions. In this study we could achieve an improved conservation of the typhoon strength by blending the Holland typhoon model result by the empirical formula onto the ambient meteorological fields of NCEP dataset. The model results are compared with the observations and the model performance is then evaluated. The computed wave spectrums for one and two dimensions are compared with the observation in Ieodo station. Results show that the wind wave significantly enhances the current intensity and surge elevation, addressing that to incorporate the wave-current interaction effect in the wave-tide-surge coupled model is important for the accurate prediction of current and sea surface elevation as well as extreme waves in shallow coastal sea regions. The resulting modeling system can be used for hindcasting and forecasting the wave-tide-surges in marine environments with complex coastlines, shallow water depth and fine sediment.
A heuristic model of sensory adaptation.
McBurney, Donald H; Balaban, Carey D
2009-11-01
Adaptation is a universal process in organisms as diverse as bacteria and humans, and across the various senses. This article proposes a simple, heuristic, mathematical model containing tonic and phasic processes. The model demonstrates properties not commonly associated with adaptation, such as increased sensitivity to changes, range shifting, and phase lead. Changes in only four parameters permit the model to predict empirical psychophysical data from different senses. The relatively prolonged time courses of responses to oral and topical capsaicin are used to illustrate and validate this mathematical modeling approach for different stimulus profiles. Other examples of phenomena elucidated by this modeling approach include the time courses of taste sensation, brightness perception, loudness perception, cross-adaptation to oral irritants, and cutaneous mechanoreception. It also predicts such apparently unrelated phenomena as perceived alcohol intoxication, habituation, and drug tolerance. Because the integration of phasic and tonic components is a conservative, highly efficacious solution to a ubiquitous biological challenge, sensory adaptation is seen as an evolutionary adaptation, and as a prominent feature of Mother Nature's small bag of tricks.
Adaptive Modeling Language and Its Derivatives
NASA Technical Reports Server (NTRS)
Chemaly, Adel
2006-01-01
Adaptive Modeling Language (AML) is the underlying language of an object-oriented, multidisciplinary, knowledge-based engineering framework. AML offers an advanced modeling paradigm with an open architecture, enabling the automation of the entire product development cycle, integrating product configuration, design, analysis, visualization, production planning, inspection, and cost estimation.
NASA Astrophysics Data System (ADS)
Li, Pak Shing; Martin, Daniel F.; Klein, Richard I.; McKee, Christopher F.
2012-02-01
Performing a stable, long-duration simulation of driven MHD turbulence with a high thermal Mach number and a strong initial magnetic field is a challenge to high-order Godunov ideal MHD schemes because of the difficulty in guaranteeing positivity of the density and pressure. We have implemented a robust combination of reconstruction schemes, Riemann solvers, limiters, and constrained transport electromotive force averaging schemes that can meet this challenge, and using this strategy, we have developed a new adaptive mesh refinement (AMR) MHD module of the ORION2 code. We investigate the effects of AMR on several statistical properties of a turbulent ideal MHD system with a thermal Mach number of 10 and a plasma β0 of 0.1 as initial conditions; our code is shown to be stable for simulations with higher Mach numbers ({{\\cal M}_rms}= 17.3) and smaller plasma beta (β0 = 0.0067) as well. Our results show that the quality of the turbulence simulation is generally related to the volume-averaged refinement. Our AMR simulations show that the turbulent dissipation coefficient for supersonic MHD turbulence is about 0.5, in agreement with unigrid simulations.
Li, Pak Shing; Klein, Richard I.; Martin, Daniel F.; McKee, Christopher F. E-mail: klein@astron.berkeley.edu E-mail: cmckee@astro.berkeley.edu
2012-02-01
Performing a stable, long-duration simulation of driven MHD turbulence with a high thermal Mach number and a strong initial magnetic field is a challenge to high-order Godunov ideal MHD schemes because of the difficulty in guaranteeing positivity of the density and pressure. We have implemented a robust combination of reconstruction schemes, Riemann solvers, limiters, and constrained transport electromotive force averaging schemes that can meet this challenge, and using this strategy, we have developed a new adaptive mesh refinement (AMR) MHD module of the ORION2 code. We investigate the effects of AMR on several statistical properties of a turbulent ideal MHD system with a thermal Mach number of 10 and a plasma {beta}{sub 0} of 0.1 as initial conditions; our code is shown to be stable for simulations with higher Mach numbers (M{sub rms}= 17.3) and smaller plasma beta ({beta}{sub 0} = 0.0067) as well. Our results show that the quality of the turbulence simulation is generally related to the volume-averaged refinement. Our AMR simulations show that the turbulent dissipation coefficient for supersonic MHD turbulence is about 0.5, in agreement with unigrid simulations.
A density driven mesh generator guided by a neural network
Lowther, D.A.; Dyck, D.N. )
1993-03-01
A neural network guided mesh generator is described. The mesh generator used density information provided by the neural network to determine the size and placement of elements. This system is coupled with an adaptive meshing and solving process and is shown to have major computational benefits compared with adaptation alone.
Adaptive approximation models in optimization
Voronin, A.N.
1995-05-01
The paper proposes a method for optimization of functions of several variables that substantially reduces the number of objective function evaluations compared to traditional methods. The method is based on the property of iterative refinement of approximation models of the optimand function in approximation domains that contract to the extremum point. It does not require subjective specification of the starting point, step length, or other parameters of the search procedure. The method is designed for efficient optimization of unimodal functions of several (not more than 10-15) variables and can be applied to find the global extremum of polymodal functions and also for optimization of scalarized forms of vector objective functions.
Multiscale mesh generation on the sphere
NASA Astrophysics Data System (ADS)
Lambrechts, Jonathan; Comblen, Richard; Legat, Vincent; Geuzaine, Christophe; Remacle, Jean-François
2008-12-01
A method for generating computational meshes for applications in ocean modeling is presented. The method uses a standard engineering approach for describing the geometry of the domain that requires meshing. The underlying sphere is parametrized using stereographic coordinates. Then, coastlines are described with cubic splines drawn in the stereographic parametric space. The mesh generation algorithm builds the mesh in the parametric plane using available techniques. The method enables to import coastlines from different data sets and, consequently, to build meshes of domains with highly variable length scales. The results include meshes together with numerical simulations of various kinds.
Gonzales, Matthew J.; Sturgeon, Gregory; Krishnamurthy, Adarsh; Hake, Johan; Jonas, René; Stark, Paul; Rappel, Wouter-Jan; Narayan, Sanjiv M.; Zhang, Yongjie; Segars, W. Paul; McCulloch, Andrew D.
2013-01-01
High-order cubic Hermite finite elements have been valuable in modeling cardiac geometry, fiber orientations, biomechanics, and electrophysiology, but their use in solving three-dimensional problems has been limited to ventricular models with simple topologies. Here, we utilized a subdivision surface scheme and derived a generalization of the “local-to-global” derivative mapping scheme of cubic Hermite finite elements to construct bicubic and tricubic Hermite models of the human atria with extraordinary vertices from computed tomography images of a patient with atrial fibrillation. To an accuracy of 0.6 millimeters, we were able to capture the left atrial geometry with only 142 bicubic Hermite finite elements, and the right atrial geometry with only 90. The left and right atrial bicubic Hermite meshes were G1 continuous everywhere except in the one-neighborhood of extraordinary vertices, where the mean dot products of normals at adjacent elements were 0.928 and 0.925. We also constructed two biatrial tricubic Hermite models and defined fiber orientation fields in agreement with diagrammatic data from the literature using only 42 angle parameters. The meshes all have good quality metrics, uniform element sizes, and elements with aspect ratios near unity, and are shared with the public. These new methods will allow for more compact and efficient patient-specific models of human atrial and whole heart physiology. PMID:23602918
Gonzales, Matthew J; Sturgeon, Gregory; Krishnamurthy, Adarsh; Hake, Johan; Jonas, René; Stark, Paul; Rappel, Wouter-Jan; Narayan, Sanjiv M; Zhang, Yongjie; Segars, W Paul; McCulloch, Andrew D
2013-07-01
High-order cubic Hermite finite elements have been valuable in modeling cardiac geometry, fiber orientations, biomechanics, and electrophysiology, but their use in solving three-dimensional problems has been limited to ventricular models with simple topologies. Here, we utilized a subdivision surface scheme and derived a generalization of the "local-to-global" derivative mapping scheme of cubic Hermite finite elements to construct bicubic and tricubic Hermite models of the human atria with extraordinary vertices from computed tomography images of a patient with atrial fibrillation. To an accuracy of 0.6 mm, we were able to capture the left atrial geometry with only 142 bicubic Hermite finite elements, and the right atrial geometry with only 90. The left and right atrial bicubic Hermite meshes were G1 continuous everywhere except in the one-neighborhood of extraordinary vertices, where the mean dot products of normals at adjacent elements were 0.928 and 0.925. We also constructed two biatrial tricubic Hermite models and defined fiber orientation fields in agreement with diagrammatic data from the literature using only 42 angle parameters. The meshes all have good quality metrics, uniform element sizes, and elements with aspect ratios near unity, and are shared with the public. These new methods will allow for more compact and efficient patient-specific models of human atrial and whole heart physiology.
Adaptive System Modeling for Spacecraft Simulation
NASA Technical Reports Server (NTRS)
Thomas, Justin
2011-01-01
This invention introduces a methodology and associated software tools for automatically learning spacecraft system models without any assumptions regarding system behavior. Data stream mining techniques were used to learn models for critical portions of the International Space Station (ISS) Electrical Power System (EPS). Evaluation on historical ISS telemetry data shows that adaptive system modeling reduces simulation error anywhere from 50 to 90 percent over existing approaches. The purpose of the methodology is to outline how someone can create accurate system models from sensor (telemetry) data. The purpose of the software is to support the methodology. The software provides analysis tools to design the adaptive models. The software also provides the algorithms to initially build system models and continuously update them from the latest streaming sensor data. The main strengths are as follows: Creates accurate spacecraft system models without in-depth system knowledge or any assumptions about system behavior. Automatically updates/calibrates system models using the latest streaming sensor data. Creates device specific models that capture the exact behavior of devices of the same type. Adapts to evolving systems. Can reduce computational complexity (faster simulations).
6th International Meshing Roundtable '97
White, D.
1997-09-01
The goal of the 6th International Meshing Roundtable is to bring together researchers and developers from industry, academia, and government labs in a stimulating, open environment for the exchange of technical information related to the meshing process. In the pas~ the Roundtable has enjoyed significant participation born each of these groups from a wide variety of countries. The Roundtable will consist of technical presentations from contributed papers and abstracts, two invited speakers, and two invited panels of experts discussing topics related to the development and use of automatic mesh generation tools. In addition, this year we will feature a "Bring Your Best Mesh" competition and poster session to encourage discussion and participation from a wide variety of mesh generation tool users. The schedule and evening social events are designed to provide numerous opportunities for informal dialog. A proceedings will be published by Sandia National Laboratories and distributed at the Roundtable. In addition, papers of exceptionally high quaIity will be submitted to a special issue of the International Journal of Computational Geometry and Applications. Papers and one page abstracts were sought that present original results on the meshing process. Potential topics include but are got limited to: Unstructured triangular and tetrahedral mesh generation Unstructured quadrilateral and hexahedral mesh generation Automated blocking and structured mesh generation Mixed element meshing Surface mesh generation Geometry decomposition and clean-up techniques Geometry modification techniques related to meshing Adaptive mesh refinement and mesh quality control Mesh visualization Special purpose meshing algorithms for particular applications Theoretical or novel ideas with practical potential Technical presentations from industrial researchers.
Quadrilateral/hexahedral finite element mesh coarsening
Staten, Matthew L; Dewey, Mark W; Scott, Michael A; Benzley, Steven E
2012-10-16
A technique for coarsening a finite element mesh ("FEM") is described. This technique includes identifying a coarsening region within the FEM to be coarsened. Perimeter chords running along perimeter boundaries of the coarsening region are identified. The perimeter chords are redirected to create an adaptive chord separating the coarsening region from a remainder of the FEM. The adaptive chord runs through mesh elements residing along the perimeter boundaries of the coarsening region. The adaptive chord is then extracted to coarsen the FEM.
A Model for Optimal Constrained Adaptive Testing.
ERIC Educational Resources Information Center
van der Linden, Wim J.; Reese, Lynda M.
1998-01-01
Proposes a model for constrained computerized adaptive testing in which the information in the test at the trait level (theta) estimate is maximized subject to the number of possible constraints on the content of the test. Test assembly relies on a linear-programming approach. Illustrates the approach through simulation with items from the Law…
NASA Astrophysics Data System (ADS)
Cai, Hongzhu; Hu, Xiangyun; Li, Jianhui; Endo, Masashi; Xiong, Bin
2017-02-01
We solve the 3D controlled-source electromagnetic (CSEM) problem using the edge-based finite element method. The modeling domain is discretized using unstructured tetrahedral mesh. We adopt the total field formulation for the quasi-static variant of Maxwell's equation and the computation cost to calculate the primary field can be saved. We adopt a new boundary condition which approximate the total field on the boundary by the primary field corresponding to the layered earth approximation of the complicated conductivity model. The primary field on the modeling boundary is calculated using fast Hankel transform. By using this new type of boundary condition, the computation cost can be reduced significantly and the modeling accuracy can be improved. We consider that the conductivity can be anisotropic. We solve the finite element system of equations using a parallelized multifrontal solver which works efficiently for multiple source and large scale electromagnetic modeling.
NASA Astrophysics Data System (ADS)
Dahdouh, S.; Varsier, N.; Serrurier, A.; De la Plata, J.-P.; Anquez, J.; Angelini, E. D.; Wiart, J.; Bloch, I.
2014-08-01
Fetal dosimetry studies require the development of accurate numerical 3D models of the pregnant woman and the fetus. This paper proposes a 3D articulated fetal growth model covering the main phases of pregnancy and a pregnant woman model combining the utero-fetal structures and a deformable non-pregnant woman body envelope. The structures of interest were automatically or semi-automatically (depending on the stage of pregnancy) segmented from a database of images and surface meshes were generated. By interpolating linearly between fetal structures, each one can be generated at any age and in any position. A method is also described to insert the utero-fetal structures in the maternal body. A validation of the fetal models is proposed, comparing a set of biometric measurements to medical reference charts. The usability of the pregnant woman model in dosimetry studies is also investigated, with respect to the influence of the abdominal fat layer.
Segmentation of the heart and great vessels in CT images using a model-based adaptation framework.
Ecabert, Olivier; Peters, Jochen; Walker, Matthew J; Ivanc, Thomas; Lorenz, Cristian; von Berg, Jens; Lessick, Jonathan; Vembar, Mani; Weese, Jürgen
2011-12-01
Recently, model-based methods for the automatic segmentation of the heart chambers have been proposed. An important application of these methods is the characterization of the heart function. Heart models are, however, increasingly used for interventional guidance making it necessary to also extract the attached great vessels. It is, for instance, important to extract the left atrium and the proximal part of the pulmonary veins to support guidance of ablation procedures for atrial fibrillation treatment. For cardiac resynchronization therapy, a heart model including the coronary sinus is needed. We present a heart model comprising the four heart chambers and the attached great vessels. By assigning individual linear transformations to the heart chambers and to short tubular segments building the great vessels, variable sizes of the heart chambers and bending of the vessels can be described in a consistent way. A configurable algorithmic framework that we call adaptation engine matches the heart model automatically to cardiac CT angiography images in a multi-stage process. First, the heart is detected using a Generalized Hough Transformation. Subsequently, the heart chambers are adapted. This stage uses parametric as well as deformable mesh adaptation techniques. In the final stage, segments of the large vascular structures are successively activated and adapted. To optimize the computational performance, the adaptation engine can vary the mesh resolution and freeze already adapted mesh parts. The data used for validation were independent from the data used for model-building. Ground truth segmentations were generated for 37 CT data sets reconstructed at several cardiac phases from 17 patients. Segmentation errors were assessed for anatomical sub-structures resulting in a mean surface-to-surface error ranging 0.50-0.82mm for the heart chambers and 0.60-1.32mm for the parts of the great vessels visible in the images.
Maximizing Adaptivity in Hierarchical Topological Models Using Cancellation Trees
Bremer, P; Pascucci, V; Hamann, B
2008-12-08
We present a highly adaptive hierarchical representation of the topology of functions defined over two-manifold domains. Guided by the theory of Morse-Smale complexes, we encode dependencies between cancellations of critical points using two independent structures: a traditional mesh hierarchy to store connectivity information and a new structure called cancellation trees to encode the configuration of critical points. Cancellation trees provide a powerful method to increase adaptivity while using a simple, easy-to-implement data structure. The resulting hierarchy is significantly more flexible than the one previously reported. In particular, the resulting hierarchy is guaranteed to be of logarithmic height.
Automated adaptive inference of phenomenological dynamical models
NASA Astrophysics Data System (ADS)
Daniels, Bryan C.; Nemenman, Ilya
2015-08-01
Dynamics of complex systems is often driven by large and intricate networks of microscopic interactions, whose sheer size obfuscates understanding. With limited experimental data, many parameters of such dynamics are unknown, and thus detailed, mechanistic models risk overfitting and making faulty predictions. At the other extreme, simple ad hoc models often miss defining features of the underlying systems. Here we develop an approach that instead constructs phenomenological, coarse-grained models of network dynamics that automatically adapt their complexity to the available data. Such adaptive models produce accurate predictions even when microscopic details are unknown. The approach is computationally tractable, even for a relatively large number of dynamical variables. Using simulated data, it correctly infers the phase space structure for planetary motion, avoids overfitting in a biological signalling system and produces accurate predictions for yeast glycolysis with tens of data points and over half of the interacting species unobserved.
Automated adaptive inference of phenomenological dynamical models
Daniels, Bryan C.; Nemenman, Ilya
2015-01-01
Dynamics of complex systems is often driven by large and intricate networks of microscopic interactions, whose sheer size obfuscates understanding. With limited experimental data, many parameters of such dynamics are unknown, and thus detailed, mechanistic models risk overfitting and making faulty predictions. At the other extreme, simple ad hoc models often miss defining features of the underlying systems. Here we develop an approach that instead constructs phenomenological, coarse-grained models of network dynamics that automatically adapt their complexity to the available data. Such adaptive models produce accurate predictions even when microscopic details are unknown. The approach is computationally tractable, even for a relatively large number of dynamical variables. Using simulated data, it correctly infers the phase space structure for planetary motion, avoids overfitting in a biological signalling system and produces accurate predictions for yeast glycolysis with tens of data points and over half of the interacting species unobserved. PMID:26293508
Criman, Erik T.; Kurata, Wendy E.; Matsumoto, Karen W.; Aubin, Harry T.; Campbell, Carmen E.
2016-01-01
Background: The reported incidence of mesh infection in contaminated operative fields is as high as 30% regardless of the material used. Recently, mesenchymal stem cells (MSCs) have been shown to possess favorable immunomodulatory properties and improve tissue incorporation when seeded onto bioprosthetics. The aim of this study was to evaluate whether seeding noncrosslinked bovine pericardium (Veritas Collagen Matrix) with allogeneic bone marrow–derived MSCs improves infection resistance in vivo after inoculation with Escherichia coli (E. coli). Methods: Rat bone marrow–derived MSCs at passage 3 were seeded onto bovine pericardium and cultured for 7 days before implantation. Additional rats (n = 24) were implanted subcutaneously with MSC-seeded or unseeded mesh and inoculated with 7 × 105 colony-forming units of E. coli or saline before wound closure (group 1, unseeded mesh/saline; group 2, unseeded mesh/E. coli; group 3, MSC-seeded mesh/E. coli; 8 rats per group). Meshes were explanted at 4 weeks and underwent microbiologic and histologic analyses. Results: MSC-seeded meshes inoculated with E. coli demonstrated superior bacterial clearance and preservation of mesh integrity compared with E. coli–inoculated unseeded meshes (87.5% versus 0% clearance; p = 0.001). Complete mesh degradation concurrent with abscess formation was observed in 100% of rats in the unseeded/E. coli group, which is in contrast to 12.5% of rats in the MSC-seeded/E. coli group. Histologic evaluation determined that remodeling characteristics of E. coli–inoculated MSC-seeded meshes were similar to those of uninfected meshes 4 weeks after implantation. Conclusions: Augmenting a bioprosthetic material with stem cells seems to markedly enhance resistance to bacterial infection in vivo and preserve mesh integrity. PMID:27482490
POD-Galerkin reduced-order modeling with adaptive finite element snapshots
NASA Astrophysics Data System (ADS)
Ullmann, Sebastian; Rotkvic, Marko; Lang, Jens
2016-11-01
We consider model order reduction by proper orthogonal decomposition (POD) for parametrized partial differential equations, where the underlying snapshots are computed with adaptive finite elements. We address computational and theoretical issues arising from the fact that the snapshots are members of different finite element spaces. We propose a method to create a POD-Galerkin model without interpolating the snapshots onto their common finite element mesh. The error of the reduced-order solution is not necessarily Galerkin orthogonal to the reduced space created from space-adapted snapshot. We analyze how this influences the error assessment for POD-Galerkin models of linear elliptic boundary value problems. As a numerical example we consider a two-dimensional convection-diffusion equation with a parametrized convective direction. To illustrate the applicability of our techniques to non-linear time-dependent problems, we present a test case of a two-dimensional viscous Burgers equation with parametrized initial data.
Controlling Reflections from Mesh Refinement Interfaces in Numerical Relativity
NASA Technical Reports Server (NTRS)
Baker, John G.; Van Meter, James R.
2005-01-01
A leading approach to improving the accuracy on numerical relativity simulations of black hole systems is through fixed or adaptive mesh refinement techniques. We describe a generic numerical error which manifests as slowly converging, artificial reflections from refinement boundaries in a broad class of mesh-refinement implementations, potentially limiting the effectiveness of mesh- refinement techniques for some numerical relativity applications. We elucidate this numerical effect by presenting a model problem which exhibits the phenomenon, but which is simple enough that its numerical error can be understood analytically. Our analysis shows that the effect is caused by variations in finite differencing error generated across low and high resolution regions, and that its slow convergence is caused by the presence of dramatic speed differences among propagation modes typical of 3+1 relativity. Lastly, we resolve the problem, presenting a class of finite-differencing stencil modifications which eliminate this pathology in both our model problem and in numerical relativity examples.
NASA Astrophysics Data System (ADS)
Wong, M.; Skamarock, W. C.
2015-12-01
Global numerical weather forecast tests were performed using the global nonhydrostatic atmospheric model, Model for Prediction Across Scales (MPAS), for the NOAA Storm Prediction Center 2015 Spring Forecast Experiment (May 2015) and the Plains Elevated Convection at Night (PECAN) field campaign (June to mid-July 2015). These two sets of forecasts were performed on 50-to-3 km and 15-to-3 km smoothly-varying horizontal meshes, respectively. Both variable-resolution meshes have nominal convection-permitting 3-km grid spacing over the entire continental US. Here we evaluate the limited-area (vs. global) spectra from these NWP simulations. We will show the simulated spectral characteristics of total kinetic energy, vertical velocity variance, and precipitation during these spring and summer periods when diurnal continental convection is most active over central US. Spectral characteristics of a high-resolution global 3-km simulation (essentially no nesting) from the 20 May 2013 Moore, OK tornado case are also shown. These characteristics include spectral scaling, shape, and anisotropy, as well as the effective resolution of continental convection representation in MPAS.
Hybrid Mesh for Nasal Airflow Studies
Zubair, Mohammed; Abdullah, Mohammed Zulkifly; Ahmad, Kamarul Arifin
2013-01-01
The accuracy of the numerical result is closely related to mesh density as well as its distribution. Mesh plays a very significant role in the outcome of numerical simulation. Many nasal airflow studies have employed unstructured mesh and more recently hybrid mesh scheme has been utilized considering the complexity of anatomical architecture. The objective of this study is to compare the results of hybrid mesh with unstructured mesh and study its effect on the flow parameters inside the nasal cavity. A three-dimensional nasal cavity model is reconstructed based on computed tomographic images of a healthy Malaysian adult nose. Navier-Stokes equation for steady airflow is solved numerically to examine inspiratory nasal flow. The pressure drop obtained using the unstructured computational grid is about 22.6 Pa for a flow rate of 20 L/min, whereas the hybrid mesh resulted in 17.8 Pa for the same flow rate. The maximum velocity obtained at the nasal valve using unstructured grid is 4.18 m/s and that with hybrid mesh is around 4.76 m/s. Hybrid mesh reported lower grid convergence index (GCI) than the unstructured mesh. Significant differences between unstructured mesh and hybrid mesh are determined highlighting the usefulness of hybrid mesh for nasal airflow studies. PMID:23983811
Adaptive Behaviour Assessment System: Indigenous Australian Adaptation Model (ABAS: IAAM)
ERIC Educational Resources Information Center
du Plessis, Santie
2015-01-01
The study objectives were to develop, trial and evaluate a cross-cultural adaptation of the Adaptive Behavior Assessment System-Second Edition Teacher Form (ABAS-II TF) ages 5-21 for use with Indigenous Australian students ages 5-14. This study introduced a multiphase mixed-method design with semi-structured and informal interviews, school…
NASA Astrophysics Data System (ADS)
Gonizzi Barsanti, S.; Guidi, G.
2017-02-01
Conservation of Cultural Heritage is a key issue and structural changes and damages can influence the mechanical behaviour of artefacts and buildings. The use of Finite Elements Methods (FEM) for mechanical analysis is largely used in modelling stress behaviour. The typical workflow involves the use of CAD 3D models made by Non-Uniform Rational B-splines (NURBS) surfaces, representing the ideal shape of the object to be simulated. Nowadays, 3D documentation of CH has been widely developed through reality-based approaches, but the models are not suitable for a direct use in FEA: the mesh has in fact to be converted to volumetric, and the density has to be reduced since the computational complexity of a FEA grows exponentially with the number of nodes. The focus of this paper is to present a new method aiming at generate the most accurate 3D representation of a real artefact from highly accurate 3D digital models derived from reality-based techniques, maintaining the accuracy of the high-resolution polygonal models in the solid ones. The approach proposed is based on a wise use of retopology procedures and a transformation of this model to a mathematical one made by NURBS surfaces suitable for being processed by volumetric meshers typically embedded in standard FEM packages. The strong simplification with little loss of consistency possible with the retopology step is used for maintaining as much coherence as possible between the original acquired mesh and the simplified model, creating in the meantime a topology that is more favourable for the automatic NURBS conversion.
Euler and Navier-Stokes computations for two-dimensional geometries using unstructured meshes
NASA Technical Reports Server (NTRS)
Mavriplis, D. J.
1990-01-01
A general purpose unstructured mesh solver for steady-state two-dimensional inviscid and viscous flows is described. The efficiency and accuracy of the method are enhanced by the simultaneous use of adaptive meshing and an unstructured multigrid technique. A method for generating highly stretched triangulations in regions of viscous flow is outlined, and a procedure for implementing an algebraic turbulence model on unstructured meshes is described. Results are shown for external and internal inviscid flows and for turbulent viscous flow over a multi-element airfoil configuration.
Numerical methods for estimating J integral in models with regular rectangular meshes
NASA Astrophysics Data System (ADS)
Kozłowiec, B.
2017-02-01
Cracks and delaminations are the common structural degradation mechanisms studied recently using numerous methods and techniques. Among them, numerical methods based on FEM analyses are in widespread commercial use. The scope of these methods has focused i.e. on energetic approach to linear elastic fracture mechanics (LEFM) theory, encompassing such quantities as the J-integral and the energy release rate G. This approach enables to introduce damage criteria of analyzed structures without dealing with the details of the physical singularities occurring at the crack tip. In this paper, two numerical methods based on LEFM are used to analyze both isotropic and orthotropic specimens and the results are compared with well-known analytical solutions as well as (in some cases) VCCT results. These methods are optimized for industrial use with simple, rectangular meshes. The verification is made based on two dimensional mode partitioning.
Adaptive Control with Reference Model Modification
NASA Technical Reports Server (NTRS)
Stepanyan, Vahram; Krishnakumar, Kalmanje
2012-01-01
This paper presents a modification of the conventional model reference adaptive control (MRAC) architecture in order to improve transient performance of the input and output signals of uncertain systems. A simple modification of the reference model is proposed by feeding back the tracking error signal. It is shown that the proposed approach guarantees tracking of the given reference command and the reference control signal (one that would be designed if the system were known) not only asymptotically but also in transient. Moreover, it prevents generation of high frequency oscillations, which are unavoidable in conventional MRAC systems for large adaptation rates. The provided design guideline makes it possible to track a reference commands of any magnitude from any initial position without re-tuning. The benefits of the method are demonstrated with a simulation example
Adaptive cyber-attack modeling system
NASA Astrophysics Data System (ADS)
Gonsalves, Paul G.; Dougherty, Edward T.
2006-05-01
The pervasiveness of software and networked information systems is evident across a broad spectrum of business and government sectors. Such reliance provides an ample opportunity not only for the nefarious exploits of lone wolf computer hackers, but for more systematic software attacks from organized entities. Much effort and focus has been placed on preventing and ameliorating network and OS attacks, a concomitant emphasis is required to address protection of mission critical software. Typical software protection technique and methodology evaluation and verification and validation (V&V) involves the use of a team of subject matter experts (SMEs) to mimic potential attackers or hackers. This manpower intensive, time-consuming, and potentially cost-prohibitive approach is not amenable to performing the necessary multiple non-subjective analyses required to support quantifying software protection levels. To facilitate the evaluation and V&V of software protection solutions, we have designed and developed a prototype adaptive cyber attack modeling system. Our approach integrates an off-line mechanism for rapid construction of Bayesian belief network (BN) attack models with an on-line model instantiation, adaptation and knowledge acquisition scheme. Off-line model construction is supported via a knowledge elicitation approach for identifying key domain requirements and a process for translating these requirements into a library of BN-based cyber-attack models. On-line attack modeling and knowledge acquisition is supported via BN evidence propagation and model parameter learning.
Conformal refinement of unstructured quadrilateral meshes
Garmella, Rao
2009-01-01
We present a multilevel adaptive refinement technique for unstructured quadrilateral meshes in which the mesh is kept conformal at all times. This means that the refined mesh, like the original, is formed of only quadrilateral elements that intersect strictly along edges or at vertices, i.e., vertices of one quadrilateral element do not lie in an edge of another quadrilateral. Elements are refined using templates based on 1:3 refinement of edges. We demonstrate that by careful design of the refinement and coarsening strategy, we can maintain high quality elements in the refined mesh. We demonstrate the method on a number of examples with dynamically changing refinement regions.
Adaptive human behavior in epidemiological models
Fenichel, Eli P.; Castillo-Chavez, Carlos; Ceddia, M. G.; Chowell, Gerardo; Parra, Paula A. Gonzalez; Hickling, Graham J.; Holloway, Garth; Horan, Richard; Morin, Benjamin; Perrings, Charles; Springborn, Michael; Velazquez, Leticia; Villalobos, Cristina
2011-01-01
The science and management of infectious disease are entering a new stage. Increasingly public policy to manage epidemics focuses on motivating people, through social distancing policies, to alter their behavior to reduce contacts and reduce public disease risk. Person-to-person contacts drive human disease dynamics. People value such contacts and are willing to accept some disease risk to gain contact-related benefits. The cost–benefit trade-offs that shape contact behavior, and hence the course of epidemics, are often only implicitly incorporated in epidemiological models. This approach creates difficulty in parsing out the effects of adaptive behavior. We use an epidemiological–economic model of disease dynamics to explicitly model the trade-offs that drive person-to-person contact decisions. Results indicate that including adaptive human behavior significantly changes the predicted course of epidemics and that this inclusion has implications for parameter estimation and interpretation and for the development of social distancing policies. Acknowledging adaptive behavior requires a shift in thinking about epidemiological processes and parameters. PMID:21444809
Adaptive human behavior in epidemiological models.
Fenichel, Eli P; Castillo-Chavez, Carlos; Ceddia, M G; Chowell, Gerardo; Parra, Paula A Gonzalez; Hickling, Graham J; Holloway, Garth; Horan, Richard; Morin, Benjamin; Perrings, Charles; Springborn, Michael; Velazquez, Leticia; Villalobos, Cristina
2011-04-12
The science and management of infectious disease are entering a new stage. Increasingly public policy to manage epidemics focuses on motivating people, through social distancing policies, to alter their behavior to reduce contacts and reduce public disease risk. Person-to-person contacts drive human disease dynamics. People value such contacts and are willing to accept some disease risk to gain contact-related benefits. The cost-benefit trade-offs that shape contact behavior, and hence the course of epidemics, are often only implicitly incorporated in epidemiological models. This approach creates difficulty in parsing out the effects of adaptive behavior. We use an epidemiological-economic model of disease dynamics to explicitly model the trade-offs that drive person-to-person contact decisions. Results indicate that including adaptive human behavior significantly changes the predicted course of epidemics and that this inclusion has implications for parameter estimation and interpretation and for the development of social distancing policies. Acknowledging adaptive behavior requires a shift in thinking about epidemiological processes and parameters.
NASA Astrophysics Data System (ADS)
Bourantas, Georgios; Lavier, Luc; Claus, Susanne; Van Dam, Tonie; Bordas, Stephane
2015-04-01
Geodynamic modeling is an important branch of Earth Sciences. Direct observation of geodynamic processes is limited in both time and space, while on the other hand numerical methods are capable of simulating millions of years in a matter of days on a desktop computer. The model equations can be reduced to a set of Partial Differential Equations with possibly discontinuous coefficients, governing mass, momentum and heat transfer over the domain. Some of the major challenges associated with such simulations are (1) geological time scales, which require long (in physical time) simulations using small time steps; (2) the presence of localization zones over which large gradients are present and which are much smaller than the overall physical dimensions of the computational domain and require much more refined discretization than for the rest of the domain, much like in fracture or shear band mechanics. An added difficulty is that such structures in the solution may appear after long periods of stagnant behaviour; (3) the definition of boundary conditions, material parameters and that of a suitable computational domain in terms of size; (4) a posteriori error estimation, sensitivity analysis and discretization adaptivity for the resulting coupled problem, including error propagation between different unknown fields. Consequently, it is arguable that any suitable numerical methods aimed at the solution of such problems on a large scale must be able to (i) provide ease of discretization refinement, including possible partition of unity enrichment; (ii) offer a large stability domain, so that "large" time steps can be chosen; (iii) ease of parallelization and good scalability. Our approach is to rely on "meshless" methods based on a point collocation strategy for the discretization of the set of PDEs. The method is hybrid Eulerian/Lagrangian, which enables to switch easily between stagnant periods and periods of localization. Mass and momentum equations are solved using a
A Model for Climate Change Adaptation
NASA Astrophysics Data System (ADS)
Pasqualini, D.; Keating, G. N.
2009-12-01
Climate models predict serious impacts on the western U.S. in the next few decades, including increased temperatures and reduced precipitation. In combination, these changes are linked to profound impacts on fundamental systems, such as water and energy supplies, agriculture, population stability, and the economy. Global and national imperatives for climate change mitigation and adaptation are made actionable at the state level, for instance through greenhouse gas (GHG) emission regulations and incentives for renewable energy sources. However, adaptation occurs at the local level, where energy and water usage can be understood relative to local patterns of agriculture, industry, and culture. In response to the greenhouse gas emission reductions required by California’s Assembly Bill 32 (2006), Sonoma County has committed to sharp emissions reductions across several sectors, including water, energy, and transportation. To assist Sonoma County develop a renewable energy (RE) portfolio to achieve this goal we have developed an integrated assessment model, CLEAR (CLimate-Energy Assessment for Resiliency) model. Building on Sonoma County’s existing baseline studies of energy use, carbon emissions and potential RE sources, the CLEAR model simulates the complex interactions among technology deployment, economics and social behavior. This model enables assessment of these and other components with specific analysis of their coupling and feedbacks because, due to the complex nature of the problem, the interrelated sectors cannot be studied independently. The goal is an approach to climate change mitigation and adaptation that is replicable for use by other interested communities. The model user interfaces helps stakeholders and policymakers understand options for technology implementation.
NASA Technical Reports Server (NTRS)
Macneice, Peter
1995-01-01
This is an introduction to numerical Particle-Mesh techniques, which are commonly used to model plasmas, gravitational N-body systems, and both compressible and incompressible fluids. The theory behind this approach is presented, and its practical implementation, both for serial and parallel machines, is discussed. This document is based on a four-hour lecture course presented by the author at the NASA Summer School for High Performance Computational Physics, held at Goddard Space Flight Center.
MHD simulations on an unstructured mesh
Strauss, H.R.; Park, W.; Belova, E.; Fu, G.Y.; Longcope, D.W.; Sugiyama, L.E.
1998-12-31
Two reasons for using an unstructured computational mesh are adaptivity, and alignment with arbitrarily shaped boundaries. Two codes which use finite element discretization on an unstructured mesh are described. FEM3D solves 2D and 3D RMHD using an adaptive grid. MH3D++, which incorporates methods of FEM3D into the MH3D generalized MHD code, can be used with shaped boundaries, which might be 3D.
Income distribution: An adaptive heterogeneous model
NASA Astrophysics Data System (ADS)
da Silva, L. C.; de Figueirêdo, P. H.
2014-02-01
In this communication an adaptive process is introduced into a many-agent model for closed economic system in order to establish general features of income distribution. In this new version agents are able to modify their exchange parameter ωi of resources through an adaptive process. The conclusions indicate that assuming an instantaneous learning behavior of all agents a Γ-distribution for income is reproduced while a frozen behavior establishes a Pareto’s distribution for income with an exponent ν=0.94±0.02. A third case occurs when a heterogeneous “inertia” behavior is introduced leading us to a Γ-distribution at the low income regime and a power-law decay for the large income values with an exponent ν=2.05±0.05. This method enables investigation of the resources flux in the economic environment and produces also bounding values for the Gini index comparable with data evidences.
Dickinson, Robert E.
1998-08-31
The characteristics of land important for climate are very heterogeneous, as are the key atmospheric inputs to land, i.e. precipitation and radiation. To adequately represent this heterogeneity, state-of-the-art climate models should represent atmospheric inputs to land, land properties, and the dynamical changes of land at the highest resolution accessible by climate models. The research funded under this project focused on the development of an alternative approach to this problem in which a sub-mesh is imposed on each atmospheric model grid square. This allows representation of the land climate dynamics at a higher resolution than that achievable in the global atmospheric models. The high spatial detail of the fine-mesh treatment provides not only a more accurate representation of land processes to the atmospheric model, but also the opportunity for direct downscaling of the surface climate. The principal objectives were: (1) To complete the development of fine-mesh data structures in the VBATS model and its link to CCM2; (2) To improve BATS model parameterizations; (3) To complete and refine fine-mesh atmospheric parameterizations; and (4) To conduct sensitivity studies. The primary shift in goals has been to include and emphasize linkages to CCM3 which has been publicly released as of May 1996.
An adaptive contextual quantum language model
NASA Astrophysics Data System (ADS)
Li, Jingfei; Zhang, Peng; Song, Dawei; Hou, Yuexian
2016-08-01
User interactions in search system represent a rich source of implicit knowledge about the user's cognitive state and information need that continuously evolves over time. Despite massive efforts that have been made to exploiting and incorporating this implicit knowledge in information retrieval, it is still a challenge to effectively capture the term dependencies and the user's dynamic information need (reflected by query modifications) in the context of user interaction. To tackle these issues, motivated by the recent Quantum Language Model (QLM), we develop a QLM based retrieval model for session search, which naturally incorporates the complex term dependencies occurring in user's historical queries and clicked documents with density matrices. In order to capture the dynamic information within users' search session, we propose a density matrix transformation framework and further develop an adaptive QLM ranking model. Extensive comparative experiments show the effectiveness of our session quantum language models.
Unstructured mesh methods for CFD
NASA Technical Reports Server (NTRS)
Peraire, J.; Morgan, K.; Peiro, J.
1990-01-01
Mesh generation methods for Computational Fluid Dynamics (CFD) are outlined. Geometric modeling is discussed. An advancing front method is described. Flow past a two engine Falcon aeroplane is studied. An algorithm and associated data structure called the alternating digital tree, which efficiently solves the geometric searching problem is described. The computation of an initial approximation to the steady state solution of a given poblem is described. Mesh generation for transient flows is described.
Adaptive h -refinement for reduced-order models: ADAPTIVE h -refinement for reduced-order models
Carlberg, Kevin T.
2014-11-05
Our work presents a method to adaptively refine reduced-order models a posteriori without requiring additional full-order-model solves. The technique is analogous to mesh-adaptive h-refinement: it enriches the reduced-basis space online by ‘splitting’ a given basis vector into several vectors with disjoint support. The splitting scheme is defined by a tree structure constructed offline via recursive k-means clustering of the state variables using snapshot data. This method identifies the vectors to split online using a dual-weighted-residual approach that aims to reduce error in an output quantity of interest. The resulting method generates a hierarchy of subspaces online without requiring large-scale operations or full-order-model solves. Furthermore, it enables the reduced-order model to satisfy any prescribed error tolerance regardless of its original fidelity, as a completely refined reduced-order model is mathematically equivalent to the original full-order model. Experiments on a parameterized inviscid Burgers equation highlight the ability of the method to capture phenomena (e.g., moving shocks) not contained in the span of the original reduced basis.
Model reference adaptive control of robots
NASA Technical Reports Server (NTRS)
Steinvorth, Rodrigo
1991-01-01
This project presents the results of controlling two types of robots using new Command Generator Tracker (CGT) based Direct Model Reference Adaptive Control (MRAC) algorithms. Two mathematical models were used to represent a single-link, flexible joint arm and a Unimation PUMA 560 arm; and these were then controlled in simulation using different MRAC algorithms. Special attention was given to the performance of the algorithms in the presence of sudden changes in the robot load. Previously used CGT based MRAC algorithms had several problems. The original algorithm that was developed guaranteed asymptotic stability only for almost strictly positive real (ASPR) plants. This condition is very restrictive, since most systems do not satisfy this assumption. Further developments to the algorithm led to an expansion of the number of plants that could be controlled, however, a steady state error was introduced in the response. These problems led to the introduction of some modifications to the algorithms so that they would be able to control a wider class of plants and at the same time would asymptotically track the reference model. This project presents the development of two algorithms that achieve the desired results and simulates the control of the two robots mentioned before. The results of the simulations are satisfactory and show that the problems stated above have been corrected in the new algorithms. In addition, the responses obtained show that the adaptively controlled processes are resistant to sudden changes in the load.
Adaptive model training system and method
Bickford, Randall L; Palnitkar, Rahul M; Lee, Vo
2014-04-15
An adaptive model training system and method for filtering asset operating data values acquired from a monitored asset for selectively choosing asset operating data values that meet at least one predefined criterion of good data quality while rejecting asset operating data values that fail to meet at least the one predefined criterion of good data quality; and recalibrating a previously trained or calibrated model having a learned scope of normal operation of the asset by utilizing the asset operating data values that meet at least the one predefined criterion of good data quality for adjusting the learned scope of normal operation of the asset for defining a recalibrated model having the adjusted learned scope of normal operation of the asset.
Adaptive model training system and method
Bickford, Randall L; Palnitkar, Rahul M
2014-11-18
An adaptive model training system and method for filtering asset operating data values acquired from a monitored asset for selectively choosing asset operating data values that meet at least one predefined criterion of good data quality while rejecting asset operating data values that fail to meet at least the one predefined criterion of good data quality; and recalibrating a previously trained or calibrated model having a learned scope of normal operation of the asset by utilizing the asset operating data values that meet at least the one predefined criterion of good data quality for adjusting the learned scope of normal operation of the asset for defining a recalibrated model having the adjusted learned scope of normal operation of the asset.
NASA Technical Reports Server (NTRS)
Lombard, C. K.; Lombard, M. P.; Menees, G. P.; Yang, J. Y.
1980-01-01
Several aspects connected with the notion of computation with flow oriented mesh systems are presented. Simple, effective approaches to the ideas discussed are demonstrated in current applications to blown forebody shock layer flow and full bluff body shock layer flow including the massively separated wake region.
Mesh quality control for multiply-refined tetrahedral grids
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Strawn, Roger
1994-01-01
A new algorithm for controlling the quality of multiply-refined tetrahedral meshes is presented in this paper. The basic dynamic mesh adaption procedure allows localized grid refinement and coarsening to efficiently capture aerodynamic flow features in computational fluid dynamics problems; however, repeated application of the procedure may significantly deteriorate the quality of the mesh. Results presented show the effectiveness of this mesh quality algorithm and its potential in the area of helicopter aerodynamics and acoustics.
Gradient scaling for nonuniform meshes
Margolin, L.G.; Ruppel, H.M.; Demuth, R.B.
1985-01-01
This paper is concerned with the effect of nonuniform meshes on the accuracy of finite-difference calculations of fluid flow. In particular, when a simple shock propagates through a nonuniform mesh, one may fail to model the jump conditions across the shock even when the equations are differenced in manifestly conservative fashion. We develop an approximate dispersion analysis of the numerical equations and identify the source of the mesh dependency with the form of the artificial viscosity. We then derive an algebraic correction to the numerical equations - a scaling factor for the pressure gradient - to essentially eliminate the mesh dependency. We present several calculations to illustrate our theory. We conclude with an alternate interpretation of our results. 14 refs., 5 figs.
Pascual, Gemma; Sotomayor, Sandra; Rodríguez, Marta; Bayon, Yves; Bellón, Juan M.
2013-01-01
Introduction Composite biomaterials designed for the repair of abdominal wall defects are composed of a mesh component and a laminar barrier in contact with the visceral peritoneum. This study assesses the behaviour of a new composite mesh by comparing it with two latest-generation composites currently used in clinical practice. Methods Defects (7x5cm) created in the anterior abdominal wall of New Zealand White rabbits were repaired using a polypropylene mesh and the composites: PhysiomeshTM; VentralightTM and a new composite mesh with a three-dimensional macroporous polyester structure and an oxidized collagen/chitosan barrier. Animals were sacrificed on days 14 and 90 postimplant. Specimens were processed to determine host tissue incorporation, gene/protein expression of neo-collagens (RT-PCR/immunofluorescence), macrophage response (RAM-11-immunolabelling) and biomechanical resistance. On postoperative days 7/14, each animal was examined laparoscopically to quantify adhesions between the visceral peritoneum and implant. Results The new composite mesh showed the lowest incidence of seroma in the short term. At each time point, the mesh surface covered with adhesions was greater in controls than composites. By day 14, the implants were fully infiltrated by a loose connective tissue that became denser over time. At 90 days, the peritoneal mesh surface was lined with a stable mesothelium. The new composite mesh induced more rapid tissue maturation than PhysiomeshTM, giving rise to a neoformed tissue containing more type I collagen. In VentralightTM the macrophage reaction was intense and significantly greater than the other composites at both follow-up times. Tensile strengths were similar for each biomaterial. Conclusions All composites showed optimal peritoneal behaviour, inducing good peritoneal regeneration and scarce postoperative adhesion formation. A greater foreign body reaction was observed for VentralightTM. All composites induced good collagen deposition
Plant adaptive behaviour in hydrological models (Invited)
NASA Astrophysics Data System (ADS)
van der Ploeg, M. J.; Teuling, R.
2013-12-01
Models that will be able to cope with future precipitation and evaporation regimes need a solid base that describes the essence of the processes involved [1]. Micro-behaviour in the soil-vegetation-atmosphere system may have a large impact on patterns emerging at larger scales. A complicating factor in the micro-behaviour is the constant interaction between vegetation and geology in which water plays a key role. The resilience of the coupled vegetation-soil system critically depends on its sensitivity to environmental changes. As a result of environmental changes vegetation may wither and die, but such environmental changes may also trigger gene adaptation. Constant exposure to environmental stresses, biotic or abiotic, influences plant physiology, gene adaptations, and flexibility in gene adaptation [2-6]. Gene expression as a result of different environmental conditions may profoundly impact drought responses across the same plant species. Differences in response to an environmental stress, has consequences for the way species are currently being treated in models (single plant to global scale). In particular, model parameters that control root water uptake and plant transpiration are generally assumed to be a property of the plant functional type. Assigning plant functional types does not allow for local plant adaptation to be reflected in the model parameters, nor does it allow for correlations that might exist between root parameters and soil type. Models potentially provide a means to link root water uptake and transport to large scale processes (e.g. Rosnay and Polcher 1998, Feddes et al. 2001, Jung 2010), especially when powered with an integrated hydrological, ecological and physiological base. We explore the experimental evidence from natural vegetation to formulate possible alternative modeling concepts. [1] Seibert, J. 2000. Multi-criteria calibration of a conceptual runoff model using a genetic algorithm. Hydrology and Earth System Sciences 4(2): 215
Adaptive Model of Wastewater Aeration Tank
NASA Astrophysics Data System (ADS)
Sniders, Andris; Laizans, Aigars
2011-01-01
The paper discusses the methodology of oxygen transfer virtual simulation in a wastewater biological treatment process, using the MATLAB/SIMULINK technology. A self-tuning adaptive model of a wastewater aeration tank, as a non-stationary object, with variable time dependent sensitivity and inertia indexes, as the functions of input variable - air pneumatic supply capacity Lg(t) (m3/min), output variable - dissolved oxygen concentration C(t) (g/m3) and oxygen expenditure, as a load - q(t) (g/min), required for wastewater complete purification, is expounded. Virtual models, applying Laplace transforms and SIMULINK blocks library, are composed in order to compare the transient processes of dissolved oxygen concentration in the simplified stationary model with constant sensitivity and inertia coefficients, and in the non-stationary model with variable sensitivity and inertia indexes. The simulation block-diagram for non-stationary model adoption to the variable parameters is developed, using informative links from input variable Lg(t), from variable load q(t) and feedback from output variable C(t) as inputs of calculation modulus, allowing to instantly re-calculate the variable indexes during simulation time. Comparison of the simplified stationary model and the non-stationary model shows that the simulation results of oxygen transfer differ up to 50%.
Patel, Deepak K; Waas, Anthony M
2016-07-13
This paper is concerned with predicting the progressive damage and failure of multi-layered hybrid textile composites subjected to uniaxial tensile loading, using a novel two-scale computational mechanics framework. These composites include three-dimensional woven textile composites (3DWTCs) with glass, carbon and Kevlar fibre tows. Progressive damage and failure of 3DWTCs at different length scales are captured in the present model by using a macroscale finite-element (FE) analysis at the representative unit cell (RUC) level, while a closed-form micromechanics analysis is implemented simultaneously at the subscale level using material properties of the constituents (fibre and matrix) as input. The N-layers concentric cylinder (NCYL) model (Zhang and Waas 2014 Acta Mech. 225, 1391-1417; Patel et al. submitted Acta Mech.) to compute local stress, srain and displacement fields in the fibre and matrix is used at the subscale. The 2-CYL fibre-matrix concentric cylinder model is extended to fibre and (N-1) matrix layers, keeping the volume fraction constant, and hence is called the NCYL model where the matrix damage can be captured locally within each discrete layer of the matrix volume. The influence of matrix microdamage at the subscale causes progressive degradation of fibre tow stiffness and matrix stiffness at the macroscale. The global RUC stiffness matrix remains positive definite, until the strain softening response resulting from different failure modes (such as fibre tow breakage, tow splitting in the transverse direction due to matrix cracking inside tow and surrounding matrix tensile failure outside of fibre tows) are initiated. At this stage, the macroscopic post-peak softening response is modelled using the mesh objective smeared crack approach (Rots et al. 1985 HERON 30, 1-48; Heinrich and Waas 2012 53rd AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics and Materials Conference, Honolulu, HI, 23-26 April 2012 AIAA 2012-1537). Manufacturing
Geometry Modeling and Adaptive Control of Air-Breathing Hypersonic Vehicles
NASA Astrophysics Data System (ADS)
Vick, Tyler Joseph
Air-breathing hypersonic vehicles have the potential to provide global reach and affordable access to space. Recent technological advancements have made scramjet-powered flight achievable, as evidenced by the successes of the X-43A and X-51A flight test programs over the last decade. Air-breathing hypersonic vehicles present unique modeling and control challenges in large part due to the fact that scramjet propulsion systems are highly integrated into the airframe, resulting in strongly coupled and often unstable dynamics. Additionally, the extreme flight conditions and inability to test fully integrated vehicle systems larger than X-51 before flight leads to inherent uncertainty in hypersonic flight. This thesis presents a means to design vehicle geometries, simulate vehicle dynamics, and develop and analyze control systems for hypersonic vehicles. First, a software tool for generating three-dimensional watertight vehicle surface meshes from simple design parameters is developed. These surface meshes are compatible with existing vehicle analysis tools, with which databases of aerodynamic and propulsive forces and moments can be constructed. A six-degree-of-freedom nonlinear dynamics simulation model which incorporates this data is presented. Inner-loop longitudinal and lateral control systems are designed and analyzed utilizing the simulation model. The first is an output feedback proportional-integral linear controller designed using linear quadratic regulator techniques. The second is a model reference adaptive controller (MRAC) which augments this baseline linear controller with an adaptive element. The performance and robustness of each controller are analyzed through simulated time responses to angle-of-attack and bank angle commands, while various uncertainties are introduced. The MRAC architecture enables the controller to adapt in a nonlinear fashion to deviations from the desired response, allowing for improved tracking performance, stability, and
Lazarov, R; Pasciak, J; Jones, J
2002-02-01
Construction, analysis and numerical testing of efficient solution techniques for solving elliptic PDEs that allow for parallel implementation have been the focus of the research. A number of discretization and solution methods for solving second order elliptic problems that include mortar and penalty approximations and domain decomposition methods for finite elements and finite volumes have been investigated and analyzed. Techniques for parallel domain decomposition algorithms in the framework of PETC and HYPRE have been studied and tested. Hierarchical parallel grid refinement and adaptive solution methods have been implemented and tested on various model problems. A parallel code implementing the mortar method with algebraically constructed multiplier spaces was developed.
The Adaptive Calibration Model of stress responsivity
Ellis, Bruce J.; Shirtcliff, Elizabeth A.
2010-01-01
This paper presents the Adaptive Calibration Model (ACM), an evolutionary-developmental theory of individual differences in the functioning of the stress response system. The stress response system has three main biological functions: (1) to coordinate the organism’s allostatic response to physical and psychosocial challenges; (2) to encode and filter information about the organism’s social and physical environment, mediating the organism’s openness to environmental inputs; and (3) to regulate the organism’s physiology and behavior in a broad range of fitness-relevant areas including defensive behaviors, competitive risk-taking, learning, attachment, affiliation and reproductive functioning. The information encoded by the system during development feeds back on the long-term calibration of the system itself, resulting in adaptive patterns of responsivity and individual differences in behavior. Drawing on evolutionary life history theory, we build a model of the development of stress responsivity across life stages, describe four prototypical responsivity patterns, and discuss the emergence and meaning of sex differences. The ACM extends the theory of biological sensitivity to context (BSC) and provides an integrative framework for future research in the field. PMID:21145350
Adaptation dynamics of the quasispecies model
NASA Astrophysics Data System (ADS)
Jain, Kavita
2009-02-01
We study the adaptation dynamics of an initially maladapted population evolving via the elementary processes of mutation and selection. The evolution occurs on rugged fitness landscapes which are defined on the multi-dimensional genotypic space and have many local peaks separated by low fitness valleys. We mainly focus on the Eigen's model that describes the deterministic dynamics of an infinite number of self-replicating molecules. In the stationary state, for small mutation rates such a population forms a {\\it quasispecies} which consists of the fittest genotype and its closely related mutants. The quasispecies dynamics on rugged fitness landscape follow a punctuated (or step-like) pattern in which a population jumps from a low fitness peak to a higher one, stays there for a considerable time before shifting the peak again and eventually reaches the global maximum of the fitness landscape. We calculate exactly several properties of this dynamical process within a simplified version of the quasispecies model.
European upper mantle tomography: adaptively parameterized models
NASA Astrophysics Data System (ADS)
Schäfer, J.; Boschi, L.
2009-04-01
We have devised a new algorithm for upper-mantle surface-wave tomography based on adaptive parameterization: i.e. the size of each parameterization pixel depends on the local density of seismic data coverage. The advantage in using this kind of parameterization is that a high resolution can be achieved in regions with dense data coverage while a lower (and cheaper) resolution is kept in regions with low coverage. This way, parameterization is everywhere optimal, both in terms of its computational cost, and of model resolution. This is especially important for data sets with inhomogenous data coverage, as it is usually the case for global seismic databases. The data set we use has an especially good coverage around Switzerland and over central Europe. We focus on periods from 35s to 150s. The final goal of the project is to determine a new model of seismic velocities for the upper mantle underlying Europe and the Mediterranean Basin, of resolution higher than what is currently found in the literature. Our inversions involve regularization via norm and roughness minimization, and this in turn requires that discrete norm and roughness operators associated with our adaptive grid be precisely defined. The discretization of the roughness damping operator in the case of adaptive parameterizations is not as trivial as it is for the uniform ones; important complications arise from the significant lateral variations in the size of pixels. We chose to first define the roughness operator in a spherical harmonic framework, and subsequently translate it to discrete pixels via a linear transformation. Since the smallest pixels we allow in our parameterization have a size of 0.625 °, the spherical-harmonic roughness operator has to be defined up to harmonic degree 899, corresponding to 810.000 harmonic coefficients. This results in considerable computational costs: we conduct the harmonic-pixel transformations on a small Beowulf cluster. We validate our implementation of adaptive
Jablonowski, Christiane
2015-07-14
The research investigates and advances strategies how to bridge the scale discrepancies between local, regional and global phenomena in climate models without the prohibitive computational costs of global cloud-resolving simulations. In particular, the research explores new frontiers in computational geoscience by introducing high-order Adaptive Mesh Refinement (AMR) techniques into climate research. AMR and statically-adapted variable-resolution approaches represent an emerging trend for atmospheric models and are likely to become the new norm in future-generation weather and climate models. The research advances the understanding of multi-scale interactions in the climate system and showcases a pathway how to model these interactions effectively with advanced computational tools, like the Chombo AMR library developed at the Lawrence Berkeley National Laboratory. The research is interdisciplinary and combines applied mathematics, scientific computing and the atmospheric sciences. In this research project, a hierarchy of high-order atmospheric models on cubed-sphere computational grids have been developed that serve as an algorithmic prototype for the finite-volume solution-adaptive Chombo-AMR approach. The foci of the investigations have lied on the characteristics of both static mesh adaptations and dynamically-adaptive grids that can capture flow fields of interest like tropical cyclones. Six research themes have been chosen. These are (1) the introduction of adaptive mesh refinement techniques into the climate sciences, (2) advanced algorithms for nonhydrostatic atmospheric dynamical cores, (3) an assessment of the interplay between resolved-scale dynamical motions and subgrid-scale physical parameterizations, (4) evaluation techniques for atmospheric model hierarchies, (5) the comparison of AMR refinement strategies and (6) tropical cyclone studies with a focus on multi-scale interactions and variable-resolution modeling. The results of this research project
Watté, Rodrigo; Aernouts, Ben; Van Beers, Robbe; Herremans, Els; Ho, Quang Tri; Verboven, Pieter; Nicolaï, Bart; Saeys, Wouter
2015-06-29
Monte Carlo methods commonly used in tissue optics are limited to a layered tissue geometry and thus provide only a very rough approximation for many complex media such as biological structures. To overcome these limitations, a Meshed Monte Carlo method with flexible phase function choice (fpf-MC) has been developed to function in a mesh. This algorithm can model the light propagation in any complexly shaped structure, by attributing optical properties to the different mesh elements. Furthermore, this code allows the use of different discretized phase functions for each tissue type, which can be simulated from the microstructural properties of the tissue, in combination with a tool for simulating the bulk optical properties of polydisperse suspensions. As a result, the scattering properties of tissues can be estimated from information on the microstructural properties of the tissue. This is important for the estimation of the bulk optical properties that can be used for the light propagation model, since many types of tissue have never been characterized in literature. The combination of these contributions, made it possible to use the MMC-fpf for modeling the light porapagation in plant tissue. The developed Meshed Monte Carlo code with flexible phase function choice (MMC-fpf) was successfully validated in simulation through comparison with the Monte Carlo code in Multi-Layered tissues (R2 > 0.9999) and experimentally by comparing the measured and simulated reflectance (RMSE = 0.015%) and transmittance (RMSE = 0.0815%) values for tomato leaves.
Adaptable Multivariate Calibration Models for Spectral Applications
THOMAS,EDWARD V.
1999-12-20
Multivariate calibration techniques have been used in a wide variety of spectroscopic situations. In many of these situations spectral variation can be partitioned into meaningful classes. For example, suppose that multiple spectra are obtained from each of a number of different objects wherein the level of the analyte of interest varies within each object over time. In such situations the total spectral variation observed across all measurements has two distinct general sources of variation: intra-object and inter-object. One might want to develop a global multivariate calibration model that predicts the analyte of interest accurately both within and across objects, including new objects not involved in developing the calibration model. However, this goal might be hard to realize if the inter-object spectral variation is complex and difficult to model. If the intra-object spectral variation is consistent across objects, an effective alternative approach might be to develop a generic intra-object model that can be adapted to each object separately. This paper contains recommendations for experimental protocols and data analysis in such situations. The approach is illustrated with an example involving the noninvasive measurement of glucose using near-infrared reflectance spectroscopy. Extensions to calibration maintenance and calibration transfer are discussed.
Robust moving mesh algorithms for hybrid stretched meshes: Application to moving boundaries problems
NASA Astrophysics Data System (ADS)
Landry, Jonathan; Soulaïmani, Azzeddine; Luke, Edward; Ben Haj Ali, Amine
2016-12-01
A robust Mesh-Mover Algorithm (MMA) approach is designed to adapt meshes of moving boundaries problems. A new methodology is developed from the best combination of well-known algorithms in order to preserve the quality of initial meshes. In most situations, MMAs distribute mesh deformation while preserving a good mesh quality. However, invalid meshes are generated when the motion is complex and/or involves multiple bodies. After studying a few MMA limitations, we propose the following approach: use the Inverse Distance Weighting (IDW) function to produce the displacement field, then apply the Geometric Element Transformation Method (GETMe) smoothing algorithms to improve the resulting mesh quality, and use an untangler to revert negative elements. The proposed approach has been proven efficient to adapt meshes for various realistic aerodynamic motions: a symmetric wing that has suffered large tip bending and twisting and the high-lift components of a swept wing that has moved to different flight stages. Finally, the fluid flow problem has been solved on meshes that have moved and they have produced results close to experimental ones. However, for situations where moving boundaries are too close to each other, more improvements need to be made or other approaches should be taken, such as an overset grid method.
An Optimal Control Modification to Model-Reference Adaptive Control for Fast Adaptation
NASA Technical Reports Server (NTRS)
Nguyen, Nhan T.; Krishnakumar, Kalmanje; Boskovic, Jovan
2008-01-01
This paper presents a method that can achieve fast adaptation for a class of model-reference adaptive control. It is well-known that standard model-reference adaptive control exhibits high-gain control behaviors when a large adaptive gain is used to achieve fast adaptation in order to reduce tracking error rapidly. High gain control creates high-frequency oscillations that can excite unmodeled dynamics and can lead to instability. The fast adaptation approach is based on the minimization of the squares of the tracking error, which is formulated as an optimal control problem. The necessary condition of optimality is used to derive an adaptive law using the gradient method. This adaptive law is shown to result in uniform boundedness of the tracking error by means of the Lyapunov s direct method. Furthermore, this adaptive law allows a large adaptive gain to be used without causing undesired high-gain control effects. The method is shown to be more robust than standard model-reference adaptive control. Simulations demonstrate the effectiveness of the proposed method.
A comparison of tetrahedral mesh improvement techniques
Freitag, L.A.; Ollivier-Gooch, C.
1996-12-01
Automatic mesh generation and adaptive refinement methods for complex three-dimensional domains have proven to be very successful tools for the efficient solution of complex applications problems. These methods can, however, produce poorly shaped elements that cause the numerical solution to be less accurate and more difficult to compute. Fortunately, the shape of the elements can be improved through several mechanisms, including face-swapping techniques that change local connectivity and optimization-based mesh smoothing methods that adjust grid point location. The authors consider several criteria for each of these two methods and compare the quality of several meshes obtained by using different combinations of swapping and smoothing. Computational experiments show that swapping is critical to the improvement of general mesh quality and that optimization-based smoothing is highly effective in eliminating very small and very large angles. The highest quality meshes are obtained by using a combination of swapping and smoothing techniques.
NASA Astrophysics Data System (ADS)
Amsallem, David; Tezaur, Radek; Farhat, Charbel
2016-12-01
A comprehensive approach for real-time computations using a database of parametric, linear, projection-based reduced-order models (ROMs) based on arbitrary underlying meshes is proposed. In the offline phase of this approach, the parameter space is sampled and linear ROMs defined by linear reduced operators are pre-computed at the sampled parameter points and stored. Then, these operators and associated ROMs are transformed into counterparts that satisfy a certain notion of consistency. In the online phase of this approach, a linear ROM is constructed in real-time at a queried but unsampled parameter point by interpolating the pre-computed linear reduced operators on matrix manifolds and therefore computing an interpolated linear ROM. The proposed overall model reduction framework is illustrated with two applications: a parametric inverse acoustic scattering problem associated with a mockup submarine, and a parametric flutter prediction problem associated with a wing-tank system. The second application is implemented on a mobile device, illustrating the capability of the proposed computational framework to operate in real-time.
High Performance Woven Mesh Heat Exchangers
NASA Astrophysics Data System (ADS)
Wirtz, Richard A.; Li, Chen; Park, Ji-Wook; Xu, Jun
2002-07-01
Simple-to-fabricate woven mesh structures, consisting of bonded laminates of two-dimensional plain-weave conductive screens, or three-dimensional orthogonal weaves are described. Geometric equations show that these porous matrices can be fabricated to have a wide range of porosity and a highly anisotropic thermal conductivity vector. A mathematical model of the thermal performance of such a mesh, deployed as a heat exchange surface, is developed. Measurements of pressure drop and overall heat transfer rate are reported and used with the performance model to develop correlation equations of mesh friction factor and Colburn j-factor as a function of coolant properties, mesh characteristics and flow rate through the mesh. A heat exchanger performance analysis delineates conditions where the two mesh technologies offer superior performance.
An Implicit 2-D Shallow Water Flow Model on Unstructured Quadtree Rectangular Mesh
2011-01-01
Hanson, H.; Wamsley, T., and Zundel, A. K., 2006. Two-dimensional depth-averaged circulation model CMS- M2D : Version 3.0, Report 2: Sediment...Militello, A.; Reed, C.W.; Zundel, A.K. and Kraus, N.C., 2004. Two-dimensional depth-averaged circulation model M2D : Version 2.0, Report 1, Technical
Monthly mean simulation experiments with a course-mesh global atmospheric model
NASA Technical Reports Server (NTRS)
Spar, J.; Klugman, R.; Lutz, R. J.; Notario, J. J.
1978-01-01
Substitution of observed monthly mean sea-surface temperatures (SSTs) as lower boundary conditions, in place of climatological SSTs, failed to improve the model simulations. While the impact of SST anomalies on the model output is greater at sea level than at upper levels the impact on the monthly mean simulations is not beneficial at any level. Shifts of one and two days in initialization time produced small, but non-trivial, changes in the model-generated monthly mean synoptic fields. No improvements in the mean simulations resulted from the use of either time-averaged initial data or re-initialization with time-averaged early model output. The noise level of the model, as determined from a multiple initial state perturbation experiment, was found to be generally low, but with a noisier response to initial state errors in high latitudes than the tropics.
Fracture-Based Mesh Size Requirements for Matrix Cracks in Continuum Damage Mechanics Models
NASA Technical Reports Server (NTRS)
Leone, Frank A.; Davila, Carlos G.; Mabson, Gerald E.; Ramnath, Madhavadas; Hyder, Imran
2017-01-01
This paper evaluates the ability of progressive damage analysis (PDA) finite element (FE) models to predict transverse matrix cracks in unidirectional composites. The results of the analyses are compared to closed-form linear elastic fracture mechanics (LEFM) solutions. Matrix cracks in fiber-reinforced composite materials subjected to mode I and mode II loading are studied using continuum damage mechanics and zero-thickness cohesive zone modeling approaches. The FE models used in this study are built parametrically so as to investigate several model input variables and the limits associated with matching the upper-bound LEFM solutions. Specifically, the sensitivity of the PDA FE model results to changes in strength and element size are investigated.
MeshEZW: an image coder using mesh and finite elements
NASA Astrophysics Data System (ADS)
Landais, Thomas; Bonnaud, Laurent; Chassery, Jean-Marc
2003-08-01
In this paper, we present a new method to compress the information in an image, called MeshEZW. The proposed approach is based on the finite elements method, a mesh construction and a zerotree method. The zerotree method is an adaptive of the EZW algorithm with two new symbols for increasing the performance. These steps allow a progressive representation of the image by the automatic construction of a bitstream. The mesh structure is adapted to the image compression domain and is defined to allow video comrpession. The coder is described and some preliminary results are discussed.
Daghighi, Seyedmojtaba; Sjollema, Jelmer; Jaspers, Valery; de Boer, Leonie; Zaat, Sebastian A J; Dijkstra, Rene J B; van Dam, Gooitzen M; van der Mei, Henny C; Busscher, Henk J
2012-11-01
Biomaterials are increasingly used for the restoration of human function, but can become infected as a result of peri- or early post-operative bacterial contamination, although biomaterial-associated infections (BAIs) can also initiate at any time from hematogenous spreading of bacteria from an infection elsewhere in the body. Infecting bacteria in BAIs not only seek shelter in their own protective biofilm matrix, but also hide in surrounding tissue. This study compares staphylococcal persistence on and around a degradable and non-degradable surgical mesh through the use of longitudinal bioluminescence imaging in a murine model, including histological evaluation of surrounding tissue after sacrifice. Surgical meshes were first contaminated with bioluminescent Staphylococcus aureus Xen29 and subsequently subcutaneously implanted in mice. Bioluminescent staphylococci persisted on and around non-degradable meshes during the 28-day course of the study, whereas bioluminescence returned to control levels and bacteria disappeared from surrounding tissues once a degradable mesh had fully dissolved. Thus the application of degradable biomaterials yields major advantages with respect to the prevention of BAIs, as dissolution of the implant not only is associated with elimination of the protective biofilm mode of growth of the infecting organisms, but also allows the immune system to clear the surrounding tissue from infecting organisms.
NASA Astrophysics Data System (ADS)
Kong, Song-Charng; Reitz, Rolf D.
2003-06-01
This study used a numerical model to investigate the combustion process in a premixed iso-octane homogeneous charge compression ignition (HCCI) engine. The engine was a supercharged Cummins C engine operated under HCCI conditions. The CHEMKIN code was implemented into an updated KIVA-3V code so that the combustion could be modelled using detailed chemistry in the context of engine CFD simulations. The model was able to accurately simulate the ignition timing and combustion phasing for various engine conditions. The unburned hydrocarbon emissions were also well predicted while the carbon monoxide emissions were under predicted. Model results showed that the majority of unburned hydrocarbon is located in the piston-ring crevice region and the carbon monoxide resides in the vicinity of the cylinder walls. A sensitivity study of the computational grid resolution indicated that the combustion predictions were relatively insensitive to the grid density. However, the piston-ring crevice region needed to be simulated with high resolution to obtain accurate emissions predictions. The model results also indicated that HCCI combustion and emissions are very sensitive to the initial mixture temperature. The computations also show that the carbon monoxide emissions prediction can be significantly improved by modifying a key oxidation reaction rate constant.
3D-Meshes aus medizinischen Volumendaten
NASA Astrophysics Data System (ADS)
Zelzer, Sascha; Meinzer, Hans-Peter
Diese Arbeit beschreibt eine template-basierte Methode zur Erzeugung von adaptiven Hexaeder-Meshes aus Volumendaten, welche komplizierte konkave Strukturen aufweisen können. Es wird ein vollständiger Satz von Templates generiert der es erlaubt, die Ränder konkaver Regionen feiner zu zerlegen als angrenzende Bereiche und somit die Gesamtzahl an Hexaeder verringert. Der Algorithmus arbeitet mit beliebigen gelabelten Volumendaten und erzeugt ein adaptives, konformes, reines Hexaeder-Mesh.
Montessori Infant and Toddler Programs: How Our Approach Meshes with Other Models
ERIC Educational Resources Information Center
Miller, Darla Ferris
2011-01-01
Today, Montessori infant & toddler programs around the country usually have a similar look and feel--low floor beds, floor space for movement, low shelves, natural materials, tiny wooden chairs and tables for eating, and not a highchair or swing in sight. But Montessori toddler programs seem to fall into two paradigms--one model seeming more…
A Sharing Item Response Theory Model for Computerized Adaptive Testing
ERIC Educational Resources Information Center
Segall, Daniel O.
2004-01-01
A new sharing item response theory (SIRT) model is presented that explicitly models the effects of sharing item content between informants and test takers. This model is used to construct adaptive item selection and scoring rules that provide increased precision and reduced score gains in instances where sharing occurs. The adaptive item selection…
Adapting the ALP Model for Student and Institutional Needs
ERIC Educational Resources Information Center
Sides, Meredith
2016-01-01
With the increasing adoption of accelerated models of learning comes the necessary step of adapting these models to fit the unique needs of the student population at each individual institution. One such college adapted the ALP (Accelerated Learning Program) model and made specific changes to the target population, structure and scheduling, and…
A Conceptual Model of Childhood Adaptation to Type 1 Diabetes
Whittemore, Robin; Jaser, Sarah; Guo, Jia; Grey, Margaret
2010-01-01
The Childhood Adaptation Model to Chronic Illness: Diabetes Mellitus was developed to identify factors that influence childhood adaptation to type 1 diabetes (T1D). Since this model was proposed, considerable research has been completed. The purpose of this paper is to update the model on childhood adaptation to T1D using research conducted since the original model was proposed. The framework suggests that individual and family characteristics, such as age, socioeconomic status, and in children with T1D, treatment modality (pump vs. injections), psychosocial responses (depressive symptoms and anxiety), and individual and family responses (self-management, coping, self-efficacy, family functioning, social competence) influence the level of adaptation. Adaptation has both physiologic (metabolic control) and psychosocial (QOL) components. This revised model provides greater specificity to the factors that influence adaptation to chronic illness in children. Research and clinical implications are discussed. PMID:20934079
Meshing human resources planning with strategic business planning: a model approach.
Baird, L; Meshoulam, I; DeGive, G
1983-01-01
Many people are touting the need to take a strategic approach to human resources management--but this is more easily said than done, point out authors Lloyd Baird, associate professor of management at the School of Management and the Human Resources Policy Institute at Boston University; Ilan Meshoulam, personnel executive at Digital Equipment Corporation; and Ghislaine DeGive, Boston University. Obviously, people cannot be moved around as easily or as speedily as can other resources to help move a company to a strategically identified future position. Thus a great deal of planning is required. The authors present an integrative human resources strategic planning model that focuses on four elements: the environment, the organization's culture, the corporate mission statement, and overall corporate strategy. Each part of the model is thoroughly discussed, and checklist questions are included to help the individual manager develop and implement human resources management plans for his or her own organization.
NASA Astrophysics Data System (ADS)
Miyamoto, K.
2005-12-01
I investigate how the intensity and the activity of mid-latitude cyclones change as a result of global warming, based on a time-slice experiment with a super-high resolution Atmospheric General Circulation Model (20-km mesh TL959L60 MRI/JMA AGCM). The model was developed by the RR2002 project "Development of Super High Resolution Global and Regional Climate Models" funded by the Japanese Ministry of Education, Culture, Sports, Science and Technology. In this context, I use a 10-year control simulation with the climatological SST and a 10-year time-slice global warming simulation using the SST anomalies derived from the SRES A1B scenario run with the MRI-CGCM2.3 (T42L30 atmosphere, 0.5-2.0 x 2.5 L23 ocean) corresponding to the end of the 21st century. I have analyzed the sea-level pressure field and the kinetic energy field of the wind at the 500 hPa pressure level associated with mid-latitude transients from October through April. According to a comparison of 10-day average fields between present and future in the North Pacific, some statistically significant changes are found in a warmer climate for the both of sea-level pressure and the kinetic energy fields. In particular, from late winter through early spring, the sea-level pressure decreases on many parts of the whole Pacific. The kinetic energy of the wind becomes higher on center of the basin. Therefore, I suppose the Aleutian Low is likely to settle in longer by about one month than the present. Hereafter, I plan to investigate what kind of phenomena may accompany the changes on mid-latitude transients.
Scale-adaptive subgrid-scale modelling for large-eddy simulation of turbulent flows
NASA Astrophysics Data System (ADS)
Yu, Changping; Xiao, Zuoli; Li, Xinliang
2017-03-01
The proportionality between the subgrid-scale (SGS) drain rate of kinetic energy and the viscous dissipation rate of the resolved motions is studied a priori by filtering a given fully resolved field and evaluating a generic form of the hypothesized energy spectrum. The ratio of the SGS drain to the resolved dissipation, on which a balance condition for the SGS dissipation across an arbitrary grid scale is founded, is shown to be independent of the turbulence Reynolds number, and can be described by a function in terms of the averaged mesh Reynolds number. Such a balance condition can serve as a physical constraint in the SGS modeling to account for the scale effects of the model coefficient(s). Scale-adaptive dynamic Smagorinsky-Lilly model and mixed nonlinear model are formulated for large-eddy simulation of transitional and/or turbulent flows in such a way that the constraint is satisfied. The newly proposed scale-adaptive dynamic SGS models are validated in simulations of homogeneous isotropic turbulence and turbulent channel flow, and prove to be superior over traditional dynamic SGS models.
Summary of results of January climate simulations with the GISS coarse-mesh model
NASA Technical Reports Server (NTRS)
Spar, J.; Cohen, C.; Wu, P.
1981-01-01
The large scale climates generated by extended runs of the model are relatively independent of the initial atmospheric conditions, if the first few months of each simulation are discarded. The perpetual January simulations with a specified SST field produced excessive snow accumulation over the continents of the Northern Hemisphere. Mass exchanges between the cold (warm) continents and the warm (cold) adjacent oceans produced significant surface pressure changes over the oceans as well as over the land. The effect of terrain and terrain elevation on the amount of precipitation was examined. The evaporation of continental moisture was calculated to cause large increases in precipitation over the continents.
Applicable Adaptive Testing Models for School Teachers.
ERIC Educational Resources Information Center
Wang, Albert Chang-hwa; Chuang, Chi-lin
2002-01-01
Describes a study conducted in Taipei (Taiwan) that investigated the attitudinal effects of SPRT (Sequential Probability Ratio Test) adaptive testing environment on junior high school students. Discusses test anxiety; student preferences; test adaptability; acceptance of test results; number of items answered; and computer experience. (Author/LRW)
MOAB : a mesh-oriented database.
Tautges, Timothy James; Ernst, Corey; Stimpson, Clint; Meyers, Ray J.; Merkley, Karl
2004-04-01
A finite element mesh is used to decompose a continuous domain into a discretized representation. The finite element method solves PDEs on this mesh by modeling complex functions as a set of simple basis functions with coefficients at mesh vertices and prescribed continuity between elements. The mesh is one of the fundamental types of data linking the various tools in the FEA process (mesh generation, analysis, visualization, etc.). Thus, the representation of mesh data and operations on those data play a very important role in FEA-based simulations. MOAB is a component for representing and evaluating mesh data. MOAB can store structured and unstructured mesh, consisting of elements in the finite element 'zoo'. The functional interface to MOAB is simple yet powerful, allowing the representation of many types of metadata commonly found on the mesh. MOAB is optimized for efficiency in space and time, based on access to mesh in chunks rather than through individual entities, while also versatile enough to support individual entity access. The MOAB data model consists of a mesh interface instance, mesh entities (vertices and elements), sets, and tags. Entities are addressed through handles rather than pointers, to allow the underlying representation of an entity to change without changing the handle to that entity. Sets are arbitrary groupings of mesh entities and other sets. Sets also support parent/child relationships as a relation distinct from sets containing other sets. The directed-graph provided by set parent/child relationships is useful for modeling topological relations from a geometric model or other metadata. Tags are named data which can be assigned to the mesh as a whole, individual entities, or sets. Tags are a mechanism for attaching data to individual entities and sets are a mechanism for describing relations between entities; the combination of these two mechanisms is a powerful yet simple interface for representing metadata or application
Application of Hyperelastic-based Active Mesh Model in Cardiac Motion Recovery
Yousefi-Banaem, Hossein; Kermani, Saeed; Daneshmehr, Alireza; Saneie, Hamid
2016-01-01
Considering the nonlinear hyperelastic or viscoelastic nature of soft tissues has an important effect on modeling results. In medical applications, accounting nonlinearity begets an ill posed problem, due to absence of external force. Myocardium can be considered as a hyperelastic material, and variational approaches are proposed to estimate stiffness matrix, which take into account the linear and nonlinear properties of myocardium. By displacement estimation of some points in the four-dimensional cardiac magnetic resonance imaging series, using a similarity criterion, the elementary deformations are estimated, then using the Moore–Penrose inverse matrix approach, all point deformations are obtained. Using this process, the cardiac wall motion is quantized to mechanically determine local parameters to investigate the cardiac wall functionality. This process was implemented and tested over 10 healthy and 20 patients with myocardial infarction. In all patients, the process was able to precisely determine the affected region. The proposed approach was also compared with linear one and the results demonstrated its superiority respect to the linear model. PMID:27563570
Modeling adaptive non-repudiation security services
NASA Astrophysics Data System (ADS)
Tunia, Marcin A.
2016-09-01
Non-repudiation security service helps to protect an electronic system against false denial of performing certain actions by the participants of communication involving that system. In the development process of such a security service it is important to implement all the necessary elements and adapt them in accordance with the defined protection scope. There are several types of non-repudiation and cases of their use. In this paper the author focuses on the situation when a nonrepudiation service is being implemented on an application server. All necessary actions of the users i.e. of people and/or machines are recorded by the service. Currently a new type of security services called "context-aware security services" is under studies. This type of service involves acquiring and processing additional information in order to provide flexible protection. This paper presents the elements of non-repudiation security service and how the elements can be modeled with context awareness support including reputation systems.
Multiclient Identification System Using Adaptive Probabilistic Model
NASA Astrophysics Data System (ADS)
Lin, Chin-Teng; Siana, Linda; Shou, Yu-Wen; Yang, Chien-Ting
2010-12-01
This paper aims at integrating detection and identification of human faces in a more practical and real-time face recognition system. The proposed face detection system is based on the cascade Adaboost method to improve the precision and robustness toward unstable surrounding lightings. Our Adaboost method innovates to adjust the environmental lighting conditions by histogram lighting normalization and to accurately locate the face regions by a region-based-clustering process as well. We also address on the problem of multi-scale faces in this paper by using 12 different scales of searching windows and 5 different orientations for each client in pursuit of the multi-view independent face identification. There are majorly two methodological parts in our face identification system, including PCA (principal component analysis) facial feature extraction and adaptive probabilistic model (APM). The structure of our implemented APM with a weighted combination of simple probabilistic functions constructs the likelihood functions by the probabilistic constraint in the similarity measures. In addition, our proposed method can online add a new client and update the information of registered clients due to the constructed APM. The experimental results eventually show the superior performance of our proposed system for both offline and real-time online testing.
NASA Astrophysics Data System (ADS)
Schmalzl, J.
2003-04-01
Convective flows govern much of the dynamics of the Earth. Examples of such flows are convection in the Earth's mantle, convection in magma chambers and much of the dynamics of the world oceans. Nowadays these time-dependent flows are often studied by means of three dimensional (3D) numerical models which solve the equations for the transport of heat and momentum alternatingly. These flows are often driven by a temperature difference. But for many flows there is also an active or passive chemical component that has to be considered. One characteristics of these flows is that the chemical diffusivity is very small. Implementing such a chemical field with a very low diffusivity into a numerical model using a field approach is difficult due to numerical diffusion introduced by the Eulerian schemes. Using Lagrangian tracers is also difficult in 3D flow since a massive amount of tracers is needed. We therefore have implemented a tracer-mesh method which tracks only the position of the interface between the two different components. Compared to a 2D tracer-line the insertion of new 3D surface-elements in highly deformed regions is however more complex. This is due to the topology of the mesh which changes because of the adaptive refinement. Luckily the refinement of polygonal meshes is a very active field of research in computer graphics and has been termed "Subdivision Surfaces". There is a wealth of different subdivision schemes with different properties. We applied the Butterfly scheme and the Loop scheme for the refinement of tracer meshes. When a density difference is connected to a chemical component it often acts as a restoring force. In many cases, the governing flow is spatially heterogenous and the spatial location of the heterogeneities is varying in time (e.g. the location of an upwelling plume). The restoring force of the density contrast may result in a situation where a highly deformed, and therefore highly refined region, returns to a simple geometry. In
NASA Astrophysics Data System (ADS)
Schmalzl, J.; Loddoch, A.
2003-12-01
Convective flows govern much of the dynamics of the Earth. Examples of such flows are convection in the Earth's mantle, convection in magma chambers and much of the dynamics of the world oceans. Nowadays these time-dependent flows are often studied by means of three dimensional (3D) numerical models which solve the equations for the transport of heat and momentum alternatingly. These flows are often driven by a temperature difference. But for many flows there is also an active or passive chemical component that has to be considered. One characteristics of these flows is that the chemical diffusivity is very small. Implementing such a chemical field with a very low diffusivity into a numerical model using a field approach is difficult due to numerical diffusion introduced by the Eulerian schemes. Using Lagrangian tracers is also difficult in 3D flow since a massive amount of tracers is needed. We therefore have implemented a tracer-mesh method which tracks only the position of the interface between the two different components. Compared to a 2D tracer-line the insertion of new 3D surface-elements in highly deformed regions is however more complex. This is due to the topology of the mesh which changes because of the adaptive refinement. Luckily the refinement of polygonal meshes is a very active field of research in computer graphics and has been termed "Subdivision Surfaces". There is a wealth of different subdivision schemes with different properties. We applied the Butterfly scheme and the Loop scheme for the refinement of tracer meshes. When a density difference is connected to a chemical component it often acts as a restoring force. In many cases, the governing flow is spatially heterogenous and the spatial location of the heterogeneities is varying in time (e.g. the location of an upwelling plume). The restoring force of the density contrast may result in a situation where a highly deformed, and therefore highly refined region, returns to a simple geometry. In
NASA Astrophysics Data System (ADS)
Bargatze, L. F.
2015-12-01
Active Data Archive Product Tracking (ADAPT) is a collection of software routines that permits one to generate XML metadata files to describe and register data products in support of the NASA Heliophysics Virtual Observatory VxO effort. ADAPT is also a philosophy. The ADAPT concept is to use any and all available metadata associated with scientific data to produce XML metadata descriptions in a consistent, uniform, and organized fashion to provide blanket access to the full complement of data stored on a targeted data server. In this poster, we present an application of ADAPT to describe all of the data products that are stored by using the Common Data File (CDF) format served out by the CDAWEB and SPDF data servers hosted at the NASA Goddard Space Flight Center. These data servers are the primary repositories for NASA Heliophysics data. For this purpose, the ADAPT routines have been used to generate data resource descriptions by using an XML schema named Space Physics Archive, Search, and Extract (SPASE). SPASE is the designated standard for documenting Heliophysics data products, as adopted by the Heliophysics Data and Model Consortium. The set of SPASE XML resource descriptions produced by ADAPT includes high-level descriptions of numerical data products, display data products, or catalogs and also includes low-level "Granule" descriptions. A SPASE Granule is effectively a universal access metadata resource; a Granule associates an individual data file (e.g. a CDF file) with a "parent" high-level data resource description, assigns a resource identifier to the file, and lists the corresponding assess URL(s). The CDAWEB and SPDF file systems were queried to provide the input required by the ADAPT software to create an initial set of SPASE metadata resource descriptions. Then, the CDAWEB and SPDF data repositories were queried subsequently on a nightly basis and the CDF file lists were checked for any changes such as the occurrence of new, modified, or deleted
Modeling Adaptation as a Flow and Stock Decsion with Mitigation
Mitigation and adaptation are the two key responses available to policymakers to reduce the risks of climate change. We model these two policies together in a new DICE-based integrated assessment model that characterizes adaptation as either short-lived flow spending or long-live...
Modeling Two Types of Adaptation to Climate Change
Mitigation and adaptation are the two key responses available to policymakers to reduce the risks of climate change. We model these two policies together in a new DICE-based integrated assessment model that characterizes adaptation as either short-lived flow spending or long-live...
Modeling Adaptation as a Flow and Stock Decision with Mitigation
Mitigation and adaptation are the two key responses available to policymakers to reduce the risks of climate change. We model these two policies together in a new DICE-based integrated assessment model that characterizes adaptation as either short-lived flow spending or long-liv...
Adaptive Networks Foundations: Modeling, Dynamics, and Applications
2013-02-13
22-Mar. 2, 2012. • Shaw, L.B., Long, Y., and Gross, T. Simultaneous spread of infection and information in adaptive networks. Casablanca ...International Workshop on Mathematical Biology, Casablanca , Morocco, Jun. 20-24, 2011. • Tunc, I. and Shaw, L.B. Dynamics of infection spreading in adaptive...Defense The number of undergraduates funded by your agreement who graduated during this period and will receive scholarships or fellowships for further
Gradient-based adaptation of continuous dynamic model structures
NASA Astrophysics Data System (ADS)
La Cava, William G.; Danai, Kourosh
2016-01-01
A gradient-based method of symbolic adaptation is introduced for a class of continuous dynamic models. The proposed model structure adaptation method starts with the first-principles model of the system and adapts its structure after adjusting its individual components in symbolic form. A key contribution of this work is its introduction of the model's parameter sensitivity as the measure of symbolic changes to the model. This measure, which is essential to defining the structural sensitivity of the model, not only accommodates algebraic evaluation of candidate models in lieu of more computationally expensive simulation-based evaluation, but also makes possible the implementation of gradient-based optimisation in symbolic adaptation. The proposed method is applied to models of several virtual and real-world systems that demonstrate its potential utility.
Brandt, Christopher J; Kammer, Daniel; Fiebeler, Anette; Klinge, Uwe
2011-12-01
Prosthetic reinforcements markedly reduce the risk of hernia recurrence. However, the implantation of meshes is related to an inflammatory foreign body reaction (FBR) with serious complications (i.e., persistent seroma, wound infection, mesh migration, entrapment, chronic pain). Adrenal hormones profoundly modify inflammatory response. Their effects on FBR however, remain ill defined. We therefore studied the FBR to polyvinylidenfluoride (PVDF) mesh material that was coated with four different substances: hydrocortisone (COR) or mifepristone (MIF), which respectively stimulate and block the glucocorticoid receptor, and aldosterone (ALD) or spironolactone (SPI), which respectively stimulate and block the mineralocorticoid receptor. The coated substances were released from the meshes within 24 h. Seven, 21, and 90 days after implantation, the specimen was evaluated for collagen formation, granuloma size, inflammatory activity, and angiogenesis. COR and SPI coating protected from inflammatory response, while ALD and MIF coating showed little difference to the control group. The COR and SPI groups showed smaller granuloma sizes at all time points, as well as a reduced number of inflammatory cells (p < 0.001) at day 90, and decreased collagen formation starting after 21 days (p < 0.05). There was a negative correlation for angiogenesis with inflammation around foreign body structures. In summary, these results suggest that early and temporary stimulation of the glucocorticoid receptor or blockade of the mineralocorticoid receptor have beneficial effects on FBR in the long term.
Consensus time and conformity in the adaptive voter model
NASA Astrophysics Data System (ADS)
Rogers, Tim; Gross, Thilo
2013-09-01
The adaptive voter model is a paradigmatic model in the study of opinion formation. Here we propose an extension for this model, in which conflicts are resolved by obtaining another opinion, and analytically study the time required for consensus to emerge. Our results shed light on the rich phenomenology of both the original and extended adaptive voter models, including a dynamical phase transition in the scaling behavior of the mean time to consensus.
CUBIT mesh generation environment. Volume 1: Users manual
Blacker, T.D.; Bohnhoff, W.J.; Edwards, T.L.
1994-05-01
The CUBIT mesh generation environment is a two- and three-dimensional finite element mesh generation tool which is being developed to pursue the goal of robust and unattended mesh generation--effectively automating the generation of quadrilateral and hexahedral elements. It is a solid-modeler based preprocessor that meshes volume and surface solid models for finite element analysis. A combination of techniques including paving, mapping, sweeping, and various other algorithms being developed are available for discretizing the geometry into a finite element mesh. CUBIT also features boundary layer meshing specifically designed for fluid flow problems. Boundary conditions can be applied to the mesh through the geometry and appropriate files for analysis generated. CUBIT is specifically designed to reduce the time required to create all-quadrilateral and all-hexahedral meshes. This manual is designed to serve as a reference and guide to creating finite element models in the CUBIT environment.
NASA Astrophysics Data System (ADS)
Schubert, J. E.; Sanders, B. F.
2011-12-01
Urban landscapes are at the forefront of current research efforts in the field of flood inundation modeling for two major reasons. First, urban areas hold relatively large economic and social importance and as such it is imperative to avoid or minimize future damages. Secondly, urban flooding is becoming more frequent as a consequence of continued development of impervious surfaces, population growth in cities, climate change magnifying rainfall intensity, sea level rise threatening coastal communities, and decaying flood defense infrastructure. In reality urban landscapes are particularly challenging to model because they include a multitude of geometrically complex features. Advances in remote sensing technologies and geographical information systems (GIS) have promulgated fine resolution data layers that offer a site characterization suitable for urban inundation modeling including a description of preferential flow paths, drainage networks and surface dependent resistances to overland flow. Recent research has focused on two-dimensional modeling of overland flow including within-curb flows and over-curb flows across developed parcels. Studies have focused on mesh design and parameterization, and sub-grid models that promise improved performance relative to accuracy and/or computational efficiency. This presentation addresses how fine-resolution data, available in Los Angeles County, are used to parameterize, initialize and execute flood inundation models for the 1963 Baldwin Hills dam break. Several commonly used model parameterization strategies including building-resistance, building-block and building hole are compared with a novel sub-grid strategy based on building-porosity. Performance of the models is assessed based on the accuracy of depth and velocity predictions, execution time, and the time and expertise required for model set-up. The objective of this study is to assess field-scale applicability, and to obtain a better understanding of advantages
Bohmer, R D; Byrne, P D; Maddern, G J
2002-07-01
A number of different materials are available for incisional hernia repair. Benefits of the various types are controversial and are partly dependent on the anatomical placement of the mesh. Composite mesh has been introduced to provide tissue ingrowth for strength and a non-adherent side to protect the bowel, these layers being laminated together. This report is on the separation of layers in an infected mesh and adherence of the expanded polytetrafluoroethylene layer to the small bowel.
Applying State-of-the-art, Unstructured mesh Models on Estuary Management in the Danish Wadden Sea
NASA Astrophysics Data System (ADS)
Bundgaard, K.; Lumborg, U.; Vested, H. J.; Edelvang, K.
2006-12-01
projects in the area will be used to show the general behavior of hydrodynamics and fine-grained sediment in the area. This will be demonstrated using measurements, aerial photos and model results. Based on this, a newly developed unstructured flexible mesh model of the area will be demonstrated. It will be shown that the model can depict the physics of the hydrodynamics and the fine-grained sediment, and that a good calibration can be established with regard to the hydrodynamics and suspended sediment. The demonstration will also include the models' capabilities as tools for estuary management. It will be shown how the calibrated model can be used for optimizing dredging procedures and determining their effect on long term sedimentation. A newly developed sediment transport model describing mixtures of cohesive and non-cohesive sediments will be demonstrated. The advantage of this type of model is that it is able to calculate both cohesive and non- cohesive sediment transport simultaneously. This allows a description of effects such as sorting, armoring and changing of erodibility. The model uses a multifraction approach, which allows fractions to be treated differently. This means that the model is capable of calculating the advection dominated transport of fine- grained sediments at the same time as the bed load dominated transport of sand. This makes it a very powerful tool for calculating sediment transport in estuaries with very varying sediment types for example estuaries with mudflats and sandy tidal channels.
Lagrange-mesh calculations in momentum space.
Lacroix, Gwendolyn; Semay, Claude; Buisseret, Fabien
2012-08-01
The Lagrange-mesh method is a powerful method to solve eigenequations written in configuration space. It is very easy to implement and very accurate. Using a Gauss quadrature rule, the method requires only the evaluation of the potential at some mesh points. The eigenfunctions are expanded in terms of regularized Lagrange functions which vanish at all mesh points except one. It is shown that this method can be adapted to solve eigenequations written in momentum space, keeping the convenience and the accuracy of the original technique. In particular, the kinetic operator is a diagonal matrix. Observables and wave functions in both configuration space and momentum space can also be easily computed with good accuracy using only eigenfunctions computed in the momentum space. The method is tested with Gaussian and Yukawa potentials, requiring, respectively, a small and a large mesh to reach convergence. Corresponding wave functions in both spaces are compared with each other using the Fourier transform.
A finite-element mesh generator based on growing neural networks.
Triantafyllidis, D G; Labridis, D P
2002-01-01
A mesh generator for the production of high-quality finite-element meshes is being proposed. The mesh generator uses an artificial neural network, which grows during the training process in order to adapt itself to a prespecified probability distribution. The initial mesh is a constrained Delaunay triangulation of the domain to be triangulated. Two new algorithms to accelerate the location of the best matching unit are introduced. The mesh generator has been found able to produce meshes of high quality in a number of classic cases examined and is highly suited for problems where the mesh density vector can be calculated in advance.
Mesh-driven vector field clustering and visualization: an image-based approach.
Peng, Zhenmin; Grundy, Edward; Laramee, Robert S; Chen, Guoning; Croft, Nick
2012-02-01
Vector field visualization techniques have evolved very rapidly over the last two decades, however, visualizing vector fields on complex boundary surfaces from computational flow dynamics (CFD) still remains a challenging task. In part, this is due to the large, unstructured, adaptive resolution characteristics of the meshes used in the modeling and simulation process. Out of the wide variety of existing flow field visualization techniques, vector field clustering algorithms offer the advantage of capturing a detailed picture of important areas of the domain while presenting a simplified view of areas of less importance. This paper presents a novel, robust, automatic vector field clustering algorithm that produces intuitive and insightful images of vector fields on large, unstructured, adaptive resolution boundary meshes from CFD. Our bottom-up, hierarchical approach is the first to combine the properties of the underlying vector field and mesh into a unified error-driven representation. The motivation behind the approach is the fact that CFD engineers may increase the resolution of model meshes according to importance. The algorithm has several advantages. Clusters are generated automatically, no surface parameterization is required, and large meshes are processed efficiently. The most suggestive and important information contained in the meshes and vector fields is preserved while less important areas are simplified in the visualization. Users can interactively control the level of detail by adjusting a range of clustering distance measure parameters. We describe two data structures to accelerate the clustering process. We also introduce novel visualizations of clusters inspired by statistical methods. We apply our method to a series of synthetic and complex, real-world CFD meshes to demonstrate the clustering algorithm results.
Coupled Elastic-Thermal Dynamics of Deployable Mesh Reflectors
NASA Technical Reports Server (NTRS)
Shi, H.; Yang, B.; Thomson, M.; Fang, H.
2011-01-01
This paper presents a coupled elastic-thermal dynamic model and a quasi-static strategy on the analysis of the reflector dynamics in the space mission. The linearized model, its natural frequencies and mode shapes are then derived upon the nonlinear static equilibrium of the structure. The numerical example is provided to fully adapt the strategy and investigate the dynamic behaviors of the structure. Finally the proposed method is applied on the sample of the deployable mesh reflector and the simulation results are presented. The research work delivered in the paper will be used to design the feedback surface in future.
Modeling Family Adaptation to Fragile X Syndrome
ERIC Educational Resources Information Center
Raspa, Melissa; Bailey, Donald, Jr.; Bann, Carla; Bishop, Ellen
2014-01-01
Using data from a survey of 1,099 families who have a child with Fragile X syndrome, we examined adaptation across 7 dimensions of family life: parenting knowledge, social support, social life, financial impact, well-being, quality of life, and overall impact. Results illustrate that although families report a high quality of life, they struggle…
Zabek, Daniel; Taylor, John; Bowen, Chris
2016-09-05
Flexible pyroelectric energy generators provide unique features for harvesting temperature fluctuations which can be effectively enhanced using meshed electrodes that improve thermal conduction, convection and radiation into the pyroelectric. In this paper, thermal radiation energy is continuously harvested with pyroelectric free standing Polyvilylidene Difluoride (PVDF) films over a large number of heat heat cycles using a novel micro-sized symmetrical patterned meshed electrode. It is shown that, for the meshed electrode geometries considered in this work, the polarisation-field (P-E), current-field (I-E) characteristics and device capacitance are unaffected since the fringing fields were generally small; this is verified using numerical simulations and comparison with experimental measurements. The use of meshed electrodes has been shown to significantly improve both the open circuit voltage (16 V to 59 V) and closed-circuit current (9 nA to 32 nA). The pyroelectric alternating current (AC) is rectified for direct current (DC) storage and 30% reduction in capacitor charging time is achieved by using the optimum meshed electrodes. The use of meshed electrodes on ferroelectric materials provides an innovative route to improve their performance in applications such as wearable devices, novel flexible sensors and large scale pyroelectric energy harvesters.hese instructions give you guidelines for preparing papers for IEEE Transactions and Journals. Use this document as a template if you are using Microsoft Word 6.0 or later. Otherwise, use this document as an instruction set. The electronic file of your paper will be formatted further at IEEE. Paper titles should be written in uppercase and lowercase letters, not all uppercase. Avoid writing long formulas with subscripts in the title; short formulas that identify the elements are fine (e.g., "Nd-Fe-B"). Do not write "(Invited)" in the title. Full names of authors are preferred in the author field, but are not
Particle Swarm Based Collective Searching Model for Adaptive Environment
Cui, Xiaohui; Patton, Robert M; Potok, Thomas E; Treadwell, Jim N
2008-01-01
This report presents a pilot study of an integration of particle swarm algorithm, social knowledge adaptation and multi-agent approaches for modeling the collective search behavior of self-organized groups in an adaptive environment. The objective of this research is to apply the particle swarm metaphor as a model of social group adaptation for the dynamic environment and to provide insight and understanding of social group knowledge discovering and strategic searching. A new adaptive environment model, which dynamically reacts to the group collective searching behaviors, is proposed in this research. The simulations in the research indicate that effective communication between groups is not the necessary requirement for whole self-organized groups to achieve the efficient collective searching behavior in the adaptive environment.